repo
stringlengths
6
65
file_url
stringlengths
81
311
file_path
stringlengths
6
227
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 15:31:58
2026-01-04 20:25:31
truncated
bool
2 classes
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/kms/mod.rs
crates/e2e_test/src/kms/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! KMS (Key Management Service) End-to-End Tests //! //! This module contains comprehensive end-to-end tests for RustFS KMS functionality, //! including tests for both Local and Vault backends. // KMS-specific common utilities #[cfg(test)] pub mod common; #[cfg(test)] mod kms_local_test; #[cfg(test)] mod kms_vault_test; #[cfg(test)] mod kms_comprehensive_test; #[cfg(test)] mod multipart_encryption_test; #[cfg(test)] mod kms_edge_cases_test; #[cfg(test)] mod kms_fault_recovery_test; #[cfg(test)] mod test_runner; #[cfg(test)] mod bucket_default_encryption_test; #[cfg(test)] mod encryption_metadata_test;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/kms/multipart_encryption_test.rs
crates/e2e_test/src/kms/multipart_encryption_test.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // #![allow(clippy::upper_case_acronyms)] // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Step-by-step test cases for sharded upload encryption //! //! This test suite will validate every step of the sharded upload encryption feature: //! 1. Test the underlying single-shard encryption (validate the encryption underlying logic) //! 2. Test multi-shard uploads (verify shard stitching logic) //! 3. Test the saving and reading of encrypted metadata //! 4. Test the complete sharded upload encryption process use super::common::LocalKMSTestEnvironment; use crate::common::{TEST_BUCKET, init_logging}; use serial_test::serial; use tracing::{debug, info}; /// Step 1: Test the basic single-file encryption function (ensure that SSE-S3 works properly in non-sharded scenarios) #[tokio::test] #[serial] async fn test_step1_basic_single_file_encryption() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { init_logging(); info!("πŸ§ͺ Step 1: Test the basic single-file encryption function"); let mut kms_env = LocalKMSTestEnvironment::new().await?; let _default_key_id = kms_env.start_rustfs_for_local_kms().await?; tokio::time::sleep(tokio::time::Duration::from_secs(3)).await; let s3_client = kms_env.base_env.create_s3_client(); kms_env.base_env.create_test_bucket(TEST_BUCKET).await?; // Test small file encryption (should be stored inline) let test_data = b"Hello, this is a small test file for SSE-S3!"; let object_key = "test-single-file-encrypted"; info!("πŸ“€ Upload a small file ({} bytes) with SSE-S3 encryption enabled", test_data.len()); let put_response = s3_client .put_object() .bucket(TEST_BUCKET) .key(object_key) .body(aws_sdk_s3::primitives::ByteStream::from(test_data.to_vec())) .server_side_encryption(aws_sdk_s3::types::ServerSideEncryption::Aes256) .send() .await?; debug!("PUT responds to ETags: {:?}", put_response.e_tag()); debug!("PUT responds to SSE: {:?}", put_response.server_side_encryption()); // Verify that the PUT response contains the correct cipher header assert_eq!( put_response.server_side_encryption(), Some(&aws_sdk_s3::types::ServerSideEncryption::Aes256) ); info!("πŸ“₯ Download the file and verify the encryption status"); let get_response = s3_client.get_object().bucket(TEST_BUCKET).key(object_key).send().await?; debug!("GET responds to SSE: {:?}", get_response.server_side_encryption()); // Verify that the GET response contains the correct cipher header assert_eq!( get_response.server_side_encryption(), Some(&aws_sdk_s3::types::ServerSideEncryption::Aes256) ); // Verify data integrity let downloaded_data = get_response.body.collect().await?.into_bytes(); assert_eq!(&downloaded_data[..], test_data); kms_env.base_env.delete_test_bucket(TEST_BUCKET).await?; info!("βœ… Step 1: The basic single file encryption function is normal"); Ok(()) } /// Step 2: Test the unencrypted shard upload (make sure the shard upload base is working properly) #[tokio::test] #[serial] async fn test_step2_basic_multipart_upload_without_encryption() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { init_logging(); info!("πŸ§ͺ Step 2: Test unencrypted shard uploads"); let mut kms_env = LocalKMSTestEnvironment::new().await?; let _default_key_id = kms_env.start_rustfs_for_local_kms().await?; tokio::time::sleep(tokio::time::Duration::from_secs(3)).await; let s3_client = kms_env.base_env.create_s3_client(); kms_env.base_env.create_test_bucket(TEST_BUCKET).await?; let object_key = "test-multipart-no-encryption"; let part_size = 5 * 1024 * 1024; // 5MB per part (S3 minimum) let total_parts = 2; let total_size = part_size * total_parts; // Generate test data (with obvious patterns for easy verification) let test_data: Vec<u8> = (0..total_size).map(|i| (i % 256) as u8).collect(); info!( "πŸš€ Start sharded upload (unencrypted): {} parts, {}MB each", total_parts, part_size / (1024 * 1024) ); // Step 1: Create a sharded upload let create_multipart_output = s3_client .create_multipart_upload() .bucket(TEST_BUCKET) .key(object_key) .send() .await?; let upload_id = create_multipart_output.upload_id().unwrap(); info!("πŸ“‹ Create a shard upload with ID: {}", upload_id); // Step 2: Upload individual shards let mut completed_parts = Vec::new(); for part_number in 1..=total_parts { let start = (part_number - 1) * part_size; let end = std::cmp::min(start + part_size, total_size); let part_data = &test_data[start..end]; info!("πŸ“€ Upload the shard {} ({} bytes)", part_number, part_data.len()); let upload_part_output = s3_client .upload_part() .bucket(TEST_BUCKET) .key(object_key) .upload_id(upload_id) .part_number(part_number as i32) .body(aws_sdk_s3::primitives::ByteStream::from(part_data.to_vec())) .send() .await?; let etag = upload_part_output.e_tag().unwrap().to_string(); completed_parts.push( aws_sdk_s3::types::CompletedPart::builder() .part_number(part_number as i32) .e_tag(&etag) .build(), ); debug!("Fragment {} upload complete,ETag: {}", part_number, etag); } // Step 3: Complete the shard upload let completed_multipart_upload = aws_sdk_s3::types::CompletedMultipartUpload::builder() .set_parts(Some(completed_parts)) .build(); info!("πŸ”— Complete the shard upload"); let complete_output = s3_client .complete_multipart_upload() .bucket(TEST_BUCKET) .key(object_key) .upload_id(upload_id) .multipart_upload(completed_multipart_upload) .send() .await?; debug!("Complete the shard upload,ETag: {:?}", complete_output.e_tag()); // Step 4: Download and verify info!("πŸ“₯ Download the file and verify data integrity"); let get_response = s3_client.get_object().bucket(TEST_BUCKET).key(object_key).send().await?; let downloaded_data = get_response.body.collect().await?.into_bytes(); assert_eq!(downloaded_data.len(), total_size); assert_eq!(&downloaded_data[..], &test_data[..]); kms_env.base_env.delete_test_bucket(TEST_BUCKET).await?; info!("βœ… Step 2: Unencrypted shard upload functions normally"); Ok(()) } /// Step 3: Test Shard Upload + SSE-S3 Encryption (Focus Test) #[tokio::test] #[serial] async fn test_step3_multipart_upload_with_sse_s3() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { init_logging(); info!("πŸ§ͺ Step 3: Test Shard Upload + SSE-S3 Encryption"); let mut kms_env = LocalKMSTestEnvironment::new().await?; let _default_key_id = kms_env.start_rustfs_for_local_kms().await?; tokio::time::sleep(tokio::time::Duration::from_secs(3)).await; let s3_client = kms_env.base_env.create_s3_client(); kms_env.base_env.create_test_bucket(TEST_BUCKET).await?; let object_key = "test-multipart-sse-s3"; let part_size = 5 * 1024 * 1024; // 5MB per part let total_parts = 2; let total_size = part_size * total_parts; // Generate test data let test_data: Vec<u8> = (0..total_size).map(|i| ((i / 1000) % 256) as u8).collect(); info!( "πŸ” Start sharded upload (SSE-S3 encryption): {} parts, {}MB each", total_parts, part_size / (1024 * 1024) ); // Step 1: Create a shard upload and enable SSE-S3 let create_multipart_output = s3_client .create_multipart_upload() .bucket(TEST_BUCKET) .key(object_key) .server_side_encryption(aws_sdk_s3::types::ServerSideEncryption::Aes256) .send() .await?; let upload_id = create_multipart_output.upload_id().unwrap(); info!("πŸ“‹ Create an encrypted shard upload with ID: {}", upload_id); // Verify the CreateMultipartUpload response (if there is an SSE header) if let Some(sse) = create_multipart_output.server_side_encryption() { debug!("CreateMultipartUpload Contains SSE responses: {:?}", sse); assert_eq!(sse, &aws_sdk_s3::types::ServerSideEncryption::Aes256); } else { debug!("CreateMultipartUpload does not contain SSE response headers (normal in some implementations)"); } // Step 2: Upload individual shards let mut completed_parts = Vec::new(); for part_number in 1..=total_parts { let start = (part_number - 1) * part_size; let end = std::cmp::min(start + part_size, total_size); let part_data = &test_data[start..end]; info!("πŸ” Upload encrypted shards {} ({} bytes)", part_number, part_data.len()); let upload_part_output = s3_client .upload_part() .bucket(TEST_BUCKET) .key(object_key) .upload_id(upload_id) .part_number(part_number as i32) .body(aws_sdk_s3::primitives::ByteStream::from(part_data.to_vec())) .send() .await?; let etag = upload_part_output.e_tag().unwrap().to_string(); completed_parts.push( aws_sdk_s3::types::CompletedPart::builder() .part_number(part_number as i32) .e_tag(&etag) .build(), ); debug!("Encrypted shard {} upload complete,ETag: {}", part_number, etag); } // Step 3: Complete the shard upload let completed_multipart_upload = aws_sdk_s3::types::CompletedMultipartUpload::builder() .set_parts(Some(completed_parts)) .build(); info!("πŸ”— Complete the encrypted shard upload"); let complete_output = s3_client .complete_multipart_upload() .bucket(TEST_BUCKET) .key(object_key) .upload_id(upload_id) .multipart_upload(completed_multipart_upload) .send() .await?; debug!("Encrypted multipart upload completed with ETag {:?}", complete_output.e_tag()); // Step 4: HEAD request to inspect metadata info!("πŸ“‹ Inspecting object metadata"); let head_response = s3_client.head_object().bucket(TEST_BUCKET).key(object_key).send().await?; debug!("HEAD response SSE: {:?}", head_response.server_side_encryption()); debug!("HEAD response metadata: {:?}", head_response.metadata()); // Step 5: GET request to download and verify info!("πŸ“₯ Downloading encrypted object for verification"); let get_response = s3_client.get_object().bucket(TEST_BUCKET).key(object_key).send().await?; debug!("GET response SSE: {:?}", get_response.server_side_encryption()); // 🎯 Critical check: GET response must include SSE-S3 headers assert_eq!( get_response.server_side_encryption(), Some(&aws_sdk_s3::types::ServerSideEncryption::Aes256) ); // Verify data integrity let downloaded_data = get_response.body.collect().await?.into_bytes(); assert_eq!(downloaded_data.len(), total_size); assert_eq!(&downloaded_data[..], &test_data[..]); kms_env.base_env.delete_test_bucket(TEST_BUCKET).await?; info!("βœ… Step 3 passed: multipart upload with SSE-S3 encryption"); Ok(()) } /// Step 4: test larger multipart uploads (streaming encryption) #[tokio::test] #[serial] async fn test_step4_large_multipart_upload_with_encryption() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { init_logging(); info!("πŸ§ͺ Step 4: test large-file multipart encryption"); let mut kms_env = LocalKMSTestEnvironment::new().await?; let _default_key_id = kms_env.start_rustfs_for_local_kms().await?; tokio::time::sleep(tokio::time::Duration::from_secs(3)).await; let s3_client = kms_env.base_env.create_s3_client(); kms_env.base_env.create_test_bucket(TEST_BUCKET).await?; let object_key = "test-large-multipart-encrypted"; let part_size = 6 * 1024 * 1024; // 6 MB per part (greater than the 1 MB encryption chunk) let total_parts = 3; // 18 MB total let total_size = part_size * total_parts; info!( "πŸ—‚οΈ Generated large-file test data: {} parts, {} MB each, {} MB total", total_parts, part_size / (1024 * 1024), total_size / (1024 * 1024) ); // Generate large test data (complex pattern for validation) let test_data: Vec<u8> = (0..total_size) .map(|i| { let part_num = i / part_size; let offset_in_part = i % part_size; ((part_num * 100 + offset_in_part / 1000) % 256) as u8 }) .collect(); info!("πŸ” Starting large-file multipart upload (SSE-S3 encryption)"); // Create multipart upload let create_multipart_output = s3_client .create_multipart_upload() .bucket(TEST_BUCKET) .key(object_key) .server_side_encryption(aws_sdk_s3::types::ServerSideEncryption::Aes256) .send() .await?; let upload_id = create_multipart_output.upload_id().unwrap(); info!("πŸ“‹ Created large encrypted multipart upload, ID: {}", upload_id); // Upload each part let mut completed_parts = Vec::new(); for part_number in 1..=total_parts { let start = (part_number - 1) * part_size; let end = std::cmp::min(start + part_size, total_size); let part_data = &test_data[start..end]; info!( "πŸ” Uploading encrypted large-file part {} ({:.2} MB)", part_number, part_data.len() as f64 / (1024.0 * 1024.0) ); let upload_part_output = s3_client .upload_part() .bucket(TEST_BUCKET) .key(object_key) .upload_id(upload_id) .part_number(part_number as i32) .body(aws_sdk_s3::primitives::ByteStream::from(part_data.to_vec())) .send() .await?; let etag = upload_part_output.e_tag().unwrap().to_string(); completed_parts.push( aws_sdk_s3::types::CompletedPart::builder() .part_number(part_number as i32) .e_tag(&etag) .build(), ); debug!("Large encrypted part {} uploaded with ETag {}", part_number, etag); } // Complete the multipart upload let completed_multipart_upload = aws_sdk_s3::types::CompletedMultipartUpload::builder() .set_parts(Some(completed_parts)) .build(); info!("πŸ”— Completing large encrypted multipart upload"); let complete_output = s3_client .complete_multipart_upload() .bucket(TEST_BUCKET) .key(object_key) .upload_id(upload_id) .multipart_upload(completed_multipart_upload) .send() .await?; debug!("Large encrypted multipart upload completed with ETag {:?}", complete_output.e_tag()); // Download and verify info!("πŸ“₯ Downloading large object for verification"); let get_response = s3_client.get_object().bucket(TEST_BUCKET).key(object_key).send().await?; // Verify encryption headers assert_eq!( get_response.server_side_encryption(), Some(&aws_sdk_s3::types::ServerSideEncryption::Aes256) ); // Verify data integrity let downloaded_data = get_response.body.collect().await?.into_bytes(); assert_eq!(downloaded_data.len(), total_size); // Validate bytes individually (stricter for large files) for (i, (&actual, &expected)) in downloaded_data.iter().zip(test_data.iter()).enumerate() { if actual != expected { panic!("Large file mismatch at byte {i}: actual={actual}, expected={expected}"); } } kms_env.base_env.delete_test_bucket(TEST_BUCKET).await?; info!("βœ… Step 4 passed: large-file multipart encryption succeeded"); Ok(()) } /// Step 5: test multipart uploads for every encryption mode #[tokio::test] #[serial] async fn test_step5_all_encryption_types_multipart() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { init_logging(); info!("πŸ§ͺ Step 5: test multipart uploads for every encryption mode"); let mut kms_env = LocalKMSTestEnvironment::new().await?; let _default_key_id = kms_env.start_rustfs_for_local_kms().await?; tokio::time::sleep(tokio::time::Duration::from_secs(3)).await; let s3_client = kms_env.base_env.create_s3_client(); kms_env.base_env.create_test_bucket(TEST_BUCKET).await?; let part_size = 5 * 1024 * 1024; // 5MB per part let total_parts = 2; let total_size = part_size * total_parts; // Test SSE-KMS info!("πŸ” Testing SSE-KMS multipart upload"); test_multipart_encryption_type( &s3_client, TEST_BUCKET, "test-multipart-sse-kms", total_size, part_size, total_parts, EncryptionType::SSEKMS, ) .await?; // Test SSE-C info!("πŸ” Testing SSE-C multipart upload"); test_multipart_encryption_type( &s3_client, TEST_BUCKET, "test-multipart-sse-c", total_size, part_size, total_parts, EncryptionType::SSEC, ) .await?; kms_env.base_env.delete_test_bucket(TEST_BUCKET).await?; info!("βœ… Step 5 passed: multipart uploads succeeded for every encryption mode"); Ok(()) } #[derive(Debug)] enum EncryptionType { SSEKMS, SSEC, } /// Helper: test multipart uploads for a specific encryption type async fn test_multipart_encryption_type( s3_client: &aws_sdk_s3::Client, bucket: &str, object_key: &str, total_size: usize, part_size: usize, total_parts: usize, encryption_type: EncryptionType, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { // Generate test data let test_data: Vec<u8> = (0..total_size).map(|i| ((i * 7) % 256) as u8).collect(); // Prepare SSE-C keys when required let (sse_c_key, sse_c_md5) = if matches!(encryption_type, EncryptionType::SSEC) { let key = "01234567890123456789012345678901"; let key_b64 = base64::Engine::encode(&base64::engine::general_purpose::STANDARD, key); let key_md5 = format!("{:x}", md5::compute(key)); (Some(key_b64), Some(key_md5)) } else { (None, None) }; info!("πŸ“‹ Creating multipart upload - {:?}", encryption_type); // Create multipart upload let mut create_request = s3_client.create_multipart_upload().bucket(bucket).key(object_key); create_request = match encryption_type { EncryptionType::SSEKMS => create_request.server_side_encryption(aws_sdk_s3::types::ServerSideEncryption::AwsKms), EncryptionType::SSEC => create_request .sse_customer_algorithm("AES256") .sse_customer_key(sse_c_key.as_ref().unwrap()) .sse_customer_key_md5(sse_c_md5.as_ref().unwrap()), }; let create_multipart_output = create_request.send().await?; let upload_id = create_multipart_output.upload_id().unwrap(); // Upload parts let mut completed_parts = Vec::new(); for part_number in 1..=total_parts { let start = (part_number - 1) * part_size; let end = std::cmp::min(start + part_size, total_size); let part_data = &test_data[start..end]; let mut upload_request = s3_client .upload_part() .bucket(bucket) .key(object_key) .upload_id(upload_id) .part_number(part_number as i32) .body(aws_sdk_s3::primitives::ByteStream::from(part_data.to_vec())); // SSE-C requires the key on each UploadPart request if matches!(encryption_type, EncryptionType::SSEC) { upload_request = upload_request .sse_customer_algorithm("AES256") .sse_customer_key(sse_c_key.as_ref().unwrap()) .sse_customer_key_md5(sse_c_md5.as_ref().unwrap()); } let upload_part_output = upload_request.send().await?; let etag = upload_part_output.e_tag().unwrap().to_string(); completed_parts.push( aws_sdk_s3::types::CompletedPart::builder() .part_number(part_number as i32) .e_tag(&etag) .build(), ); debug!("{:?} part {} uploaded", encryption_type, part_number); } // Complete the multipart upload let completed_multipart_upload = aws_sdk_s3::types::CompletedMultipartUpload::builder() .set_parts(Some(completed_parts)) .build(); let _complete_output = s3_client .complete_multipart_upload() .bucket(bucket) .key(object_key) .upload_id(upload_id) .multipart_upload(completed_multipart_upload) .send() .await?; // Download and verify let mut get_request = s3_client.get_object().bucket(bucket).key(object_key); // SSE-C requires the key on GET requests if matches!(encryption_type, EncryptionType::SSEC) { get_request = get_request .sse_customer_algorithm("AES256") .sse_customer_key(sse_c_key.as_ref().unwrap()) .sse_customer_key_md5(sse_c_md5.as_ref().unwrap()); } let get_response = get_request.send().await?; // Verify encryption headers match encryption_type { EncryptionType::SSEKMS => { assert_eq!( get_response.server_side_encryption(), Some(&aws_sdk_s3::types::ServerSideEncryption::AwsKms) ); } EncryptionType::SSEC => { assert_eq!(get_response.sse_customer_algorithm(), Some("AES256")); } } // Verify data integrity let downloaded_data = get_response.body.collect().await?.into_bytes(); assert_eq!(downloaded_data.len(), total_size); assert_eq!(&downloaded_data[..], &test_data[..]); info!("βœ… {:?} multipart upload test passed", encryption_type); Ok(()) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/kms/kms_vault_test.rs
crates/e2e_test/src/kms/kms_vault_test.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! End-to-end tests for Vault KMS backend //! //! These tests mirror the local KMS coverage but target the Vault backend. //! They validate Vault bootstrap, admin API flows, encryption modes, and //! multipart upload behaviour. use crate::common::{TEST_BUCKET, init_logging}; use md5::compute; use serial_test::serial; use tokio::time::{Duration, sleep}; use tracing::{error, info}; use super::common::{ VAULT_KEY_NAME, VaultTestEnvironment, get_kms_status, start_kms, test_all_multipart_encryption_types, test_error_scenarios, test_kms_key_management, test_sse_c_encryption, test_sse_kms_encryption, test_sse_s3_encryption, }; /// Helper that brings up Vault, configures RustFS, and starts the KMS service. struct VaultKmsTestContext { env: VaultTestEnvironment, } impl VaultKmsTestContext { async fn new() -> Result<Self, Box<dyn std::error::Error + Send + Sync>> { let mut env = VaultTestEnvironment::new().await?; env.start_vault().await?; env.setup_vault_transit().await?; env.start_rustfs_for_vault().await?; env.configure_vault_kms().await?; start_kms(&env.base_env.url, &env.base_env.access_key, &env.base_env.secret_key).await?; // Allow Vault to finish initialising token auth and transit engine. sleep(Duration::from_secs(2)).await; Ok(Self { env }) } fn base_env(&self) -> &crate::common::RustFSTestEnvironment { &self.env.base_env } fn s3_client(&self) -> aws_sdk_s3::Client { self.env.base_env.create_s3_client() } } #[tokio::test] #[serial] async fn test_vault_kms_end_to_end() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { init_logging(); info!("Starting Vault KMS End-to-End Test with default key {}", VAULT_KEY_NAME); let context = VaultKmsTestContext::new().await?; match get_kms_status(&context.base_env().url, &context.base_env().access_key, &context.base_env().secret_key).await { Ok(status) => info!("Vault KMS status after startup: {}", status), Err(err) => { error!("Failed to query Vault KMS status: {}", err); return Err(err); } } let s3_client = context.s3_client(); context .base_env() .create_test_bucket(TEST_BUCKET) .await .expect("Failed to create test bucket"); test_kms_key_management(&context.base_env().url, &context.base_env().access_key, &context.base_env().secret_key) .await .expect("Vault KMS key management test failed"); test_sse_c_encryption(&s3_client, TEST_BUCKET) .await .expect("Vault SSE-C encryption test failed"); test_sse_s3_encryption(&s3_client, TEST_BUCKET) .await .expect("Vault SSE-S3 encryption test failed"); test_sse_kms_encryption(&s3_client, TEST_BUCKET) .await .expect("Vault SSE-KMS encryption test failed"); test_error_scenarios(&s3_client, TEST_BUCKET) .await .expect("Vault KMS error scenario test failed"); context .base_env() .delete_test_bucket(TEST_BUCKET) .await .expect("Failed to delete test bucket"); info!("Vault KMS End-to-End Test completed successfully"); Ok(()) } #[tokio::test] #[serial] async fn test_vault_kms_key_isolation() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { init_logging(); info!("Starting Vault KMS SSE-C key isolation test"); let context = VaultKmsTestContext::new().await?; let s3_client = context.s3_client(); context .base_env() .create_test_bucket(TEST_BUCKET) .await .expect("Failed to create test bucket"); let key1 = "01234567890123456789012345678901"; let key2 = "98765432109876543210987654321098"; let key1_b64 = base64::Engine::encode(&base64::engine::general_purpose::STANDARD, key1); let key2_b64 = base64::Engine::encode(&base64::engine::general_purpose::STANDARD, key2); let key1_md5 = format!("{:x}", compute(key1)); let key2_md5 = format!("{:x}", compute(key2)); let data1 = b"Vault data encrypted with key 1"; let data2 = b"Vault data encrypted with key 2"; s3_client .put_object() .bucket(TEST_BUCKET) .key("vault-object1") .body(aws_sdk_s3::primitives::ByteStream::from(data1.to_vec())) .sse_customer_algorithm("AES256") .sse_customer_key(&key1_b64) .sse_customer_key_md5(&key1_md5) .send() .await .expect("Failed to upload object1 with key1"); s3_client .put_object() .bucket(TEST_BUCKET) .key("vault-object2") .body(aws_sdk_s3::primitives::ByteStream::from(data2.to_vec())) .sse_customer_algorithm("AES256") .sse_customer_key(&key2_b64) .sse_customer_key_md5(&key2_md5) .send() .await .expect("Failed to upload object2 with key2"); let object1 = s3_client .get_object() .bucket(TEST_BUCKET) .key("vault-object1") .sse_customer_algorithm("AES256") .sse_customer_key(&key1_b64) .sse_customer_key_md5(&key1_md5) .send() .await .expect("Failed to download object1 with key1"); let downloaded1 = object1.body.collect().await.expect("Failed to read object1").into_bytes(); assert_eq!(downloaded1.as_ref(), data1); let wrong_key = s3_client .get_object() .bucket(TEST_BUCKET) .key("vault-object1") .sse_customer_algorithm("AES256") .sse_customer_key(&key2_b64) .sse_customer_key_md5(&key2_md5) .send() .await; assert!(wrong_key.is_err(), "Object1 should not decrypt with key2"); context .base_env() .delete_test_bucket(TEST_BUCKET) .await .expect("Failed to delete test bucket"); info!("Vault KMS SSE-C key isolation test completed successfully"); Ok(()) } #[tokio::test] #[serial] async fn test_vault_kms_large_file() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { init_logging(); info!("Starting Vault KMS large file SSE-S3 test"); let context = VaultKmsTestContext::new().await?; let s3_client = context.s3_client(); context .base_env() .create_test_bucket(TEST_BUCKET) .await .expect("Failed to create test bucket"); let large_data = vec![0xCDu8; 1024 * 1024]; let object_key = "vault-large-encrypted-file"; let put_response = s3_client .put_object() .bucket(TEST_BUCKET) .key(object_key) .body(aws_sdk_s3::primitives::ByteStream::from(large_data.clone())) .server_side_encryption(aws_sdk_s3::types::ServerSideEncryption::Aes256) .send() .await .expect("Failed to upload large SSE-S3 object"); assert_eq!( put_response.server_side_encryption(), Some(&aws_sdk_s3::types::ServerSideEncryption::Aes256) ); let get_response = s3_client .get_object() .bucket(TEST_BUCKET) .key(object_key) .send() .await .expect("Failed to download large SSE-S3 object"); assert_eq!( get_response.server_side_encryption(), Some(&aws_sdk_s3::types::ServerSideEncryption::Aes256) ); let downloaded = get_response .body .collect() .await .expect("Failed to read large object body") .into_bytes(); assert_eq!(downloaded.len(), large_data.len()); assert_eq!(downloaded.as_ref(), large_data.as_slice()); context .base_env() .delete_test_bucket(TEST_BUCKET) .await .expect("Failed to delete test bucket"); info!("Vault KMS large file test completed successfully"); Ok(()) } #[tokio::test] #[serial] async fn test_vault_kms_multipart_upload() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { init_logging(); info!("Starting Vault KMS multipart upload encryption suite"); let context = VaultKmsTestContext::new().await?; let s3_client = context.s3_client(); context .base_env() .create_test_bucket(TEST_BUCKET) .await .expect("Failed to create test bucket"); test_all_multipart_encryption_types(&s3_client, TEST_BUCKET, "vault-multipart") .await .expect("Vault multipart encryption test suite failed"); context .base_env() .delete_test_bucket(TEST_BUCKET) .await .expect("Failed to delete test bucket"); info!("Vault KMS multipart upload tests completed successfully"); Ok(()) } #[tokio::test] #[serial] async fn test_vault_kms_key_operations() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { init_logging(); info!("Starting Vault KMS key operations test (CRUD)"); let context = VaultKmsTestContext::new().await?; test_vault_kms_key_crud(&context.base_env().url, &context.base_env().access_key, &context.base_env().secret_key).await?; info!("Vault KMS key operations test completed successfully"); Ok(()) } async fn test_vault_kms_key_crud( base_url: &str, access_key: &str, secret_key: &str, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { info!("Testing Vault KMS key CRUD operations"); // Create with key name in tags let test_key_name = "test-vault-key-crud"; let create_key_body = serde_json::json!({ "key_usage": "EncryptDecrypt", "description": "Test key for CRUD operations", "tags": { "name": test_key_name, "algorithm": "AES-256", "created_by": "e2e_test", "test_type": "crud" } }) .to_string(); let create_response = crate::common::awscurl_post(&format!("{base_url}/rustfs/admin/v3/kms/keys"), &create_key_body, access_key, secret_key) .await?; let create_result: serde_json::Value = serde_json::from_str(&create_response)?; let key_id = create_result["key_id"] .as_str() .ok_or("Failed to get key_id from create response")?; info!("βœ… Create: Created key with ID: {}", key_id); // Read let describe_response = crate::common::awscurl_get(&format!("{base_url}/rustfs/admin/v3/kms/keys/{key_id}"), access_key, secret_key).await?; let describe_result: serde_json::Value = serde_json::from_str(&describe_response)?; assert_eq!(describe_result["key_metadata"]["key_id"], key_id); assert_eq!(describe_result["key_metadata"]["key_usage"], "EncryptDecrypt"); assert_eq!(describe_result["key_metadata"]["key_state"], "Enabled"); // Verify that the key name was properly stored - MUST be present let tags = describe_result["key_metadata"]["tags"] .as_object() .expect("Tags field must be present in key metadata"); let stored_name = tags .get("name") .and_then(|v| v.as_str()) .expect("Key name must be preserved in tags"); assert_eq!(stored_name, test_key_name, "Key name must match the name provided during creation"); // Verify other tags are also preserved assert_eq!( tags.get("algorithm") .and_then(|v| v.as_str()) .expect("Algorithm tag must be present"), "AES-256" ); assert_eq!( tags.get("created_by") .and_then(|v| v.as_str()) .expect("Created_by tag must be present"), "e2e_test" ); assert_eq!( tags.get("test_type") .and_then(|v| v.as_str()) .expect("Test_type tag must be present"), "crud" ); info!("βœ… Read: Successfully described key: {}", key_id); // Read let list_response = crate::common::awscurl_get(&format!("{base_url}/rustfs/admin/v3/kms/keys"), access_key, secret_key).await?; let list_result: serde_json::Value = serde_json::from_str(&list_response)?; let keys = list_result["keys"] .as_array() .ok_or("Failed to get keys array from list response")?; let found_key = keys.iter().find(|k| k["key_id"].as_str() == Some(key_id)); assert!(found_key.is_some(), "Created key not found in list"); // Verify key name in list response - MUST be present let key = found_key.expect("Created key must be found in list"); let list_tags = key["tags"].as_object().expect("Tags field must be present in list response"); let listed_name = list_tags .get("name") .and_then(|v| v.as_str()) .expect("Key name must be preserved in list response"); assert_eq!( listed_name, test_key_name, "Key name in list must match the name provided during creation" ); info!("βœ… Read: Successfully listed keys, found test key"); // Delete let delete_response = crate::common::execute_awscurl( &format!("{base_url}/rustfs/admin/v3/kms/keys/delete?keyId={key_id}"), "DELETE", None, access_key, secret_key, ) .await?; // Parse and validate the delete response let delete_result: serde_json::Value = serde_json::from_str(&delete_response)?; assert_eq!(delete_result["success"], true, "Delete operation must return success=true"); info!("βœ… Delete: Successfully deleted key: {}", key_id); // Verify key state after deletion let describe_deleted_response = crate::common::awscurl_get(&format!("{base_url}/rustfs/admin/v3/kms/keys/{key_id}"), access_key, secret_key).await?; let describe_result: serde_json::Value = serde_json::from_str(&describe_deleted_response)?; let key_state = describe_result["key_metadata"]["key_state"] .as_str() .expect("Key state must be present after deletion"); // After deletion, key must not be in Enabled state assert_ne!(key_state, "Enabled", "Deleted key must not remain in Enabled state"); // Key should be in PendingDeletion state after deletion assert_eq!(key_state, "PendingDeletion", "Deleted key must be in PendingDeletion state"); info!("βœ… Delete verification: Key state correctly changed to: {}", key_state); // Force Delete - Force immediate deletion for PendingDeletion key let force_delete_response = crate::common::execute_awscurl( &format!("{base_url}/rustfs/admin/v3/kms/keys/delete?keyId={key_id}&force_immediate=true"), "DELETE", None, access_key, secret_key, ) .await?; // Parse and validate the force delete response let force_delete_result: serde_json::Value = serde_json::from_str(&force_delete_response)?; assert_eq!(force_delete_result["success"], true, "Force delete operation must return success=true"); info!("βœ… Force Delete: Successfully force deleted key: {}", key_id); // Verify key no longer exists after force deletion (should return error) let describe_force_deleted_result = crate::common::awscurl_get(&format!("{base_url}/rustfs/admin/v3/kms/keys/{key_id}"), access_key, secret_key).await; // After force deletion, key should not be found (GET should fail) assert!(describe_force_deleted_result.is_err(), "Force deleted key should not be found"); info!("βœ… Force Delete verification: Key was permanently deleted and is no longer accessible"); info!("Vault KMS key CRUD operations completed successfully"); Ok(()) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/kms/common.rs
crates/e2e_test/src/kms/common.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #![allow(dead_code)] #![allow(clippy::upper_case_acronyms)] //! KMS-specific utilities for end-to-end tests //! //! This module provides KMS-specific functionality including: //! - Vault server management and configuration //! - KMS backend configuration (Local and Vault) //! - SSE encryption testing utilities use crate::common::{RustFSTestEnvironment, awscurl_get, awscurl_post, init_logging as common_init_logging}; use aws_sdk_s3::Client; use aws_sdk_s3::primitives::ByteStream; use aws_sdk_s3::types::ServerSideEncryption; use base64::Engine; use serde_json; use std::process::{Child, Command}; use std::time::Duration; use tokio::fs; use tokio::net::TcpStream; use tokio::time::sleep; use tracing::{debug, error, info}; // KMS-specific constants pub const TEST_BUCKET: &str = "kms-test-bucket"; // Vault constants pub const VAULT_URL: &str = "http://127.0.0.1:8200"; pub const VAULT_ADDRESS: &str = "127.0.0.1:8200"; pub const VAULT_TOKEN: &str = "dev-root-token"; pub const VAULT_TRANSIT_PATH: &str = "transit"; pub const VAULT_KEY_NAME: &str = "rustfs-master-key"; /// Initialize tracing for KMS tests with KMS-specific log levels pub fn init_logging() { common_init_logging(); // Additional KMS-specific logging configuration can be added here if needed } // KMS-specific helper functions /// Configure KMS backend via admin API pub async fn configure_kms( base_url: &str, config_json: &str, access_key: &str, secret_key: &str, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { let url = format!("{base_url}/rustfs/admin/v3/kms/configure"); awscurl_post(&url, config_json, access_key, secret_key).await?; info!("KMS configured successfully"); Ok(()) } /// Start KMS service via admin API pub async fn start_kms( base_url: &str, access_key: &str, secret_key: &str, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { let url = format!("{base_url}/rustfs/admin/v3/kms/start"); awscurl_post(&url, "{}", access_key, secret_key).await?; info!("KMS started successfully"); Ok(()) } /// Get KMS status via admin API pub async fn get_kms_status( base_url: &str, access_key: &str, secret_key: &str, ) -> Result<String, Box<dyn std::error::Error + Send + Sync>> { let url = format!("{base_url}/rustfs/admin/v3/kms/status"); let status = awscurl_get(&url, access_key, secret_key).await?; info!("KMS status retrieved: {}", status); Ok(status) } /// Create a default KMS key for testing and return the created key ID pub async fn create_default_key( base_url: &str, access_key: &str, secret_key: &str, ) -> Result<String, Box<dyn std::error::Error + Send + Sync>> { let create_key_body = serde_json::json!({ "KeyUsage": "ENCRYPT_DECRYPT", "Description": "Default key for e2e testing" }) .to_string(); let url = format!("{base_url}/rustfs/admin/v3/kms/keys"); let response = awscurl_post(&url, &create_key_body, access_key, secret_key).await?; // Parse response to get the actual key ID let create_result: serde_json::Value = serde_json::from_str(&response)?; let key_id = create_result["key_id"] .as_str() .ok_or("Failed to get key_id from create response")? .to_string(); info!("Default KMS key created: {}", key_id); Ok(key_id) } /// Create a KMS key with a specific ID (by directly writing to the key directory) pub async fn create_key_with_specific_id(key_dir: &str, key_id: &str) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { use rand::RngCore; use std::collections::HashMap; use tokio::fs; // Create a 32-byte AES key let mut key_data = [0u8; 32]; rand::rng().fill_bytes(&mut key_data); // Create the stored key structure that Local KMS backend expects let stored_key = serde_json::json!({ "key_id": key_id, "version": 1u32, "algorithm": "AES_256", "usage": "EncryptDecrypt", "status": "Active", "metadata": HashMap::<String, String>::new(), "created_at": chrono::Utc::now().to_rfc3339(), "rotated_at": serde_json::Value::Null, "created_by": "e2e-test", "encrypted_key_material": key_data.to_vec(), "nonce": Vec::<u8>::new() }); // Write the key to file with the specified ID as JSON let key_path = format!("{key_dir}/{key_id}.key"); let content = serde_json::to_vec_pretty(&stored_key)?; fs::write(&key_path, &content).await?; info!("Created KMS key with ID '{}' at path: {}", key_id, key_path); Ok(()) } /// Test SSE-C encryption with the given S3 client pub async fn test_sse_c_encryption(s3_client: &Client, bucket: &str) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { info!("Testing SSE-C encryption"); let test_key = "01234567890123456789012345678901"; // 32-byte key let test_key_b64 = base64::engine::general_purpose::STANDARD.encode(test_key); let test_key_md5 = format!("{:x}", md5::compute(test_key)); let test_data = b"Hello, KMS SSE-C World!"; let object_key = "test-sse-c-object"; // Upload with SSE-C (customer-provided key encryption) // Note: For SSE-C, we should NOT set server_side_encryption, only the customer key headers let put_response = s3_client .put_object() .bucket(bucket) .key(object_key) .body(ByteStream::from(test_data.to_vec())) .sse_customer_algorithm("AES256") .sse_customer_key(&test_key_b64) .sse_customer_key_md5(&test_key_md5) .send() .await?; info!("SSE-C upload successful, ETag: {:?}", put_response.e_tag()); // For SSE-C, server_side_encryption should be None since customer provides the key // The encryption algorithm is specified via SSE-C headers instead // Download with SSE-C info!("Starting SSE-C download test"); let get_response = s3_client .get_object() .bucket(bucket) .key(object_key) .sse_customer_algorithm("AES256") .sse_customer_key(&test_key_b64) .sse_customer_key_md5(&test_key_md5) .send() .await?; info!("SSE-C download successful"); info!("Starting to collect response body"); let downloaded_data = get_response.body.collect().await?.into_bytes(); info!("Downloaded data length: {}, expected length: {}", downloaded_data.len(), test_data.len()); assert_eq!(downloaded_data.as_ref(), test_data); // For SSE-C, we don't check server_side_encryption since it's customer-managed info!("SSE-C encryption test completed successfully"); Ok(()) } /// Test SSE-S3 encryption (server-managed keys) pub async fn test_sse_s3_encryption(s3_client: &Client, bucket: &str) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { info!("Testing SSE-S3 encryption"); let test_data = b"Hello, KMS SSE-S3 World!"; let object_key = "test-sse-s3-object"; // Upload with SSE-S3 let put_response = s3_client .put_object() .bucket(bucket) .key(object_key) .body(ByteStream::from(test_data.to_vec())) .server_side_encryption(ServerSideEncryption::Aes256) .send() .await?; info!("SSE-S3 upload successful, ETag: {:?}", put_response.e_tag()); assert_eq!(put_response.server_side_encryption(), Some(&ServerSideEncryption::Aes256)); // Download object let get_response = s3_client.get_object().bucket(bucket).key(object_key).send().await?; let encryption = get_response.server_side_encryption().cloned(); let downloaded_data = get_response.body.collect().await?.into_bytes(); assert_eq!(downloaded_data.as_ref(), test_data); assert_eq!(encryption, Some(ServerSideEncryption::Aes256)); info!("SSE-S3 encryption test completed successfully"); Ok(()) } /// Test SSE-KMS encryption (KMS-managed keys) pub async fn test_sse_kms_encryption( s3_client: &aws_sdk_s3::Client, bucket: &str, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { info!("Testing SSE-KMS encryption"); let object_key = "test-sse-kms-object"; let test_data = b"Hello, SSE-KMS World! This data should be encrypted with KMS-managed keys."; // Upload object with SSE-KMS encryption let put_response = s3_client .put_object() .bucket(bucket) .key(object_key) .body(aws_sdk_s3::primitives::ByteStream::from(test_data.to_vec())) .server_side_encryption(ServerSideEncryption::AwsKms) .send() .await?; info!("SSE-KMS upload successful, ETag: {:?}", put_response.e_tag()); assert_eq!(put_response.server_side_encryption(), Some(&ServerSideEncryption::AwsKms)); // Download object let get_response = s3_client.get_object().bucket(bucket).key(object_key).send().await?; let encryption = get_response.server_side_encryption().cloned(); let downloaded_data = get_response.body.collect().await?.into_bytes(); assert_eq!(downloaded_data.as_ref(), test_data); assert_eq!(encryption, Some(ServerSideEncryption::AwsKms)); info!("SSE-KMS encryption test completed successfully"); Ok(()) } /// Test KMS key management APIs pub async fn test_kms_key_management( base_url: &str, access_key: &str, secret_key: &str, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { info!("Testing KMS key management APIs"); // Test CreateKey let create_key_body = serde_json::json!({ "KeyUsage": "EncryptDecrypt", "Description": "Test key for e2e testing" }) .to_string(); let create_response = awscurl_post(&format!("{base_url}/rustfs/admin/v3/kms/keys"), &create_key_body, access_key, secret_key).await?; let create_result: serde_json::Value = serde_json::from_str(&create_response)?; let key_id = create_result["key_id"] .as_str() .ok_or("Failed to get key_id from create response")?; info!("Created key with ID: {}", key_id); // Test DescribeKey let describe_response = awscurl_get(&format!("{base_url}/rustfs/admin/v3/kms/keys/{key_id}"), access_key, secret_key).await?; info!("DescribeKey response: {}", describe_response); let describe_result: serde_json::Value = serde_json::from_str(&describe_response)?; info!("Parsed describe result: {:?}", describe_result); assert_eq!(describe_result["key_metadata"]["key_id"], key_id); info!("Successfully described key: {}", key_id); // Test ListKeys let list_response = awscurl_get(&format!("{base_url}/rustfs/admin/v3/kms/keys"), access_key, secret_key).await?; let list_result: serde_json::Value = serde_json::from_str(&list_response)?; let keys = list_result["keys"] .as_array() .ok_or("Failed to get keys array from list response")?; let found_key = keys.iter().any(|k| k["key_id"].as_str() == Some(key_id)); assert!(found_key, "Created key not found in list"); info!("Successfully listed keys, found created key"); info!("KMS key management API tests completed successfully"); Ok(()) } /// Test error scenarios pub async fn test_error_scenarios(s3_client: &Client, bucket: &str) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { info!("Testing error scenarios"); // Test SSE-C with wrong key for download let test_key = "01234567890123456789012345678901"; let wrong_key = "98765432109876543210987654321098"; let test_key_b64 = base64::engine::general_purpose::STANDARD.encode(test_key); let wrong_key_b64 = base64::engine::general_purpose::STANDARD.encode(wrong_key); let test_key_md5 = format!("{:x}", md5::compute(test_key)); let wrong_key_md5 = format!("{:x}", md5::compute(wrong_key)); let test_data = b"Test data for error scenarios"; let object_key = "test-error-object"; // Upload with correct key (SSE-C) s3_client .put_object() .bucket(bucket) .key(object_key) .body(ByteStream::from(test_data.to_vec())) .sse_customer_algorithm("AES256") .sse_customer_key(&test_key_b64) .sse_customer_key_md5(&test_key_md5) .send() .await?; // Try to download with wrong key - should fail let wrong_key_result = s3_client .get_object() .bucket(bucket) .key(object_key) .sse_customer_algorithm("AES256") .sse_customer_key(&wrong_key_b64) .sse_customer_key_md5(&wrong_key_md5) .send() .await; assert!(wrong_key_result.is_err(), "Download with wrong SSE-C key should fail"); info!("βœ… Correctly rejected download with wrong SSE-C key"); info!("Error scenario tests completed successfully"); Ok(()) } /// Vault test environment management pub struct VaultTestEnvironment { pub base_env: RustFSTestEnvironment, pub vault_process: Option<Child>, } impl VaultTestEnvironment { /// Create a new Vault test environment pub async fn new() -> Result<Self, Box<dyn std::error::Error + Send + Sync>> { let base_env = RustFSTestEnvironment::new().await?; Ok(Self { base_env, vault_process: None, }) } /// Start Vault server in development mode pub async fn start_vault(&mut self) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { info!("Starting Vault server in development mode"); let vault_process = Command::new("vault") .args([ "server", "-dev", "-dev-root-token-id", VAULT_TOKEN, "-dev-listen-address", VAULT_ADDRESS, ]) .spawn()?; self.vault_process = Some(vault_process); // Wait for Vault to start self.wait_for_vault_ready().await?; Ok(()) } async fn wait_for_vault_ready(&self) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { info!("Waiting for Vault server to be ready..."); for i in 0..30 { let port_check = TcpStream::connect(VAULT_ADDRESS).await.is_ok(); if port_check { // Additional check by making a health request if let Ok(response) = reqwest::get(&format!("{VAULT_URL}/v1/sys/health")).await && response.status().is_success() { info!("Vault server is ready after {} seconds", i); return Ok(()); } } if i == 29 { return Err("Vault server failed to become ready".into()); } sleep(Duration::from_secs(1)).await; } Ok(()) } /// Setup Vault transit secrets engine pub async fn setup_vault_transit(&self) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { let client = reqwest::Client::new(); info!("Enabling Vault transit secrets engine"); // Enable transit secrets engine let enable_response = client .post(format!("{VAULT_URL}/v1/sys/mounts/{VAULT_TRANSIT_PATH}")) .header("X-Vault-Token", VAULT_TOKEN) .json(&serde_json::json!({ "type": "transit" })) .send() .await?; if !enable_response.status().is_success() && enable_response.status() != 400 { let error_text = enable_response.text().await?; return Err(format!("Failed to enable transit engine: {error_text}").into()); } info!("Creating Vault encryption key"); // Create encryption key let key_response = client .post(format!("{VAULT_URL}/v1/{VAULT_TRANSIT_PATH}/keys/{VAULT_KEY_NAME}")) .header("X-Vault-Token", VAULT_TOKEN) .json(&serde_json::json!({ "type": "aes256-gcm96" })) .send() .await?; if !key_response.status().is_success() && key_response.status() != 400 { let error_text = key_response.text().await?; return Err(format!("Failed to create encryption key: {error_text}").into()); } info!("Vault transit engine setup completed"); Ok(()) } /// Start RustFS server for Vault backend; dynamic configuration will be applied later. pub async fn start_rustfs_for_vault(&mut self) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { self.base_env.start_rustfs_server(Vec::new()).await } /// Configure Vault KMS backend pub async fn configure_vault_kms(&self) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { let kms_config = serde_json::json!({ "backend_type": "vault", "address": VAULT_URL, "auth_method": { "Token": { "token": VAULT_TOKEN } }, "mount_path": VAULT_TRANSIT_PATH, "kv_mount": "secret", "key_path_prefix": "rustfs/kms/keys", "default_key_id": VAULT_KEY_NAME, "skip_tls_verify": true }) .to_string(); configure_kms(&self.base_env.url, &kms_config, &self.base_env.access_key, &self.base_env.secret_key).await } } impl Drop for VaultTestEnvironment { fn drop(&mut self) { if let Some(mut process) = self.vault_process.take() { info!("Terminating Vault process"); if let Err(e) = process.kill() { error!("Failed to kill Vault process: {}", e); } else { let _ = process.wait(); } } } } /// Encryption types for multipart upload testing #[derive(Debug, Clone)] pub enum EncryptionType { None, SSES3, SSEKMS, SSEC { key: String, key_md5: String }, } /// Configuration for multipart upload tests #[derive(Debug, Clone)] pub struct MultipartTestConfig { pub object_key: String, pub part_size: usize, pub total_parts: usize, pub encryption_type: EncryptionType, } impl MultipartTestConfig { pub fn new(object_key: impl Into<String>, part_size: usize, total_parts: usize, encryption_type: EncryptionType) -> Self { Self { object_key: object_key.into(), part_size, total_parts, encryption_type, } } pub fn total_size(&self) -> usize { self.part_size * self.total_parts } } /// Perform a comprehensive multipart upload test with the specified configuration pub async fn test_multipart_upload_with_config( s3_client: &Client, bucket: &str, config: &MultipartTestConfig, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { let total_size = config.total_size(); info!("πŸ§ͺ Starting multipart upload test - {:?}", config.encryption_type); info!( " Object: {}, parts: {}, part size: {} MB, total: {} MB", config.object_key, config.total_parts, config.part_size / (1024 * 1024), total_size / (1024 * 1024) ); // Generate test data with patterns for verification let test_data: Vec<u8> = (0..total_size) .map(|i| { let part_num = i / config.part_size; let offset_in_part = i % config.part_size; ((part_num * 100 + offset_in_part / 1000) % 256) as u8 }) .collect(); // Prepare encryption parameters let (sse_c_key_b64, sse_c_key_md5) = match &config.encryption_type { EncryptionType::SSEC { key, key_md5 } => { let key_b64 = base64::engine::general_purpose::STANDARD.encode(key); (Some(key_b64), Some(key_md5.clone())) } _ => (None, None), }; // Step 1: Create multipart upload let mut create_request = s3_client.create_multipart_upload().bucket(bucket).key(&config.object_key); create_request = match &config.encryption_type { EncryptionType::None => create_request, EncryptionType::SSES3 => create_request.server_side_encryption(ServerSideEncryption::Aes256), EncryptionType::SSEKMS => create_request.server_side_encryption(ServerSideEncryption::AwsKms), EncryptionType::SSEC { .. } => create_request .sse_customer_algorithm("AES256") .sse_customer_key(sse_c_key_b64.as_ref().unwrap()) .sse_customer_key_md5(sse_c_key_md5.as_ref().unwrap()), }; let create_multipart_output = create_request.send().await?; let upload_id = create_multipart_output.upload_id().unwrap(); info!("πŸ“‹ Created multipart upload, ID: {}", upload_id); // Step 2: Upload parts let mut completed_parts = Vec::new(); for part_number in 1..=config.total_parts { let start = (part_number - 1) * config.part_size; let end = std::cmp::min(start + config.part_size, total_size); let part_data = &test_data[start..end]; info!("πŸ“€ Uploading part {} ({:.2} MB)", part_number, part_data.len() as f64 / (1024.0 * 1024.0)); let mut upload_request = s3_client .upload_part() .bucket(bucket) .key(&config.object_key) .upload_id(upload_id) .part_number(part_number as i32) .body(ByteStream::from(part_data.to_vec())); // Add encryption headers for SSE-C parts if let EncryptionType::SSEC { .. } = &config.encryption_type { upload_request = upload_request .sse_customer_algorithm("AES256") .sse_customer_key(sse_c_key_b64.as_ref().unwrap()) .sse_customer_key_md5(sse_c_key_md5.as_ref().unwrap()); } let upload_part_output = upload_request.send().await?; let etag = upload_part_output.e_tag().unwrap().to_string(); completed_parts.push( aws_sdk_s3::types::CompletedPart::builder() .part_number(part_number as i32) .e_tag(&etag) .build(), ); debug!("Part {} uploaded with ETag {}", part_number, etag); } // Step 3: Complete multipart upload let completed_multipart_upload = aws_sdk_s3::types::CompletedMultipartUpload::builder() .set_parts(Some(completed_parts)) .build(); info!("πŸ”— Completing multipart upload"); let complete_output = s3_client .complete_multipart_upload() .bucket(bucket) .key(&config.object_key) .upload_id(upload_id) .multipart_upload(completed_multipart_upload) .send() .await?; debug!("Multipart upload finalized with ETag {:?}", complete_output.e_tag()); // Step 4: Download and verify info!("πŸ“₯ Downloading object for verification"); let mut get_request = s3_client.get_object().bucket(bucket).key(&config.object_key); // Add encryption headers for SSE-C GET if let EncryptionType::SSEC { .. } = &config.encryption_type { get_request = get_request .sse_customer_algorithm("AES256") .sse_customer_key(sse_c_key_b64.as_ref().unwrap()) .sse_customer_key_md5(sse_c_key_md5.as_ref().unwrap()); } let get_response = get_request.send().await?; // Verify encryption headers match &config.encryption_type { EncryptionType::None => { assert_eq!(get_response.server_side_encryption(), None); } EncryptionType::SSES3 => { assert_eq!(get_response.server_side_encryption(), Some(&ServerSideEncryption::Aes256)); } EncryptionType::SSEKMS => { assert_eq!(get_response.server_side_encryption(), Some(&ServerSideEncryption::AwsKms)); } EncryptionType::SSEC { .. } => { assert_eq!(get_response.sse_customer_algorithm(), Some("AES256")); } } // Verify data integrity let downloaded_data = get_response.body.collect().await?.into_bytes(); assert_eq!(downloaded_data.len(), total_size); assert_eq!(&downloaded_data[..], &test_data[..]); info!("βœ… Multipart upload test passed - {:?}", config.encryption_type); Ok(()) } /// Create a standard SSE-C encryption configuration for testing pub fn create_sse_c_config() -> EncryptionType { let key = "01234567890123456789012345678901"; // 32-byte key let key_md5 = format!("{:x}", md5::compute(key)); EncryptionType::SSEC { key: key.to_string(), key_md5, } } /// Test all encryption types for multipart uploads pub async fn test_all_multipart_encryption_types( s3_client: &Client, bucket: &str, base_object_key: &str, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { info!("πŸ§ͺ Testing multipart uploads for every encryption type"); let part_size = 5 * 1024 * 1024; // 5MB per part let total_parts = 2; // Test configurations for all encryption types let test_configs = vec![ MultipartTestConfig::new(format!("{base_object_key}-no-encryption"), part_size, total_parts, EncryptionType::None), MultipartTestConfig::new(format!("{base_object_key}-sse-s3"), part_size, total_parts, EncryptionType::SSES3), MultipartTestConfig::new(format!("{base_object_key}-sse-kms"), part_size, total_parts, EncryptionType::SSEKMS), MultipartTestConfig::new(format!("{base_object_key}-sse-c"), part_size, total_parts, create_sse_c_config()), ]; // Run tests for each encryption type for config in test_configs { test_multipart_upload_with_config(s3_client, bucket, &config).await?; } info!("βœ… Multipart uploads succeeded for every encryption type"); Ok(()) } /// Local KMS test environment management pub struct LocalKMSTestEnvironment { pub base_env: RustFSTestEnvironment, pub kms_keys_dir: String, } impl LocalKMSTestEnvironment { /// Create a new Local KMS test environment pub async fn new() -> Result<Self, Box<dyn std::error::Error + Send + Sync>> { let base_env = RustFSTestEnvironment::new().await?; let kms_keys_dir = format!("{}/kms-keys", base_env.temp_dir); fs::create_dir_all(&kms_keys_dir).await?; Ok(Self { base_env, kms_keys_dir }) } /// Start RustFS server configured for Local KMS backend with a default key pub async fn start_rustfs_for_local_kms(&mut self) -> Result<String, Box<dyn std::error::Error + Send + Sync>> { // Create a default key first let default_key_id = "rustfs-e2e-test-default-key"; create_key_with_specific_id(&self.kms_keys_dir, default_key_id).await?; let extra_args = vec![ "--kms-enable", "--kms-backend", "local", "--kms-key-dir", &self.kms_keys_dir, "--kms-default-key-id", default_key_id, ]; self.base_env.start_rustfs_server(extra_args).await?; Ok(default_key_id.to_string()) } /// Configure Local KMS backend with a predefined default key pub async fn configure_local_kms(&self) -> Result<String, Box<dyn std::error::Error + Send + Sync>> { // Use a fixed, predictable default key ID let default_key_id = "rustfs-e2e-test-default-key"; // Create the default key file first using our manual method create_key_with_specific_id(&self.kms_keys_dir, default_key_id).await?; // Configure KMS with the default key in one step let kms_config = serde_json::json!({ "backend_type": "local", "key_dir": self.kms_keys_dir, "file_permissions": 0o600, "default_key_id": default_key_id }) .to_string(); configure_kms(&self.base_env.url, &kms_config, &self.base_env.access_key, &self.base_env.secret_key).await?; Ok(default_key_id.to_string()) } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/kms/test_runner.rs
crates/e2e_test/src/kms/test_runner.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // #![allow(dead_code)] // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Unified KMS test suite runner //! //! This module provides a unified interface for running KMS tests with categorization, //! filtering, and comprehensive reporting capabilities. use crate::common::init_logging; use serial_test::serial; use std::time::Instant; use tokio::time::{Duration, sleep}; use tracing::{debug, error, info, warn}; /// Test category for organization and filtering #[derive(Debug, Clone, PartialEq, Eq, Hash)] pub enum TestCategory { CoreFunctionality, MultipartEncryption, EdgeCases, FaultRecovery, Comprehensive, Performance, } impl TestCategory { pub fn as_str(&self) -> &'static str { match self { TestCategory::CoreFunctionality => "core-functionality", TestCategory::MultipartEncryption => "multipart-encryption", TestCategory::EdgeCases => "edge-cases", TestCategory::FaultRecovery => "fault-recovery", TestCategory::Comprehensive => "comprehensive", TestCategory::Performance => "performance", } } } /// Test definition with metadata #[derive(Debug, Clone)] pub struct TestDefinition { pub name: String, pub description: String, pub category: TestCategory, pub estimated_duration: Duration, pub is_critical: bool, } impl TestDefinition { pub fn new( name: impl Into<String>, description: impl Into<String>, category: TestCategory, estimated_duration: Duration, is_critical: bool, ) -> Self { Self { name: name.into(), description: description.into(), category, estimated_duration, is_critical, } } } /// Test execution result #[derive(Debug, Clone)] pub struct TestResult { pub test_name: String, pub category: TestCategory, pub success: bool, pub duration: Duration, pub error_message: Option<String>, } impl TestResult { pub fn success(test_name: String, category: TestCategory, duration: Duration) -> Self { Self { test_name, category, success: true, duration, error_message: None, } } pub fn failure(test_name: String, category: TestCategory, duration: Duration, error: String) -> Self { Self { test_name, category, success: false, duration, error_message: Some(error), } } } /// Comprehensive test suite configuration #[derive(Debug, Clone)] pub struct TestSuiteConfig { pub categories: Vec<TestCategory>, pub include_critical_only: bool, pub max_duration: Option<Duration>, pub parallel_execution: bool, } impl Default for TestSuiteConfig { fn default() -> Self { Self { categories: vec![ TestCategory::CoreFunctionality, TestCategory::MultipartEncryption, TestCategory::EdgeCases, TestCategory::FaultRecovery, TestCategory::Comprehensive, ], include_critical_only: false, max_duration: None, parallel_execution: false, } } } /// Unified KMS test suite runner pub struct KMSTestSuite { tests: Vec<TestDefinition>, config: TestSuiteConfig, } impl KMSTestSuite { /// Create a new test suite with default configuration pub fn new() -> Self { let tests = vec![ // Core Functionality Tests TestDefinition::new( "test_local_kms_end_to_end", "End-to-end KMS test with all encryption types", TestCategory::CoreFunctionality, Duration::from_secs(60), true, ), TestDefinition::new( "test_local_kms_key_isolation", "Test KMS key isolation and security", TestCategory::CoreFunctionality, Duration::from_secs(45), true, ), // Multipart Encryption Tests TestDefinition::new( "test_local_kms_multipart_upload", "Test large file multipart upload with encryption", TestCategory::MultipartEncryption, Duration::from_secs(120), true, ), TestDefinition::new( "test_step1_basic_single_file_encryption", "Basic single file encryption test", TestCategory::MultipartEncryption, Duration::from_secs(30), false, ), TestDefinition::new( "test_step2_basic_multipart_upload_without_encryption", "Basic multipart upload without encryption", TestCategory::MultipartEncryption, Duration::from_secs(45), false, ), TestDefinition::new( "test_step3_multipart_upload_with_sse_s3", "Multipart upload with SSE-S3 encryption", TestCategory::MultipartEncryption, Duration::from_secs(60), true, ), TestDefinition::new( "test_step4_large_multipart_upload_with_encryption", "Large file multipart upload with encryption", TestCategory::MultipartEncryption, Duration::from_secs(90), false, ), TestDefinition::new( "test_step5_all_encryption_types_multipart", "All encryption types multipart test", TestCategory::MultipartEncryption, Duration::from_secs(120), true, ), // Edge Cases Tests TestDefinition::new( "test_kms_zero_byte_file_encryption", "Test encryption of zero-byte files", TestCategory::EdgeCases, Duration::from_secs(20), false, ), TestDefinition::new( "test_kms_single_byte_file_encryption", "Test encryption of single-byte files", TestCategory::EdgeCases, Duration::from_secs(20), false, ), TestDefinition::new( "test_kms_multipart_boundary_conditions", "Test multipart upload boundary conditions", TestCategory::EdgeCases, Duration::from_secs(45), false, ), TestDefinition::new( "test_kms_invalid_key_scenarios", "Test invalid key scenarios", TestCategory::EdgeCases, Duration::from_secs(30), false, ), TestDefinition::new( "test_kms_concurrent_encryption", "Test concurrent encryption operations", TestCategory::EdgeCases, Duration::from_secs(60), false, ), TestDefinition::new( "test_kms_key_validation_security", "Test key validation security", TestCategory::EdgeCases, Duration::from_secs(30), false, ), // Fault Recovery Tests TestDefinition::new( "test_kms_key_directory_unavailable", "Test KMS when key directory is unavailable", TestCategory::FaultRecovery, Duration::from_secs(45), false, ), TestDefinition::new( "test_kms_corrupted_key_files", "Test KMS with corrupted key files", TestCategory::FaultRecovery, Duration::from_secs(30), false, ), TestDefinition::new( "test_kms_multipart_upload_interruption", "Test multipart upload interruption recovery", TestCategory::FaultRecovery, Duration::from_secs(60), false, ), TestDefinition::new( "test_kms_resource_constraints", "Test KMS under resource constraints", TestCategory::FaultRecovery, Duration::from_secs(90), false, ), // Comprehensive Tests TestDefinition::new( "test_comprehensive_kms_full_workflow", "Full KMS workflow comprehensive test", TestCategory::Comprehensive, Duration::from_secs(300), true, ), TestDefinition::new( "test_comprehensive_stress_test", "KMS stress test with large datasets", TestCategory::Comprehensive, Duration::from_secs(400), false, ), TestDefinition::new( "test_comprehensive_key_isolation", "Comprehensive key isolation test", TestCategory::Comprehensive, Duration::from_secs(180), false, ), TestDefinition::new( "test_comprehensive_concurrent_operations", "Comprehensive concurrent operations test", TestCategory::Comprehensive, Duration::from_secs(240), false, ), TestDefinition::new( "test_comprehensive_performance_benchmark", "KMS performance benchmark test", TestCategory::Comprehensive, Duration::from_secs(360), false, ), ]; Self { tests, config: TestSuiteConfig::default(), } } /// Configure the test suite pub fn with_config(mut self, config: TestSuiteConfig) -> Self { self.config = config; self } /// Filter tests based on category pub fn filter_by_category(&self, category: &TestCategory) -> Vec<&TestDefinition> { self.tests.iter().filter(|test| &test.category == category).collect() } /// Filter tests based on criticality pub fn filter_critical_tests(&self) -> Vec<&TestDefinition> { self.tests.iter().filter(|test| test.is_critical).collect() } /// Get test summary by category pub fn get_category_summary(&self) -> std::collections::HashMap<TestCategory, Vec<&TestDefinition>> { let mut summary = std::collections::HashMap::new(); for test in &self.tests { summary.entry(test.category.clone()).or_insert_with(Vec::new).push(test); } summary } /// Run the complete test suite pub async fn run_test_suite(&self) -> Vec<TestResult> { init_logging(); info!("πŸš€ Starting unified KMS test suite"); let start_time = Instant::now(); let mut results = Vec::new(); // Filter tests based on configuration let tests_to_run: Vec<&TestDefinition> = self .tests .iter() .filter(|test| self.config.categories.contains(&test.category)) .filter(|test| !self.config.include_critical_only || test.is_critical) .collect(); info!("πŸ“Š Test plan: {} test(s) scheduled", tests_to_run.len()); for (i, test) in tests_to_run.iter().enumerate() { info!(" {}. {} ({})", i + 1, test.name, test.category.as_str()); } // Execute tests for (i, test_def) in tests_to_run.iter().enumerate() { info!("πŸ§ͺ Running test {}/{}: {}", i + 1, tests_to_run.len(), test_def.name); info!(" πŸ“ Description: {}", test_def.description); info!(" 🏷️ Category: {}", test_def.category.as_str()); info!(" ⏱️ Estimated duration: {:?}", test_def.estimated_duration); let test_start = Instant::now(); let result = self.run_single_test(test_def).await; let test_duration = test_start.elapsed(); match result { Ok(_) => { info!("βœ… Test passed: {} ({:.2}s)", test_def.name, test_duration.as_secs_f64()); results.push(TestResult::success(test_def.name.clone(), test_def.category.clone(), test_duration)); } Err(e) => { error!("❌ Test failed: {} ({:.2}s): {}", test_def.name, test_duration.as_secs_f64(), e); results.push(TestResult::failure( test_def.name.clone(), test_def.category.clone(), test_duration, e.to_string(), )); } } // Add delay between tests to avoid resource conflicts if i < tests_to_run.len() - 1 { debug!("⏸️ Waiting two seconds before the next test..."); sleep(Duration::from_secs(2)).await; } } let total_duration = start_time.elapsed(); self.print_test_summary(&results, total_duration); results } /// Run a single test by dispatching to the appropriate test function async fn run_single_test(&self, test_def: &TestDefinition) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { // This is a placeholder for test dispatch logic // In a real implementation, this would dispatch to actual test functions warn!("⚠️ Test '{}' is not implemented in the unified runner; skipping", test_def.name); Ok(()) } /// Print comprehensive test summary fn print_test_summary(&self, results: &[TestResult], total_duration: Duration) { info!("πŸ“Š KMS test suite summary"); info!("⏱️ Total duration: {:.2} seconds", total_duration.as_secs_f64()); info!("πŸ“ˆ Total tests: {}", results.len()); let passed = results.iter().filter(|r| r.success).count(); let failed = results.iter().filter(|r| !r.success).count(); info!("βœ… Passed: {}", passed); info!("❌ Failed: {}", failed); info!("πŸ“Š Success rate: {:.1}%", (passed as f64 / results.len() as f64) * 100.0); // Summary by category let mut category_summary: std::collections::HashMap<TestCategory, (usize, usize)> = std::collections::HashMap::new(); for result in results { let (total, passed_count) = category_summary.entry(result.category.clone()).or_insert((0, 0)); *total += 1; if result.success { *passed_count += 1; } } info!("πŸ“Š Category summary:"); for (category, (total, passed_count)) in category_summary { info!( " 🏷️ {}: {}/{} ({:.1}%)", category.as_str(), passed_count, total, (passed_count as f64 / total as f64) * 100.0 ); } // List failed tests if failed > 0 { warn!("❌ Failing tests:"); for result in results.iter().filter(|r| !r.success) { warn!( " - {}: {}", result.test_name, result.error_message.as_ref().unwrap_or(&"Unknown error".to_string()) ); } } } } /// Quick test suite for critical tests only #[tokio::test] #[serial] async fn test_kms_critical_suite() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { let config = TestSuiteConfig { categories: vec![TestCategory::CoreFunctionality, TestCategory::MultipartEncryption], include_critical_only: true, max_duration: Some(Duration::from_secs(600)), // 10 minutes max parallel_execution: false, }; let suite = KMSTestSuite::new().with_config(config); let results = suite.run_test_suite().await; let failed_count = results.iter().filter(|r| !r.success).count(); if failed_count > 0 { return Err(format!("Critical test suite failed: {failed_count} tests failed").into()); } info!("βœ… All critical tests passed"); Ok(()) } /// Full comprehensive test suite #[tokio::test] #[serial] async fn test_kms_full_suite() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { let suite = KMSTestSuite::new(); let results = suite.run_test_suite().await; let total_tests = results.len(); let failed_count = results.iter().filter(|r| !r.success).count(); let success_rate = ((total_tests - failed_count) as f64 / total_tests as f64) * 100.0; info!("πŸ“Š Full suite success rate: {:.1}%", success_rate); // Allow up to 10% failure rate for non-critical tests if success_rate < 90.0 { return Err(format!("Test suite success rate too low: {success_rate:.1}%").into()); } info!("βœ… Full test suite succeeded"); Ok(()) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/protocols/test_env.rs
crates/e2e_test/src/protocols/test_env.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Protocol test environment for FTPS and SFTP use std::net::TcpStream; use std::time::Duration; use tokio::time::sleep; use tracing::{info, warn}; /// Default credentials pub const DEFAULT_ACCESS_KEY: &str = "rustfsadmin"; pub const DEFAULT_SECRET_KEY: &str = "rustfsadmin"; /// Custom test environment that doesn't automatically stop servers pub struct ProtocolTestEnvironment { pub temp_dir: String, } impl ProtocolTestEnvironment { /// Create a new test environment /// This environment won't stop any server when dropped pub fn new() -> Result<Self, Box<dyn std::error::Error + Send + Sync>> { let temp_dir = format!("/tmp/rustfs_protocol_test_{}", uuid::Uuid::new_v4()); std::fs::create_dir_all(&temp_dir)?; Ok(Self { temp_dir }) } /// Wait for server to be ready pub async fn wait_for_port_ready(port: u16, max_attempts: u32) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { let address = format!("127.0.0.1:{}", port); info!("Waiting for server to be ready on {}", address); for i in 0..max_attempts { if TcpStream::connect(&address).is_ok() { info!("Server is ready after {} s", i + 1); return Ok(()); } if i == max_attempts - 1 { return Err(format!("Server did not become ready within {} s", max_attempts).into()); } sleep(Duration::from_secs(1)).await; } Ok(()) } } // Implement Drop trait that doesn't stop servers impl Drop for ProtocolTestEnvironment { fn drop(&mut self) { // Clean up temp directory only, don't stop any server if let Err(e) = std::fs::remove_dir_all(&self.temp_dir) { warn!("Failed to clean up temp directory {}: {}", self.temp_dir, e); } } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/protocols/mod.rs
crates/e2e_test/src/protocols/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Protocol tests for FTPS and SFTP pub mod ftps_core; pub mod test_env; pub mod test_runner;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/protocols/ftps_core.rs
crates/e2e_test/src/protocols/ftps_core.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Core FTPS tests use crate::common::rustfs_binary_path; use crate::protocols::test_env::{DEFAULT_ACCESS_KEY, DEFAULT_SECRET_KEY, ProtocolTestEnvironment}; use anyhow::Result; use rcgen::generate_simple_self_signed; use rustls::crypto::aws_lc_rs::default_provider; use rustls::{ClientConfig, RootCertStore}; use std::io::Cursor; use std::path::PathBuf; use std::sync::Arc; use suppaftp::RustlsConnector; use suppaftp::RustlsFtpStream; use tokio::process::Command; use tracing::info; // Fixed FTPS port for testing const FTPS_PORT: u16 = 9021; const FTPS_ADDRESS: &str = "127.0.0.1:9021"; /// Test FTPS: put, ls, mkdir, rmdir, delete operations pub async fn test_ftps_core_operations() -> Result<()> { let env = ProtocolTestEnvironment::new().map_err(|e| anyhow::anyhow!("{}", e))?; // Generate and write certificate let cert = generate_simple_self_signed(vec!["localhost".to_string(), "127.0.0.1".to_string()])?; let cert_path = PathBuf::from(&env.temp_dir).join("ftps.crt"); let key_path = PathBuf::from(&env.temp_dir).join("ftps.key"); let cert_pem = cert.cert.pem(); let key_pem = cert.signing_key.serialize_pem(); tokio::fs::write(&cert_path, &cert_pem).await?; tokio::fs::write(&key_path, &key_pem).await?; // Start server manually info!("Starting FTPS server on {}", FTPS_ADDRESS); let binary_path = rustfs_binary_path(); let mut server_process = Command::new(&binary_path) .env("RUSTFS_FTPS_ENABLE", "true") .env("RUSTFS_FTPS_ADDRESS", FTPS_ADDRESS) .env("RUSTFS_FTPS_CERTS_FILE", cert_path.to_str().unwrap()) .env("RUSTFS_FTPS_KEY_FILE", key_path.to_str().unwrap()) .arg(&env.temp_dir) .spawn()?; // Ensure server is cleaned up even on failure let result = async { // Wait for server to be ready ProtocolTestEnvironment::wait_for_port_ready(FTPS_PORT, 30) .await .map_err(|e| anyhow::anyhow!("{}", e))?; // Install the aws-lc-rs crypto provider default_provider() .install_default() .map_err(|e| anyhow::anyhow!("Failed to install crypto provider: {:?}", e))?; // Create a simple rustls config that accepts any certificate for testing let mut root_store = RootCertStore::empty(); // Add the self-signed certificate to the trust store for e2e // Note: In a real environment, you'd use proper root certificates let cert_pem = cert.cert.pem(); let cert_der = rustls_pemfile::certs(&mut Cursor::new(cert_pem)) .collect::<Result<Vec<_>, _>>() .map_err(|e| anyhow::anyhow!("Failed to parse cert: {}", e))?; root_store.add_parsable_certificates(cert_der); let config = ClientConfig::builder() .with_root_certificates(root_store) .with_no_client_auth(); // Wrap in suppaftp's RustlsConnector let tls_connector = RustlsConnector::from(Arc::new(config)); // Connect to FTPS server let ftp_stream = RustlsFtpStream::connect(FTPS_ADDRESS).map_err(|e| anyhow::anyhow!("Failed to connect: {}", e))?; // Upgrade to secure connection let mut ftp_stream = ftp_stream .into_secure(tls_connector, "127.0.0.1") .map_err(|e| anyhow::anyhow!("Failed to upgrade to TLS: {}", e))?; ftp_stream.login(DEFAULT_ACCESS_KEY, DEFAULT_SECRET_KEY)?; info!("Testing FTPS: mkdir bucket"); let bucket_name = "testbucket"; ftp_stream.mkdir(bucket_name)?; info!("PASS: mkdir bucket '{}' successful", bucket_name); info!("Testing FTPS: cd to bucket"); ftp_stream.cwd(bucket_name)?; info!("PASS: cd to bucket '{}' successful", bucket_name); info!("Testing FTPS: put file"); let filename = "test.txt"; let content = "Hello, FTPS!"; ftp_stream.put_file(filename, &mut Cursor::new(content.as_bytes()))?; info!("PASS: put file '{}' ({} bytes) successful", filename, content.len()); info!("Testing FTPS: download file"); let downloaded_content = ftp_stream.retr(filename, |stream| { let mut buffer = Vec::new(); stream.read_to_end(&mut buffer).map_err(suppaftp::FtpError::ConnectionError)?; Ok(buffer) })?; let downloaded_str = String::from_utf8(downloaded_content)?; assert_eq!(downloaded_str, content, "Downloaded content should match uploaded content"); info!("PASS: download file '{}' successful, content matches", filename); info!("Testing FTPS: ls list objects in bucket"); let list = ftp_stream.list(None)?; assert!(list.iter().any(|line| line.contains(filename)), "File should appear in list"); info!("PASS: ls command successful, file '{}' found in bucket", filename); info!("Testing FTPS: ls . (list current directory)"); let list_dot = ftp_stream.list(Some(".")).unwrap_or_else(|_| ftp_stream.list(None).unwrap()); assert!(list_dot.iter().any(|line| line.contains(filename)), "File should appear in ls ."); info!("PASS: ls . successful, file '{}' found", filename); info!("Testing FTPS: ls / (list root directory)"); let list_root = ftp_stream.list(Some("/")).unwrap(); assert!(list_root.iter().any(|line| line.contains(bucket_name)), "Bucket should appear in ls /"); assert!(!list_root.iter().any(|line| line.contains(filename)), "File should not appear in ls /"); info!( "PASS: ls / successful, bucket '{}' found, file '{}' not found in root", bucket_name, filename ); info!("Testing FTPS: ls /. (list root directory with /.)"); let list_root_dot = ftp_stream .list(Some("/.")) .unwrap_or_else(|_| ftp_stream.list(Some("/")).unwrap()); assert!( list_root_dot.iter().any(|line| line.contains(bucket_name)), "Bucket should appear in ls /." ); info!("PASS: ls /. successful, bucket '{}' found", bucket_name); info!("Testing FTPS: ls /bucket (list bucket by absolute path)"); let list_bucket = ftp_stream.list(Some(&format!("/{}", bucket_name))).unwrap(); assert!(list_bucket.iter().any(|line| line.contains(filename)), "File should appear in ls /bucket"); info!("PASS: ls /{} successful, file '{}' found", bucket_name, filename); info!("Testing FTPS: cd . (stay in current directory)"); ftp_stream.cwd(".")?; info!("PASS: cd . successful (stays in current directory)"); info!("Testing FTPS: ls after cd . (should still see file)"); let list_after_dot = ftp_stream.list(None)?; assert!( list_after_dot.iter().any(|line| line.contains(filename)), "File should still appear in list after cd ." ); info!("PASS: ls after cd . successful, file '{}' still found in bucket", filename); info!("Testing FTPS: cd / (go to root directory)"); ftp_stream.cwd("/")?; info!("PASS: cd / successful (back to root directory)"); info!("Testing FTPS: ls after cd / (should see bucket only)"); let root_list_after = ftp_stream.list(None)?; assert!( !root_list_after.iter().any(|line| line.contains(filename)), "File should not appear in root ls" ); assert!( root_list_after.iter().any(|line| line.contains(bucket_name)), "Bucket should appear in root ls" ); info!("PASS: ls after cd / successful, file not in root, bucket '{}' found in root", bucket_name); info!("Testing FTPS: cd back to bucket"); ftp_stream.cwd(bucket_name)?; info!("PASS: cd back to bucket '{}' successful", bucket_name); info!("Testing FTPS: delete object"); ftp_stream.rm(filename)?; info!("PASS: delete object '{}' successful", filename); info!("Testing FTPS: ls verify object deleted"); let list_after = ftp_stream.list(None)?; assert!(!list_after.iter().any(|line| line.contains(filename)), "File should be deleted"); info!("PASS: ls after delete successful, file '{}' is not found", filename); info!("Testing FTPS: cd up to root directory"); ftp_stream.cdup()?; info!("PASS: cd up to root directory successful"); info!("Testing FTPS: cd to nonexistent bucket (should fail)"); let nonexistent_bucket = "nonexistent-bucket"; let cd_result = ftp_stream.cwd(nonexistent_bucket); assert!(cd_result.is_err(), "cd to nonexistent bucket should fail"); info!("PASS: cd to nonexistent bucket '{}' failed as expected", nonexistent_bucket); info!("Testing FTPS: ls verify bucket exists in root"); let root_list = ftp_stream.list(None)?; assert!(root_list.iter().any(|line| line.contains(bucket_name)), "Bucket should exist in root"); info!("PASS: ls root successful, bucket '{}' found in root", bucket_name); info!("Testing FTPS: rmdir delete bucket"); ftp_stream.rmdir(bucket_name)?; info!("PASS: rmdir bucket '{}' successful", bucket_name); info!("Testing FTPS: ls verify bucket deleted"); let root_list_after = ftp_stream.list(None)?; assert!(!root_list_after.iter().any(|line| line.contains(bucket_name)), "Bucket should be deleted"); info!("PASS: ls root after delete successful, bucket '{}' is not found", bucket_name); ftp_stream.quit()?; info!("FTPS core tests passed"); Ok(()) } .await; // Always cleanup server process let _ = server_process.kill().await; let _ = server_process.wait().await; result }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/protocols/test_runner.rs
crates/e2e_test/src/protocols/test_runner.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Protocol test runner use crate::common::init_logging; use crate::protocols::ftps_core::test_ftps_core_operations; use std::time::Instant; use tokio::time::{Duration, sleep}; use tracing::{error, info}; /// Test result #[derive(Debug, Clone)] pub struct TestResult { pub test_name: String, pub success: bool, pub error_message: Option<String>, } impl TestResult { pub fn success(test_name: String) -> Self { Self { test_name, success: true, error_message: None, } } pub fn failure(test_name: String, error: String) -> Self { Self { test_name, success: false, error_message: Some(error), } } } /// Protocol test suite pub struct ProtocolTestSuite { tests: Vec<TestDefinition>, } #[derive(Debug, Clone)] struct TestDefinition { name: String, } impl ProtocolTestSuite { /// Create default test suite pub fn new() -> Self { let tests = vec![ TestDefinition { name: "test_ftps_core_operations".to_string(), }, // TestDefinition { name: "test_sftp_core_operations".to_string() }, ]; Self { tests } } /// Run test suite pub async fn run_test_suite(&self) -> Vec<TestResult> { init_logging(); info!("Starting Protocol test suite"); let start_time = Instant::now(); let mut results = Vec::new(); info!("Scheduled {} tests", self.tests.len()); // Run tests for (i, test_def) in self.tests.iter().enumerate() { let test_description = match test_def.name.as_str() { "test_ftps_core_operations" => { info!("=== Starting FTPS Module Test ==="); "FTPS core operations (put, ls, mkdir, rmdir, delete)" } "test_sftp_core_operations" => { info!("=== Starting SFTP Module Test ==="); "SFTP core operations (put, ls, mkdir, rmdir, delete)" } _ => "", }; info!("Test {}/{} - {}", i + 1, self.tests.len(), test_description); info!("Running: {}", test_def.name); let test_start = Instant::now(); let result = self.run_single_test(test_def).await; let test_duration = test_start.elapsed(); match result { Ok(_) => { info!("Test passed: {} ({:.2}s)", test_def.name, test_duration.as_secs_f64()); results.push(TestResult::success(test_def.name.clone())); } Err(e) => { error!("Test failed: {} ({:.2}s): {}", test_def.name, test_duration.as_secs_f64(), e); results.push(TestResult::failure(test_def.name.clone(), e.to_string())); } } // Delay between tests to avoid resource conflicts if i < self.tests.len() - 1 { sleep(Duration::from_secs(2)).await; } } // Print summary self.print_summary(&results, start_time.elapsed()); results } /// Run a single test async fn run_single_test(&self, test_def: &TestDefinition) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { match test_def.name.as_str() { "test_ftps_core_operations" => test_ftps_core_operations().await.map_err(|e| e.into()), // "test_sftp_core_operations" => test_sftp_core_operations().await.map_err(|e| e.into()), _ => Err(format!("Test {} not implemented", test_def.name).into()), } } /// Print test summary fn print_summary(&self, results: &[TestResult], total_duration: Duration) { info!("=== Test Suite Summary ==="); info!("Total duration: {:.2}s", total_duration.as_secs_f64()); info!("Total tests: {}", results.len()); let passed = results.iter().filter(|r| r.success).count(); let failed = results.len() - passed; let success_rate = (passed as f64 / results.len() as f64) * 100.0; info!("Passed: {} | Failed: {}", passed, failed); info!("Success rate: {:.1}%", success_rate); if failed > 0 { error!("Failed tests:"); for result in results.iter().filter(|r| !r.success) { error!(" - {}: {}", result.test_name, result.error_message.as_ref().unwrap()); } } } } /// Test suite #[tokio::test] async fn test_protocol_core_suite() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { let suite = ProtocolTestSuite::new(); let results = suite.run_test_suite().await; let failed = results.iter().filter(|r| !r.success).count(); if failed > 0 { return Err(format!("Protocol tests failed: {failed} failures").into()); } info!("All protocol tests passed"); Ok(()) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/reliant/conditional_writes.rs
crates/e2e_test/src/reliant/conditional_writes.rs
#![cfg(test)] use aws_config::meta::region::RegionProviderChain; use aws_sdk_s3::Client; use aws_sdk_s3::config::{Credentials, Region}; use aws_sdk_s3::error::SdkError; use aws_sdk_s3::types::{CompletedMultipartUpload, CompletedPart}; use bytes::Bytes; use serial_test::serial; use std::error::Error; const ENDPOINT: &str = "http://localhost:9000"; const ACCESS_KEY: &str = "rustfsadmin"; const SECRET_KEY: &str = "rustfsadmin"; const BUCKET: &str = "api-test"; async fn create_aws_s3_client() -> Result<Client, Box<dyn Error>> { let region_provider = RegionProviderChain::default_provider().or_else(Region::new("us-east-1")); let shared_config = aws_config::defaults(aws_config::BehaviorVersion::latest()) .region(region_provider) .credentials_provider(Credentials::new(ACCESS_KEY, SECRET_KEY, None, None, "static")) .endpoint_url(ENDPOINT) .load() .await; let client = Client::from_conf( aws_sdk_s3::Config::from(&shared_config) .to_builder() .force_path_style(true) .build(), ); Ok(client) } /// Setup test bucket, creating it if it doesn't exist async fn setup_test_bucket(client: &Client) -> Result<(), Box<dyn Error>> { match client.create_bucket().bucket(BUCKET).send().await { Ok(_) => {} Err(SdkError::ServiceError(e)) => { let e = e.into_err(); let error_code = e.meta().code().unwrap_or(""); if !error_code.eq("BucketAlreadyExists") { return Err(e.into()); } } Err(e) => { return Err(e.into()); } } Ok(()) } /// Generate test data of specified size fn generate_test_data(size: usize) -> Vec<u8> { let pattern = b"0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz"; let mut data = Vec::with_capacity(size); for i in 0..size { data.push(pattern[i % pattern.len()]); } data } /// Upload an object and return its ETag async fn upload_object_with_metadata(client: &Client, bucket: &str, key: &str, data: &[u8]) -> Result<String, Box<dyn Error>> { let response = client .put_object() .bucket(bucket) .key(key) .body(Bytes::from(data.to_vec()).into()) .send() .await?; let etag = response.e_tag().unwrap_or("").to_string(); Ok(etag) } /// Cleanup test objects from bucket async fn cleanup_objects(client: &Client, bucket: &str, keys: &[&str]) { for key in keys { let _ = client.delete_object().bucket(bucket).key(*key).send().await; } } /// Generate unique test object key fn generate_test_key(prefix: &str) -> String { use std::time::{SystemTime, UNIX_EPOCH}; let timestamp = SystemTime::now().duration_since(UNIX_EPOCH).unwrap().as_nanos(); format!("{prefix}-{timestamp}") } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_conditional_put_okay() -> Result<(), Box<dyn std::error::Error>> { let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; let test_key = generate_test_key("conditional-put-ok"); let initial_data = generate_test_data(1024); // 1KB test data let updated_data = generate_test_data(2048); // 2KB updated data // Upload initial object and get its ETag let initial_etag = upload_object_with_metadata(&client, BUCKET, &test_key, &initial_data).await?; // Test 1: PUT with matching If-Match condition (should succeed) let response1 = client .put_object() .bucket(BUCKET) .key(&test_key) .body(Bytes::from(updated_data.clone()).into()) .if_match(&initial_etag) .send() .await; assert!(response1.is_ok(), "PUT with matching If-Match should succeed"); // Test 2: PUT with non-matching If-None-Match condition (should succeed) let fake_etag = "\"fake-etag-12345\""; let response2 = client .put_object() .bucket(BUCKET) .key(&test_key) .body(Bytes::from(updated_data.clone()).into()) .if_none_match(fake_etag) .send() .await; assert!(response2.is_ok(), "PUT with non-matching If-None-Match should succeed"); // Cleanup cleanup_objects(&client, BUCKET, &[&test_key]).await; Ok(()) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_conditional_put_failed() -> Result<(), Box<dyn std::error::Error>> { let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; let test_key = generate_test_key("conditional-put-failed"); let initial_data = generate_test_data(1024); let updated_data = generate_test_data(2048); // Upload initial object and get its ETag let initial_etag = upload_object_with_metadata(&client, BUCKET, &test_key, &initial_data).await?; // Test 1: PUT with non-matching If-Match condition (should fail with 412) let fake_etag = "\"fake-etag-should-not-match\""; let response1 = client .put_object() .bucket(BUCKET) .key(&test_key) .body(Bytes::from(updated_data.clone()).into()) .if_match(fake_etag) .send() .await; assert!(response1.is_err(), "PUT with non-matching If-Match should fail"); if let Err(e) = response1 { if let SdkError::ServiceError(e) = e { let e = e.into_err(); let error_code = e.meta().code().unwrap_or(""); assert_eq!("PreconditionFailed", error_code); } else { panic!("Unexpected error: {e:?}"); } } // Test 2: PUT with matching If-None-Match condition (should fail with 412) let response2 = client .put_object() .bucket(BUCKET) .key(&test_key) .body(Bytes::from(updated_data.clone()).into()) .if_none_match(&initial_etag) .send() .await; assert!(response2.is_err(), "PUT with matching If-None-Match should fail"); if let Err(e) = response2 { if let SdkError::ServiceError(e) = e { let e = e.into_err(); let error_code = e.meta().code().unwrap_or(""); assert_eq!("PreconditionFailed", error_code); } else { panic!("Unexpected error: {e:?}"); } } // Cleanup - only need to clean up the initial object since failed PUTs shouldn't create objects cleanup_objects(&client, BUCKET, &[&test_key]).await; Ok(()) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_conditional_put_when_object_does_not_exist() -> Result<(), Box<dyn std::error::Error>> { let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; let key = "some_key"; cleanup_objects(&client, BUCKET, &[key]).await; // When the object does not exist, the If-Match condition should always fail let response1 = client .put_object() .bucket(BUCKET) .key(key) .body(Bytes::from(generate_test_data(1024)).into()) .if_match("*") .send() .await; assert!(response1.is_err()); if let Err(e) = response1 { if let SdkError::ServiceError(e) = e { let e = e.into_err(); let error_code = e.meta().code().unwrap_or(""); assert_eq!("NoSuchKey", error_code); } else { panic!("Unexpected error: {e:?}"); } } // When the object does not exist, the If-None-Match condition should be able to succeed let response2 = client .put_object() .bucket(BUCKET) .key(key) .body(Bytes::from(generate_test_data(1024)).into()) .if_none_match("*") .send() .await; assert!(response2.is_ok()); cleanup_objects(&client, BUCKET, &[key]).await; Ok(()) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_conditional_multi_part_upload() -> Result<(), Box<dyn std::error::Error>> { let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; let test_key = generate_test_key("multipart-upload-ok"); let test_data = generate_test_data(1024); let initial_etag = upload_object_with_metadata(&client, BUCKET, &test_key, &test_data).await?; let part_size = 5 * 1024 * 1024; // 5MB per part (minimum for multipart) let num_parts = 3; let mut parts = Vec::new(); // Initiate multipart upload let initiate_response = client.create_multipart_upload().bucket(BUCKET).key(&test_key).send().await?; let upload_id = initiate_response .upload_id() .ok_or(std::io::Error::other("No upload ID returned"))?; // Upload parts for part_number in 1..=num_parts { let part_data = generate_test_data(part_size); let upload_part_response = client .upload_part() .bucket(BUCKET) .key(&test_key) .upload_id(upload_id) .part_number(part_number) .body(Bytes::from(part_data).into()) .send() .await?; let part_etag = upload_part_response .e_tag() .ok_or(std::io::Error::other("Do not have etag"))? .to_string(); let completed_part = CompletedPart::builder().part_number(part_number).e_tag(part_etag).build(); parts.push(completed_part); } // Complete multipart upload let completed_upload = CompletedMultipartUpload::builder().set_parts(Some(parts)).build(); // Test 1: Multipart upload with wildcard If-None-Match, should fail let complete_response = client .complete_multipart_upload() .bucket(BUCKET) .key(&test_key) .upload_id(upload_id) .multipart_upload(completed_upload.clone()) .if_none_match("*") .send() .await; assert!(complete_response.is_err()); // Test 2: Multipart upload with matching If-None-Match, should fail let complete_response = client .complete_multipart_upload() .bucket(BUCKET) .key(&test_key) .upload_id(upload_id) .multipart_upload(completed_upload.clone()) .if_none_match(initial_etag.clone()) .send() .await; assert!(complete_response.is_err()); // Test 3: Multipart upload with unmatching If-Match, should fail let complete_response = client .complete_multipart_upload() .bucket(BUCKET) .key(&test_key) .upload_id(upload_id) .multipart_upload(completed_upload.clone()) .if_match("\"abcdef\"") .send() .await; assert!(complete_response.is_err()); // Test 4: Multipart upload with matching If-Match, should succeed let complete_response = client .complete_multipart_upload() .bucket(BUCKET) .key(&test_key) .upload_id(upload_id) .multipart_upload(completed_upload.clone()) .if_match(initial_etag) .send() .await; assert!(complete_response.is_ok()); // Cleanup cleanup_objects(&client, BUCKET, &[&test_key]).await; Ok(()) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/reliant/lock.rs
crates/e2e_test/src/reliant/lock.rs
#![cfg(test)] // Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use async_trait::async_trait; use rustfs_ecstore::disk::endpoint::Endpoint; use rustfs_lock::client::{LockClient, local::LocalClient, remote::RemoteClient}; use rustfs_lock::types::{LockInfo, LockResponse, LockStats}; use rustfs_lock::{LockId, LockMetadata, LockPriority, LockType}; use rustfs_lock::{LockRequest, NamespaceLock, NamespaceLockManager}; use rustfs_protos::{node_service_time_out_client, proto_gen::node_service::GenerallyLockRequest}; use serial_test::serial; use std::{collections::HashMap, error::Error, sync::Arc, time::Duration}; use tokio::time::sleep; use tonic::Request; use url::Url; const CLUSTER_ADDR: &str = "http://localhost:9000"; fn get_cluster_endpoints() -> Vec<Endpoint> { vec![Endpoint { url: Url::parse(CLUSTER_ADDR).unwrap(), is_local: false, pool_idx: 0, set_idx: 0, disk_idx: 0, }] } async fn create_unique_clients(endpoints: &[Endpoint]) -> Result<Vec<Arc<dyn LockClient>>, Box<dyn Error>> { let mut unique_endpoints: HashMap<String, &Endpoint> = HashMap::new(); for endpoint in endpoints { if endpoint.is_local { unique_endpoints.insert("local".to_string(), endpoint); } else { let host_port = format!( "{}:{}", endpoint.url.host_str().unwrap_or("localhost"), endpoint.url.port().unwrap_or(9000) ); unique_endpoints.insert(host_port, endpoint); } } let mut clients = Vec::new(); for (_key, endpoint) in unique_endpoints { if endpoint.is_local { clients.push(Arc::new(LocalClient::new()) as Arc<dyn LockClient>); } else { clients.push(Arc::new(RemoteClient::new(endpoint.url.to_string())) as Arc<dyn LockClient>); } } Ok(clients) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_guard_drop_releases_exclusive_lock_local() -> Result<(), Box<dyn Error>> { // Single local client; no external server required let client: Arc<dyn LockClient> = Arc::new(LocalClient::new()); let ns_lock = NamespaceLock::with_clients("e2e_guard_local".to_string(), vec![client]); // Acquire exclusive guard let g1 = ns_lock .lock_guard("guard_exclusive", "owner1", Duration::from_millis(100), Duration::from_secs(5)) .await?; assert!(g1.is_some(), "first guard acquisition should succeed"); // While g1 is alive, second exclusive acquisition should fail let g2 = ns_lock .lock_guard("guard_exclusive", "owner2", Duration::from_millis(50), Duration::from_secs(5)) .await?; assert!(g2.is_none(), "second guard acquisition should fail while first is held"); // Drop first guard to trigger background release drop(g1); // Give the background unlock worker a short moment to process sleep(Duration::from_millis(80)).await; // Now acquisition should succeed let g3 = ns_lock .lock_guard("guard_exclusive", "owner2", Duration::from_millis(100), Duration::from_secs(5)) .await?; assert!(g3.is_some(), "acquisition should succeed after guard drop releases the lock"); drop(g3); Ok(()) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_guard_shared_then_write_after_drop() -> Result<(), Box<dyn Error>> { // Two shared read guards should coexist; write should be blocked until they drop let client: Arc<dyn LockClient> = Arc::new(LocalClient::new()); let ns_lock = NamespaceLock::with_clients("e2e_guard_rw".to_string(), vec![client]); // Acquire two read guards let r1 = ns_lock .rlock_guard("rw_resource", "reader1", Duration::from_millis(100), Duration::from_secs(5)) .await?; let r2 = ns_lock .rlock_guard("rw_resource", "reader2", Duration::from_millis(100), Duration::from_secs(5)) .await?; assert!(r1.is_some() && r2.is_some(), "both read guards should be acquired"); // Attempt write while readers hold the lock should fail let w_fail = ns_lock .lock_guard("rw_resource", "writer", Duration::from_millis(50), Duration::from_secs(5)) .await?; assert!(w_fail.is_none(), "write should be blocked when read guards are active"); // Drop read guards to release drop(r1); drop(r2); sleep(Duration::from_millis(80)).await; // Now write should succeed let w_ok = ns_lock .lock_guard("rw_resource", "writer", Duration::from_millis(150), Duration::from_secs(5)) .await?; assert!(w_ok.is_some(), "write should succeed after read guards are dropped"); drop(w_ok); Ok(()) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_lock_unlock_rpc() -> Result<(), Box<dyn Error>> { let args = LockRequest { lock_id: LockId::new_deterministic("dandan"), resource: "dandan".to_string(), lock_type: LockType::Exclusive, owner: "dd".to_string(), acquire_timeout: Duration::from_secs(30), ttl: Duration::from_secs(30), metadata: LockMetadata::default(), priority: LockPriority::Normal, deadlock_detection: false, }; let args = serde_json::to_string(&args)?; let mut client = node_service_time_out_client(&CLUSTER_ADDR.to_string()).await?; println!("got client"); let request = Request::new(GenerallyLockRequest { args: args.clone() }); println!("start request"); let response = client.lock(request).await?.into_inner(); println!("request ended"); if let Some(error_info) = response.error_info { panic!("can not get lock: {error_info}"); } let request = Request::new(GenerallyLockRequest { args }); let response = client.un_lock(request).await?.into_inner(); if let Some(error_info) = response.error_info { panic!("can not get un_lock: {error_info}"); } Ok(()) } /// Mock client that simulates remote node failures #[derive(Debug)] struct FailingMockClient { local_client: Arc<dyn LockClient>, should_fail_acquire: bool, should_fail_release: bool, } impl FailingMockClient { fn new(should_fail_acquire: bool, should_fail_release: bool) -> Self { Self { local_client: Arc::new(LocalClient::new()), should_fail_acquire, should_fail_release, } } } #[async_trait] impl LockClient for FailingMockClient { async fn acquire_exclusive(&self, request: &LockRequest) -> rustfs_lock::error::Result<LockResponse> { if self.should_fail_acquire { // Simulate network timeout or remote node failure return Ok(LockResponse::failure("Simulated remote node failure", Duration::from_millis(100))); } self.local_client.acquire_exclusive(request).await } async fn acquire_shared(&self, request: &LockRequest) -> rustfs_lock::error::Result<LockResponse> { if self.should_fail_acquire { return Ok(LockResponse::failure("Simulated remote node failure", Duration::from_millis(100))); } self.local_client.acquire_shared(request).await } async fn release(&self, lock_id: &LockId) -> rustfs_lock::error::Result<bool> { if self.should_fail_release { return Err(rustfs_lock::error::LockError::internal("Simulated release failure")); } self.local_client.release(lock_id).await } async fn refresh(&self, lock_id: &LockId) -> rustfs_lock::error::Result<bool> { self.local_client.refresh(lock_id).await } async fn force_release(&self, lock_id: &LockId) -> rustfs_lock::error::Result<bool> { self.local_client.force_release(lock_id).await } async fn check_status(&self, lock_id: &LockId) -> rustfs_lock::error::Result<Option<LockInfo>> { self.local_client.check_status(lock_id).await } async fn get_stats(&self) -> rustfs_lock::error::Result<LockStats> { self.local_client.get_stats().await } async fn close(&self) -> rustfs_lock::error::Result<()> { self.local_client.close().await } async fn is_online(&self) -> bool { if self.should_fail_acquire { return false; // Simulate offline node } true // Simulate online node } async fn is_local(&self) -> bool { false // Simulate remote client } } #[tokio::test] #[serial] async fn test_transactional_lock_with_remote_failure() -> Result<(), Box<dyn Error>> { println!("πŸ§ͺ Testing transactional lock with simulated remote node failure"); // Create a two-node cluster: one local (success) + one remote (failure) let local_client: Arc<dyn LockClient> = Arc::new(LocalClient::new()); let failing_remote_client: Arc<dyn LockClient> = Arc::new(FailingMockClient::new(true, false)); let clients = vec![local_client, failing_remote_client]; let ns_lock = NamespaceLock::with_clients("test_transactional".to_string(), clients); let resource = "critical_resource".to_string(); // Test single lock operation with 2PC println!("πŸ“ Testing single lock with remote failure..."); let request = LockRequest::new(&resource, LockType::Exclusive, "test_owner").with_ttl(Duration::from_secs(30)); let response = ns_lock.acquire_lock(&request).await?; // Should fail because quorum (2/2) is not met due to remote failure assert!(!response.success, "Lock should fail due to remote node failure"); println!("βœ… Single lock correctly failed due to remote node failure"); // Verify no locks are left behind on the local node let local_client_direct = LocalClient::new(); let lock_id = LockId::new_deterministic(&ns_lock.get_resource_key(&resource)); let lock_status = local_client_direct.check_status(&lock_id).await?; assert!(lock_status.is_none(), "No lock should remain on local node after rollback"); println!("βœ… Verified rollback: no locks left on local node"); Ok(()) } #[tokio::test] #[serial] async fn test_transactional_batch_lock_with_mixed_failures() -> Result<(), Box<dyn Error>> { println!("πŸ§ͺ Testing transactional batch lock with mixed node failures"); // Create a cluster with different failure patterns let local_client: Arc<dyn LockClient> = Arc::new(LocalClient::new()); let failing_remote_client: Arc<dyn LockClient> = Arc::new(FailingMockClient::new(true, false)); let clients = vec![local_client, failing_remote_client]; let ns_lock = NamespaceLock::with_clients("test_batch_transactional".to_string(), clients); let resources = vec!["resource_1".to_string(), "resource_2".to_string(), "resource_3".to_string()]; println!("πŸ“ Testing batch lock with remote failure..."); let result = ns_lock .lock_batch(&resources, "batch_owner", Duration::from_millis(100), Duration::from_secs(30)) .await?; // Should fail because remote node cannot acquire locks assert!(!result, "Batch lock should fail due to remote node failure"); println!("βœ… Batch lock correctly failed due to remote node failure"); // Verify no locks are left behind on any resource let local_client_direct = LocalClient::new(); for resource in &resources { let lock_id = LockId::new_deterministic(&ns_lock.get_resource_key(resource)); let lock_status = local_client_direct.check_status(&lock_id).await?; assert!(lock_status.is_none(), "No lock should remain for resource: {resource}"); } println!("βœ… Verified rollback: no locks left on any resource"); Ok(()) } #[tokio::test] #[serial] async fn test_transactional_lock_with_quorum_success() -> Result<(), Box<dyn Error>> { println!("πŸ§ͺ Testing transactional lock with quorum success"); // Create a three-node cluster where 2 succeed and 1 fails (quorum = 2 automatically) let local_client1: Arc<dyn LockClient> = Arc::new(LocalClient::new()); let local_client2: Arc<dyn LockClient> = Arc::new(LocalClient::new()); let failing_remote_client: Arc<dyn LockClient> = Arc::new(FailingMockClient::new(true, false)); let clients = vec![local_client1, local_client2, failing_remote_client]; let ns_lock = NamespaceLock::with_clients("test_quorum".to_string(), clients); let resource = "quorum_resource".to_string(); println!("πŸ“ Testing lock with automatic quorum=2, 2 success + 1 failure..."); let request = LockRequest::new(&resource, LockType::Exclusive, "quorum_owner").with_ttl(Duration::from_secs(30)); let response = ns_lock.acquire_lock(&request).await?; // Should fail because we require all nodes to succeed for consistency // (even though quorum is met, the implementation requires all nodes for consistency) assert!(!response.success, "Lock should fail due to consistency requirement"); println!("βœ… Lock correctly failed due to consistency requirement (partial success rolled back)"); Ok(()) } #[tokio::test] #[serial] async fn test_transactional_lock_rollback_on_release_failure() -> Result<(), Box<dyn Error>> { println!("πŸ§ͺ Testing rollback behavior when release fails"); // Create clients where acquire succeeds but release fails let local_client: Arc<dyn LockClient> = Arc::new(LocalClient::new()); let failing_release_client: Arc<dyn LockClient> = Arc::new(FailingMockClient::new(false, true)); let clients = vec![local_client, failing_release_client]; let ns_lock = NamespaceLock::with_clients("test_release_failure".to_string(), clients); let resource = "release_test_resource".to_string(); println!("πŸ“ Testing lock acquisition with release failure handling..."); let request = LockRequest::new(&resource, LockType::Exclusive, "test_owner").with_ttl(Duration::from_secs(30)); // This should fail because both LocalClient instances share the same global lock map // The first client (LocalClient) will acquire the lock, but the second client // (FailingMockClient's internal LocalClient) will fail to acquire the same resource let response = ns_lock.acquire_lock(&request).await?; // The operation should fail due to lock contention between the two LocalClient instances assert!( !response.success, "Lock should fail due to lock contention between LocalClient instances sharing global lock map" ); println!("βœ… Lock correctly failed due to lock contention (both clients use same global lock map)"); // Verify no locks are left behind after rollback let local_client_direct = LocalClient::new(); let lock_id = LockId::new_deterministic(&ns_lock.get_resource_key(&resource)); let lock_status = local_client_direct.check_status(&lock_id).await?; assert!(lock_status.is_none(), "No lock should remain after rollback"); println!("βœ… Verified rollback: no locks left after failed acquisition"); Ok(()) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_lock_unlock_ns_lock() -> Result<(), Box<dyn Error>> { let endpoints = get_cluster_endpoints(); let clients = create_unique_clients(&endpoints).await?; let ns_lock = NamespaceLock::with_clients("test".to_string(), clients); let resources = vec!["foo".to_string()]; let result = ns_lock .lock_batch(&resources, "dandan", Duration::from_secs(5), Duration::from_secs(10)) .await; match &result { Ok(success) => println!("Lock result: {success}"), Err(e) => println!("Lock error: {e}"), } let result = result?; assert!(result, "Lock should succeed, but got: {result}"); ns_lock.unlock_batch(&resources, "dandan").await?; Ok(()) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_concurrent_lock_attempts() -> Result<(), Box<dyn Error>> { let endpoints = get_cluster_endpoints(); let clients = create_unique_clients(&endpoints).await?; let ns_lock = NamespaceLock::with_clients("test".to_string(), clients); let resource = vec!["concurrent_resource".to_string()]; // First lock should succeed println!("Attempting first lock..."); let result1 = ns_lock .lock_batch(&resource, "owner1", Duration::from_secs(5), Duration::from_secs(10)) .await?; println!("First lock result: {result1}"); assert!(result1, "First lock should succeed"); // Second lock should fail (resource already locked) println!("Attempting second lock..."); let result2 = ns_lock .lock_batch(&resource, "owner2", Duration::from_secs(1), Duration::from_secs(10)) .await?; println!("Second lock result: {result2}"); assert!(!result2, "Second lock should fail"); // Unlock by first owner println!("Unlocking first lock..."); ns_lock.unlock_batch(&resource, "owner1").await?; println!("First lock unlocked"); // Now second owner should be able to lock println!("Attempting third lock..."); let result3 = ns_lock .lock_batch(&resource, "owner2", Duration::from_secs(5), Duration::from_secs(10)) .await?; println!("Third lock result: {result3}"); assert!(result3, "Lock should succeed after unlock"); // Clean up println!("Cleaning up..."); ns_lock.unlock_batch(&resource, "owner2").await?; println!("Test completed"); Ok(()) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_read_write_lock_compatibility() -> Result<(), Box<dyn Error>> { let endpoints = get_cluster_endpoints(); let clients = create_unique_clients(&endpoints).await?; let ns_lock = NamespaceLock::with_clients("test_rw".to_string(), clients); let resource = vec!["rw_resource".to_string()]; // First read lock should succeed let result1 = ns_lock .rlock_batch(&resource, "reader1", Duration::from_secs(5), Duration::from_secs(10)) .await?; assert!(result1, "First read lock should succeed"); // Second read lock should also succeed (read locks are compatible) let result2 = ns_lock .rlock_batch(&resource, "reader2", Duration::from_secs(5), Duration::from_secs(10)) .await?; assert!(result2, "Second read lock should succeed"); // Write lock should fail (read locks are held) let result3 = ns_lock .lock_batch(&resource, "writer1", Duration::from_secs(1), Duration::from_secs(10)) .await?; assert!(!result3, "Write lock should fail when read locks are held"); // Release read locks ns_lock.runlock_batch(&resource, "reader1").await?; ns_lock.runlock_batch(&resource, "reader2").await?; // Now write lock should succeed let result4 = ns_lock .lock_batch(&resource, "writer1", Duration::from_secs(5), Duration::from_secs(10)) .await?; assert!(result4, "Write lock should succeed after read locks released"); // Clean up ns_lock.unlock_batch(&resource, "writer1").await?; Ok(()) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_lock_timeout() -> Result<(), Box<dyn Error>> { let endpoints = get_cluster_endpoints(); let clients = create_unique_clients(&endpoints).await?; let ns_lock = NamespaceLock::with_clients("test_timeout".to_string(), clients); let resource = vec!["timeout_resource".to_string()]; // First lock with short timeout let result1 = ns_lock .lock_batch(&resource, "owner1", Duration::from_secs(2), Duration::from_secs(1)) .await?; assert!(result1, "First lock should succeed"); // Wait for lock to expire sleep(Duration::from_secs(5)).await; // Second lock should succeed after timeout let result2 = ns_lock .lock_batch(&resource, "owner2", Duration::from_secs(5), Duration::from_secs(1)) .await?; assert!(result2, "Lock should succeed after timeout"); // Clean up ns_lock.unlock_batch(&resource, "owner2").await?; Ok(()) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_batch_lock_operations() -> Result<(), Box<dyn Error>> { let endpoints = get_cluster_endpoints(); let clients = create_unique_clients(&endpoints).await?; let ns_lock = NamespaceLock::with_clients("test_batch".to_string(), clients); let resources = vec![ "batch_resource1".to_string(), "batch_resource2".to_string(), "batch_resource3".to_string(), ]; // Lock all resources let result = ns_lock .lock_batch(&resources, "batch_owner", Duration::from_secs(5), Duration::from_secs(10)) .await?; assert!(result, "Batch lock should succeed"); // Try to lock one of the resources with different owner - should fail let single_resource = vec!["batch_resource2".to_string()]; let result2 = ns_lock .lock_batch(&single_resource, "other_owner", Duration::from_secs(1), Duration::from_secs(10)) .await?; assert!(!result2, "Lock should fail for already locked resource"); // Unlock all resources ns_lock.unlock_batch(&resources, "batch_owner").await?; // Now should be able to lock single resource let result3 = ns_lock .lock_batch(&single_resource, "other_owner", Duration::from_secs(5), Duration::from_secs(10)) .await?; assert!(result3, "Lock should succeed after batch unlock"); // Clean up ns_lock.unlock_batch(&single_resource, "other_owner").await?; Ok(()) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_multiple_namespaces() -> Result<(), Box<dyn Error>> { let endpoints = get_cluster_endpoints(); let clients = create_unique_clients(&endpoints).await?; let ns_lock1 = NamespaceLock::with_clients("namespace1".to_string(), clients.clone()); let ns_lock2 = NamespaceLock::with_clients("namespace2".to_string(), clients); let resource = vec!["shared_resource".to_string()]; // Lock same resource in different namespaces - both should succeed let result1 = ns_lock1 .lock_batch(&resource, "owner1", Duration::from_secs(5), Duration::from_secs(10)) .await?; assert!(result1, "Lock in namespace1 should succeed"); let result2 = ns_lock2 .lock_batch(&resource, "owner2", Duration::from_secs(5), Duration::from_secs(10)) .await?; assert!(result2, "Lock in namespace2 should succeed"); // Clean up ns_lock1.unlock_batch(&resource, "owner1").await?; ns_lock2.unlock_batch(&resource, "owner2").await?; Ok(()) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_rpc_read_lock() -> Result<(), Box<dyn Error>> { let args = LockRequest { lock_id: LockId::new_deterministic("read_resource"), resource: "read_resource".to_string(), lock_type: LockType::Shared, owner: "reader1".to_string(), acquire_timeout: Duration::from_secs(30), ttl: Duration::from_secs(30), metadata: LockMetadata::default(), priority: LockPriority::Normal, deadlock_detection: false, }; let args_str = serde_json::to_string(&args)?; let mut client = node_service_time_out_client(&CLUSTER_ADDR.to_string()).await?; // First read lock let request = Request::new(GenerallyLockRequest { args: args_str.clone() }); let response = client.r_lock(request).await?.into_inner(); if let Some(error_info) = response.error_info { panic!("can not get read lock: {error_info}"); } // Second read lock with different owner should also succeed let args2 = LockRequest { lock_id: LockId::new_deterministic("read_resource"), resource: "read_resource".to_string(), lock_type: LockType::Shared, owner: "reader2".to_string(), acquire_timeout: Duration::from_secs(30), ttl: Duration::from_secs(30), metadata: LockMetadata::default(), priority: LockPriority::Normal, deadlock_detection: false, }; let args2_str = serde_json::to_string(&args2)?; let request2 = Request::new(GenerallyLockRequest { args: args2_str }); let response2 = client.r_lock(request2).await?.into_inner(); if let Some(error_info) = response2.error_info { panic!("can not get second read lock: {error_info}"); } // Unlock both let request = Request::new(GenerallyLockRequest { args: args_str }); let response = client.r_un_lock(request).await?.into_inner(); if let Some(error_info) = response.error_info { panic!("can not unlock read lock: {error_info}"); } Ok(()) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_lock_refresh() -> Result<(), Box<dyn Error>> { let args = LockRequest { lock_id: LockId::new_deterministic("refresh_resource"), resource: "refresh_resource".to_string(), lock_type: LockType::Exclusive, owner: "refresh_owner".to_string(), acquire_timeout: Duration::from_secs(30), ttl: Duration::from_secs(30), metadata: LockMetadata::default(), priority: LockPriority::Normal, deadlock_detection: false, }; let args_str = serde_json::to_string(&args)?; let mut client = node_service_time_out_client(&CLUSTER_ADDR.to_string()).await?; // Acquire lock let request = Request::new(GenerallyLockRequest { args: args_str.clone() }); let response = client.lock(request).await?.into_inner(); if let Some(error_info) = response.error_info { panic!("can not get lock: {error_info}"); } // Refresh lock let request = Request::new(GenerallyLockRequest { args: args_str.clone() }); let response = client.refresh(request).await?.into_inner(); if let Some(error_info) = response.error_info { panic!("can not refresh lock: {error_info}"); } assert!(response.success, "Lock refresh should succeed"); // Unlock let request = Request::new(GenerallyLockRequest { args: args_str }); let response = client.un_lock(request).await?.into_inner(); if let Some(error_info) = response.error_info { panic!("can not unlock: {error_info}"); } Ok(()) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_force_unlock() -> Result<(), Box<dyn Error>> { let args = LockRequest { lock_id: LockId::new_deterministic("force_resource"), resource: "force_resource".to_string(), lock_type: LockType::Exclusive, owner: "force_owner".to_string(), acquire_timeout: Duration::from_secs(30), ttl: Duration::from_secs(30), metadata: LockMetadata::default(), priority: LockPriority::Normal, deadlock_detection: false, }; let args_str = serde_json::to_string(&args)?; let mut client = node_service_time_out_client(&CLUSTER_ADDR.to_string()).await?; // Acquire lock let request = Request::new(GenerallyLockRequest { args: args_str.clone() }); let response = client.lock(request).await?.into_inner(); if let Some(error_info) = response.error_info { panic!("can not get lock: {error_info}"); } // Force unlock (even by different owner) let force_args = LockRequest { lock_id: LockId::new_deterministic("force_resource"), resource: "force_resource".to_string(), lock_type: LockType::Exclusive, owner: "admin".to_string(), acquire_timeout: Duration::from_secs(30), ttl: Duration::from_secs(30), metadata: LockMetadata::default(), priority: LockPriority::Normal, deadlock_detection: false, }; let force_args_str = serde_json::to_string(&force_args)?; let request = Request::new(GenerallyLockRequest { args: force_args_str }); let response = client.force_un_lock(request).await?.into_inner(); if let Some(error_info) = response.error_info { panic!("can not force unlock: {error_info}"); } assert!(response.success, "Force unlock should succeed"); Ok(()) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_global_lock_map_sharing() -> Result<(), Box<dyn Error>> { let endpoints = get_cluster_endpoints(); let clients = create_unique_clients(&endpoints).await?; let ns_lock1 = NamespaceLock::with_clients("global_test".to_string(), clients.clone()); let ns_lock2 = NamespaceLock::with_clients("global_test".to_string(), clients); let resource = vec!["global_test_resource".to_string()]; // First instance acquires lock println!("First lock map attempting to acquire lock..."); let result1 = ns_lock1 .lock_batch(&resource, "owner1", std::time::Duration::from_secs(5), std::time::Duration::from_secs(10)) .await?; println!("First lock result: {result1}"); assert!(result1, "First lock should succeed"); // Second instance should fail to acquire the same lock println!("Second lock map attempting to acquire lock..."); let result2 = ns_lock2 .lock_batch(&resource, "owner2", std::time::Duration::from_secs(1), std::time::Duration::from_secs(10)) .await?; println!("Second lock result: {result2}"); assert!(!result2, "Second lock should fail because resource is already locked"); // Release lock from first instance println!("First lock map releasing lock..."); ns_lock1.unlock_batch(&resource, "owner1").await?; // Now second instance should be able to acquire lock println!("Second lock map attempting to acquire lock again..."); let result3 = ns_lock2 .lock_batch(&resource, "owner2", std::time::Duration::from_secs(5), std::time::Duration::from_secs(10)) .await?; println!("Third lock result: {result3}"); assert!(result3, "Lock should succeed after first lock is released"); // Clean up ns_lock2.unlock_batch(&resource, "owner2").await?; Ok(()) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/reliant/lifecycle.rs
crates/e2e_test/src/reliant/lifecycle.rs
#![cfg(test)] // Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use aws_config::meta::region::RegionProviderChain; use aws_sdk_s3::Client; use aws_sdk_s3::config::{Credentials, Region}; use bytes::Bytes; use serial_test::serial; use std::error::Error; use tokio::time::sleep; const ENDPOINT: &str = "http://localhost:9000"; const ACCESS_KEY: &str = "rustfsadmin"; const SECRET_KEY: &str = "rustfsadmin"; const BUCKET: &str = "test-basic-bucket"; async fn create_aws_s3_client() -> Result<Client, Box<dyn Error>> { let region_provider = RegionProviderChain::default_provider().or_else(Region::new("us-east-1")); let shared_config = aws_config::defaults(aws_config::BehaviorVersion::latest()) .region(region_provider) .credentials_provider(Credentials::new(ACCESS_KEY, SECRET_KEY, None, None, "static")) .endpoint_url(ENDPOINT) .load() .await; let client = Client::from_conf( aws_sdk_s3::Config::from(&shared_config) .to_builder() .force_path_style(true) .build(), ); Ok(client) } async fn setup_test_bucket(client: &Client) -> Result<(), Box<dyn Error>> { match client.create_bucket().bucket(BUCKET).send().await { Ok(_) => {} Err(e) => { let error_str = e.to_string(); if !error_str.contains("BucketAlreadyOwnedByYou") && !error_str.contains("BucketAlreadyExists") { return Err(e.into()); } } } Ok(()) } #[tokio::test(flavor = "multi_thread", worker_threads = 2)] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_bucket_lifecycle_configuration() -> Result<(), Box<dyn std::error::Error>> { use aws_sdk_s3::types::{BucketLifecycleConfiguration, LifecycleExpiration, LifecycleRule, LifecycleRuleFilter}; use tokio::time::Duration; let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; // Upload test object first let test_content = "Test object for lifecycle expiration"; let lifecycle_object_key = "lifecycle-test-object.txt"; client .put_object() .bucket(BUCKET) .key(lifecycle_object_key) .body(Bytes::from(test_content.as_bytes()).into()) .send() .await?; // Verify object exists initially let resp = client.get_object().bucket(BUCKET).key(lifecycle_object_key).send().await?; assert!(resp.content_length().unwrap_or(0) > 0); // Configure lifecycle rule: expire after current time + 3 seconds let expiration = LifecycleExpiration::builder().days(0).build(); let filter = LifecycleRuleFilter::builder().prefix(lifecycle_object_key).build(); let rule = LifecycleRule::builder() .id("expire-test-object") .filter(filter) .expiration(expiration) .status(aws_sdk_s3::types::ExpirationStatus::Enabled) .build()?; let lifecycle = BucketLifecycleConfiguration::builder().rules(rule).build()?; client .put_bucket_lifecycle_configuration() .bucket(BUCKET) .lifecycle_configuration(lifecycle) .send() .await?; // Verify lifecycle configuration was set let resp = client.get_bucket_lifecycle_configuration().bucket(BUCKET).send().await?; let rules = resp.rules(); assert!(rules.iter().any(|r| r.id().unwrap_or("") == "expire-test-object")); // Wait for lifecycle processing (scanner runs every 1 second) sleep(Duration::from_secs(3)).await; // After lifecycle processing, the object should be deleted by the lifecycle rule let get_result = client.get_object().bucket(BUCKET).key(lifecycle_object_key).send().await; match get_result { Ok(_) => { panic!("Expected object to be deleted by lifecycle rule, but it still exists"); } Err(e) => { if let Some(service_error) = e.as_service_error() { if service_error.is_no_such_key() { println!("Lifecycle configuration test completed - object was successfully deleted by lifecycle rule"); } else { panic!("Expected NoSuchKey error, but got: {e:?}"); } } else { panic!("Expected service error, but got: {e:?}"); } } } println!("Lifecycle configuration test completed."); Ok(()) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/reliant/sql.rs
crates/e2e_test/src/reliant/sql.rs
#![cfg(test)] // Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use aws_config::meta::region::RegionProviderChain; use aws_sdk_s3::Client; use aws_sdk_s3::config::{Credentials, Region}; use aws_sdk_s3::types::{ CsvInput, CsvOutput, ExpressionType, FileHeaderInfo, InputSerialization, JsonInput, JsonOutput, JsonType, OutputSerialization, }; use bytes::Bytes; use serial_test::serial; use std::error::Error; const ENDPOINT: &str = "http://localhost:9000"; const ACCESS_KEY: &str = "rustfsadmin"; const SECRET_KEY: &str = "rustfsadmin"; const BUCKET: &str = "test-sql-bucket"; const CSV_OBJECT: &str = "test-data.csv"; const JSON_OBJECT: &str = "test-data.json"; async fn create_aws_s3_client() -> Result<Client, Box<dyn Error>> { let region_provider = RegionProviderChain::default_provider().or_else(Region::new("us-east-1")); let shared_config = aws_config::defaults(aws_config::BehaviorVersion::latest()) .region(region_provider) .credentials_provider(Credentials::new(ACCESS_KEY, SECRET_KEY, None, None, "static")) .endpoint_url(ENDPOINT) .load() .await; let client = Client::from_conf( aws_sdk_s3::Config::from(&shared_config) .to_builder() .force_path_style(true) // Important for S3-compatible services .build(), ); Ok(client) } async fn setup_test_bucket(client: &Client) -> Result<(), Box<dyn Error>> { match client.create_bucket().bucket(BUCKET).send().await { Ok(_) => {} Err(e) => { let error_str = e.to_string(); if !error_str.contains("BucketAlreadyOwnedByYou") && !error_str.contains("BucketAlreadyExists") { return Err(e.into()); } } } Ok(()) } async fn upload_test_csv(client: &Client) -> Result<(), Box<dyn Error>> { let csv_data = "name,age,city\nAlice,30,New York\nBob,25,Los Angeles\nCharlie,35,Chicago\nDiana,28,Boston"; client .put_object() .bucket(BUCKET) .key(CSV_OBJECT) .body(Bytes::from(csv_data.as_bytes()).into()) .send() .await?; Ok(()) } async fn upload_test_json(client: &Client) -> Result<(), Box<dyn Error>> { let json_data = r#"{"name":"Alice","age":30,"city":"New York"} {"name":"Bob","age":25,"city":"Los Angeles"} {"name":"Charlie","age":35,"city":"Chicago"} {"name":"Diana","age":28,"city":"Boston"}"#; client .put_object() .bucket(BUCKET) .key(JSON_OBJECT) .body(Bytes::from(json_data.as_bytes()).into()) .send() .await?; Ok(()) } async fn process_select_response( mut event_stream: aws_sdk_s3::operation::select_object_content::SelectObjectContentOutput, ) -> Result<String, Box<dyn Error>> { let mut total_data = Vec::new(); while let Ok(Some(event)) = event_stream.payload.recv().await { match event { aws_sdk_s3::types::SelectObjectContentEventStream::Records(records_event) => { if let Some(payload) = records_event.payload { let data = payload.into_inner(); total_data.extend_from_slice(&data); } } aws_sdk_s3::types::SelectObjectContentEventStream::End(_) => { break; } _ => { // Handle other event types (Stats, Progress, Cont, etc.) } } } Ok(String::from_utf8(total_data)?) } #[tokio::test(flavor = "multi_thread", worker_threads = 4)] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_select_object_content_csv_basic() -> Result<(), Box<dyn Error>> { let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; upload_test_csv(&client).await?; // Construct SelectObjectContent request - basic query let sql = "SELECT * FROM S3Object WHERE age > 28"; let csv_input = CsvInput::builder().file_header_info(FileHeaderInfo::Use).build(); let input_serialization = InputSerialization::builder().csv(csv_input).build(); let csv_output = CsvOutput::builder().build(); let output_serialization = OutputSerialization::builder().csv(csv_output).build(); let response = client .select_object_content() .bucket(BUCKET) .key(CSV_OBJECT) .expression(sql) .expression_type(ExpressionType::Sql) .input_serialization(input_serialization) .output_serialization(output_serialization) .send() .await?; let result_str = process_select_response(response).await?; println!("CSV Select result: {result_str}"); // Verify results contain records with age > 28 assert!(result_str.contains("Alice,30,New York")); assert!(result_str.contains("Charlie,35,Chicago")); assert!(!result_str.contains("Bob,25,Los Angeles")); assert!(!result_str.contains("Diana,28,Boston")); Ok(()) } #[tokio::test(flavor = "multi_thread", worker_threads = 4)] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_select_object_content_csv_aggregation() -> Result<(), Box<dyn Error>> { let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; upload_test_csv(&client).await?; // Construct aggregation query - use simpler approach let sql = "SELECT name, age FROM S3Object WHERE age >= 25"; let csv_input = CsvInput::builder().file_header_info(FileHeaderInfo::Use).build(); let input_serialization = InputSerialization::builder().csv(csv_input).build(); let csv_output = CsvOutput::builder().build(); let output_serialization = OutputSerialization::builder().csv(csv_output).build(); let response = client .select_object_content() .bucket(BUCKET) .key(CSV_OBJECT) .expression(sql) .expression_type(ExpressionType::Sql) .input_serialization(input_serialization) .output_serialization(output_serialization) .send() .await?; let result_str = process_select_response(response).await?; println!("CSV Aggregation result: {result_str}"); // Verify query results - should include records with age >= 25 assert!(result_str.contains("Alice")); assert!(result_str.contains("Bob")); assert!(result_str.contains("Charlie")); assert!(result_str.contains("Diana")); assert!(result_str.contains("30")); assert!(result_str.contains("25")); assert!(result_str.contains("35")); assert!(result_str.contains("28")); Ok(()) } #[tokio::test(flavor = "multi_thread", worker_threads = 4)] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_select_object_content_json_basic() -> Result<(), Box<dyn Error>> { let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; upload_test_json(&client).await?; // Construct JSON query let sql = "SELECT s.name, s.age FROM S3Object s WHERE s.age > 28"; let json_input = JsonInput::builder().set_type(Some(JsonType::Document)).build(); let input_serialization = InputSerialization::builder().json(json_input).build(); let json_output = JsonOutput::builder().build(); let output_serialization = OutputSerialization::builder().json(json_output).build(); let response = client .select_object_content() .bucket(BUCKET) .key(JSON_OBJECT) .expression(sql) .expression_type(ExpressionType::Sql) .input_serialization(input_serialization) .output_serialization(output_serialization) .send() .await?; let result_str = process_select_response(response).await?; println!("JSON Select result: {result_str}"); // Verify JSON query results assert!(result_str.contains("Alice")); assert!(result_str.contains("Charlie")); assert!(result_str.contains("30")); assert!(result_str.contains("35")); Ok(()) } #[tokio::test(flavor = "multi_thread", worker_threads = 4)] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_select_object_content_csv_limit() -> Result<(), Box<dyn Error>> { let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; upload_test_csv(&client).await?; // Test LIMIT clause let sql = "SELECT * FROM S3Object LIMIT 2"; let csv_input = CsvInput::builder().file_header_info(FileHeaderInfo::Use).build(); let input_serialization = InputSerialization::builder().csv(csv_input).build(); let csv_output = CsvOutput::builder().build(); let output_serialization = OutputSerialization::builder().csv(csv_output).build(); let response = client .select_object_content() .bucket(BUCKET) .key(CSV_OBJECT) .expression(sql) .expression_type(ExpressionType::Sql) .input_serialization(input_serialization) .output_serialization(output_serialization) .send() .await?; let result_str = process_select_response(response).await?; println!("CSV Limit result: {result_str}"); // Verify only first 2 records are returned let lines: Vec<&str> = result_str.lines().filter(|line| !line.trim().is_empty()).collect(); assert_eq!(lines.len(), 2, "Should return exactly 2 records"); Ok(()) } #[tokio::test(flavor = "multi_thread", worker_threads = 4)] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_select_object_content_csv_order_by() -> Result<(), Box<dyn Error>> { let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; upload_test_csv(&client).await?; // Test ORDER BY clause let sql = "SELECT name, age FROM S3Object ORDER BY age DESC LIMIT 2"; let csv_input = CsvInput::builder().file_header_info(FileHeaderInfo::Use).build(); let input_serialization = InputSerialization::builder().csv(csv_input).build(); let csv_output = CsvOutput::builder().build(); let output_serialization = OutputSerialization::builder().csv(csv_output).build(); let response = client .select_object_content() .bucket(BUCKET) .key(CSV_OBJECT) .expression(sql) .expression_type(ExpressionType::Sql) .input_serialization(input_serialization) .output_serialization(output_serialization) .send() .await?; let result_str = process_select_response(response).await?; println!("CSV Order By result: {result_str}"); // Verify ordered by age descending let lines: Vec<&str> = result_str.lines().filter(|line| !line.trim().is_empty()).collect(); assert!(lines.len() >= 2, "Should return at least 2 records"); // Check if contains highest age records assert!(result_str.contains("Charlie,35")); assert!(result_str.contains("Alice,30")); Ok(()) } #[tokio::test(flavor = "multi_thread", worker_threads = 4)] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_select_object_content_error_handling() -> Result<(), Box<dyn Error>> { let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; upload_test_csv(&client).await?; // Test invalid SQL query let sql = "SELECT * FROM S3Object WHERE invalid_column > 10"; let csv_input = CsvInput::builder().file_header_info(FileHeaderInfo::Use).build(); let input_serialization = InputSerialization::builder().csv(csv_input).build(); let csv_output = CsvOutput::builder().build(); let output_serialization = OutputSerialization::builder().csv(csv_output).build(); // This query should fail because invalid_column doesn't exist let result = client .select_object_content() .bucket(BUCKET) .key(CSV_OBJECT) .expression(sql) .expression_type(ExpressionType::Sql) .input_serialization(input_serialization) .output_serialization(output_serialization) .send() .await; // Verify query fails (expected behavior) assert!(result.is_err(), "Query with invalid column should fail"); Ok(()) } #[tokio::test(flavor = "multi_thread", worker_threads = 4)] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_select_object_content_nonexistent_object() -> Result<(), Box<dyn Error>> { let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; // Test query on nonexistent object let sql = "SELECT * FROM S3Object"; let csv_input = CsvInput::builder().file_header_info(FileHeaderInfo::Use).build(); let input_serialization = InputSerialization::builder().csv(csv_input).build(); let csv_output = CsvOutput::builder().build(); let output_serialization = OutputSerialization::builder().csv(csv_output).build(); let result = client .select_object_content() .bucket(BUCKET) .key("nonexistent.csv") .expression(sql) .expression_type(ExpressionType::Sql) .input_serialization(input_serialization) .output_serialization(output_serialization) .send() .await; // Verify query fails (expected behavior) assert!(result.is_err(), "Query on nonexistent object should fail"); Ok(()) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/reliant/node_interact_test.rs
crates/e2e_test/src/reliant/node_interact_test.rs
#![cfg(test)] // Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use crate::common::workspace_root; use futures::future::join_all; use rmp_serde::{Deserializer, Serializer}; use rustfs_ecstore::disk::{VolumeInfo, WalkDirOptions}; use rustfs_filemeta::{MetaCacheEntry, MetacacheReader, MetacacheWriter}; use rustfs_protos::proto_gen::node_service::WalkDirRequest; use rustfs_protos::{ models::{PingBody, PingBodyBuilder}, node_service_time_out_client, proto_gen::node_service::{ ListVolumesRequest, LocalStorageInfoRequest, MakeVolumeRequest, PingRequest, PingResponse, ReadAllRequest, }, }; use serde::{Deserialize, Serialize}; use std::error::Error; use std::io::Cursor; use std::path::PathBuf; use tokio::spawn; use tonic::Request; use tonic::codegen::tokio_stream::StreamExt; const CLUSTER_ADDR: &str = "http://localhost:9000"; #[tokio::test] #[ignore = "requires running RustFS server at localhost:9000"] async fn ping() -> Result<(), Box<dyn Error>> { let mut fbb = flatbuffers::FlatBufferBuilder::new(); let payload = fbb.create_vector(b"hello world"); let mut builder = PingBodyBuilder::new(&mut fbb); builder.add_payload(payload); let root = builder.finish(); fbb.finish(root, None); let finished_data = fbb.finished_data(); let decoded_payload = flatbuffers::root::<PingBody>(finished_data); assert!(decoded_payload.is_ok()); // Create client let mut client = node_service_time_out_client(&CLUSTER_ADDR.to_string()).await?; // Construct PingRequest let request = Request::new(PingRequest { version: 1, body: bytes::Bytes::copy_from_slice(finished_data), }); // Send request and get response let response: PingResponse = client.ping(request).await?.into_inner(); // Print response let ping_response_body = flatbuffers::root::<PingBody>(&response.body); if let Err(e) = ping_response_body { eprintln!("{e}"); } else { println!("ping_resp:body(flatbuffer): {ping_response_body:?}"); } Ok(()) } #[tokio::test] #[ignore = "requires running RustFS server at localhost:9000"] async fn make_volume() -> Result<(), Box<dyn Error>> { let mut client = node_service_time_out_client(&CLUSTER_ADDR.to_string()).await?; let request = Request::new(MakeVolumeRequest { disk: "data".to_string(), volume: "dandan".to_string(), }); let response = client.make_volume(request).await?.into_inner(); if response.success { println!("success"); } else { println!("failed: {:?}", response.error); } Ok(()) } #[tokio::test] #[ignore = "requires running RustFS server at localhost:9000"] async fn list_volumes() -> Result<(), Box<dyn Error>> { let mut client = node_service_time_out_client(&CLUSTER_ADDR.to_string()).await?; let request = Request::new(ListVolumesRequest { disk: "data".to_string(), }); let response = client.list_volumes(request).await?.into_inner(); let volume_infos: Vec<VolumeInfo> = response .volume_infos .into_iter() .filter_map(|json_str| serde_json::from_str::<VolumeInfo>(&json_str).ok()) .collect(); println!("{volume_infos:?}"); Ok(()) } #[tokio::test] #[ignore = "requires running RustFS server at localhost:9000"] async fn walk_dir() -> Result<(), Box<dyn Error>> { println!("walk_dir"); // TODO: use writer let opts = WalkDirOptions { bucket: "dandan".to_owned(), base_dir: "".to_owned(), recursive: true, ..Default::default() }; let (rd, mut wr) = tokio::io::duplex(1024); let mut buf = Vec::new(); opts.serialize(&mut Serializer::new(&mut buf))?; let mut client = node_service_time_out_client(&CLUSTER_ADDR.to_string()).await?; let disk_path = std::env::var_os("RUSTFS_DISK_PATH").map(PathBuf::from).unwrap_or_else(|| { let mut path = workspace_root(); path.push("target"); path.push(if cfg!(debug_assertions) { "debug" } else { "release" }); path.push("data"); path }); let request = Request::new(WalkDirRequest { disk: disk_path.to_string_lossy().into_owned(), walk_dir_options: buf.into(), }); let mut response = client.walk_dir(request).await?.into_inner(); let job1 = spawn(async move { let mut out = MetacacheWriter::new(&mut wr); loop { match response.next().await { Some(Ok(resp)) => { if !resp.success { println!("{}", resp.error_info.unwrap_or("".to_string())); } let entry = serde_json::from_str::<MetaCacheEntry>(&resp.meta_cache_entry) .map_err(|_e| std::io::Error::other(format!("Unexpected response: {response:?}"))) .unwrap(); out.write_obj(&entry).await.unwrap(); } None => { let _ = out.close().await; break; } _ => { println!("Unexpected response: {response:?}"); let _ = out.close().await; break; } } } }); let job2 = spawn(async move { let mut reader = MetacacheReader::new(rd); while let Ok(Some(entry)) = reader.peek().await { println!("{entry:?}"); } }); join_all(vec![job1, job2]).await; Ok(()) } #[tokio::test] #[ignore = "requires running RustFS server at localhost:9000"] async fn read_all() -> Result<(), Box<dyn Error>> { let mut client = node_service_time_out_client(&CLUSTER_ADDR.to_string()).await?; let request = Request::new(ReadAllRequest { disk: "data".to_string(), volume: "ff".to_string(), path: "format.json".to_string(), }); let response = client.read_all(request).await?.into_inner(); let volume_infos = response.data; println!("{}", response.success); println!("{volume_infos:?}"); Ok(()) } #[tokio::test] #[ignore = "requires running RustFS server at localhost:9000"] async fn storage_info() -> Result<(), Box<dyn Error>> { let mut client = node_service_time_out_client(&CLUSTER_ADDR.to_string()).await?; let request = Request::new(LocalStorageInfoRequest { metrics: true }); let response = client.local_storage_info(request).await?.into_inner(); if !response.success { println!("{:?}", response.error_info); return Ok(()); } let info = response.storage_info; let mut buf = Deserializer::new(Cursor::new(info)); let storage_info: rustfs_madmin::StorageInfo = Deserialize::deserialize(&mut buf).unwrap(); println!("{storage_info:?}"); Ok(()) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/reliant/head_deleted_object_versioning_test.rs
crates/e2e_test/src/reliant/head_deleted_object_versioning_test.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Test for HeadObject on deleted objects with versioning enabled //! //! This test reproduces the issue where getting a deleted object returns //! 200 OK instead of 404 NoSuchKey when versioning is enabled. #![cfg(test)] use aws_config::meta::region::RegionProviderChain; use aws_sdk_s3::Client; use aws_sdk_s3::config::{Credentials, Region}; use aws_sdk_s3::error::SdkError; use aws_sdk_s3::types::{BucketVersioningStatus, VersioningConfiguration}; use bytes::Bytes; use serial_test::serial; use std::error::Error; use tracing::info; const ENDPOINT: &str = "http://localhost:9000"; const ACCESS_KEY: &str = "rustfsadmin"; const SECRET_KEY: &str = "rustfsadmin"; const BUCKET: &str = "test-head-deleted-versioning-bucket"; async fn create_aws_s3_client() -> Result<Client, Box<dyn Error>> { let region_provider = RegionProviderChain::default_provider().or_else(Region::new("us-east-1")); let shared_config = aws_config::defaults(aws_config::BehaviorVersion::latest()) .region(region_provider) .credentials_provider(Credentials::new(ACCESS_KEY, SECRET_KEY, None, None, "static")) .endpoint_url(ENDPOINT) .load() .await; let client = Client::from_conf( aws_sdk_s3::Config::from(&shared_config) .to_builder() .force_path_style(true) .build(), ); Ok(client) } /// Setup test bucket, creating it if it doesn't exist, and enable versioning async fn setup_test_bucket(client: &Client) -> Result<(), Box<dyn Error>> { match client.create_bucket().bucket(BUCKET).send().await { Ok(_) => {} Err(SdkError::ServiceError(e)) => { let e = e.into_err(); let error_code = e.meta().code().unwrap_or(""); if !error_code.eq("BucketAlreadyExists") && !error_code.eq("BucketAlreadyOwnedByYou") { return Err(e.into()); } } Err(e) => { return Err(e.into()); } } // Enable versioning client .put_bucket_versioning() .bucket(BUCKET) .versioning_configuration( VersioningConfiguration::builder() .status(BucketVersioningStatus::Enabled) .build(), ) .send() .await?; Ok(()) } /// Test that HeadObject on a deleted object returns NoSuchKey when versioning is enabled #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_head_deleted_object_versioning_returns_nosuchkey() -> Result<(), Box<dyn std::error::Error>> { let _ = tracing_subscriber::fmt() .with_max_level(tracing::Level::INFO) .with_test_writer() .try_init(); info!("πŸ§ͺ Starting test_head_deleted_object_versioning_returns_nosuchkey"); let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; let key = "test-head-deleted-versioning.txt"; let content = b"Test content for HeadObject with versioning"; // Upload and verify client .put_object() .bucket(BUCKET) .key(key) .body(Bytes::from_static(content).into()) .send() .await?; // Delete the object (creates a delete marker) client.delete_object().bucket(BUCKET).key(key).send().await?; // Try to head the deleted object (latest version is delete marker) let head_result = client.head_object().bucket(BUCKET).key(key).send().await; assert!(head_result.is_err(), "HeadObject on deleted object should return an error"); match head_result.unwrap_err() { SdkError::ServiceError(service_err) => { let s3_err = service_err.into_err(); assert!( s3_err.meta().code() == Some("NoSuchKey") || s3_err.meta().code() == Some("NotFound") || s3_err.meta().code() == Some("404"), "Error should be NoSuchKey or NotFound, got: {s3_err:?}" ); info!("βœ… HeadObject correctly returns NoSuchKey/NotFound"); } other_err => { panic!("Expected ServiceError but got: {other_err:?}"); } } Ok(()) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/reliant/mod.rs
crates/e2e_test/src/reliant/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. mod conditional_writes; mod get_deleted_object_test; mod head_deleted_object_versioning_test; mod lifecycle; mod lock; mod node_interact_test; mod sql;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/reliant/get_deleted_object_test.rs
crates/e2e_test/src/reliant/get_deleted_object_test.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Test for GetObject on deleted objects //! //! This test reproduces the issue where getting a deleted object returns //! a networking error instead of NoSuchKey. #![cfg(test)] use aws_config::meta::region::RegionProviderChain; use aws_sdk_s3::Client; use aws_sdk_s3::config::{Credentials, Region}; use aws_sdk_s3::error::SdkError; use bytes::Bytes; use serial_test::serial; use std::error::Error; use tracing::info; const ENDPOINT: &str = "http://localhost:9000"; const ACCESS_KEY: &str = "rustfsadmin"; const SECRET_KEY: &str = "rustfsadmin"; const BUCKET: &str = "test-get-deleted-bucket"; async fn create_aws_s3_client() -> Result<Client, Box<dyn Error>> { let region_provider = RegionProviderChain::default_provider().or_else(Region::new("us-east-1")); let shared_config = aws_config::defaults(aws_config::BehaviorVersion::latest()) .region(region_provider) .credentials_provider(Credentials::new(ACCESS_KEY, SECRET_KEY, None, None, "static")) .endpoint_url(ENDPOINT) .load() .await; let client = Client::from_conf( aws_sdk_s3::Config::from(&shared_config) .to_builder() .force_path_style(true) .build(), ); Ok(client) } /// Setup test bucket, creating it if it doesn't exist async fn setup_test_bucket(client: &Client) -> Result<(), Box<dyn Error>> { match client.create_bucket().bucket(BUCKET).send().await { Ok(_) => {} Err(SdkError::ServiceError(e)) => { let e = e.into_err(); let error_code = e.meta().code().unwrap_or(""); if !error_code.eq("BucketAlreadyExists") && !error_code.eq("BucketAlreadyOwnedByYou") { return Err(e.into()); } } Err(e) => { return Err(e.into()); } } Ok(()) } #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_get_deleted_object_returns_nosuchkey() -> Result<(), Box<dyn std::error::Error>> { // Initialize logging let _ = tracing_subscriber::fmt() .with_max_level(tracing::Level::INFO) .with_test_writer() .try_init(); info!("πŸ§ͺ Starting test_get_deleted_object_returns_nosuchkey"); let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; // Upload a test object let key = "test-file-to-delete.txt"; let content = b"This will be deleted soon!"; info!("Uploading object: {}", key); client .put_object() .bucket(BUCKET) .key(key) .body(Bytes::from_static(content).into()) .send() .await?; // Verify object exists info!("Verifying object exists"); let get_result = client.get_object().bucket(BUCKET).key(key).send().await; assert!(get_result.is_ok(), "Object should exist after upload"); // Delete the object info!("Deleting object: {}", key); client.delete_object().bucket(BUCKET).key(key).send().await?; // Try to get the deleted object - should return NoSuchKey error info!("Attempting to get deleted object - expecting NoSuchKey error"); let get_result = client.get_object().bucket(BUCKET).key(key).send().await; // Check that we get an error assert!(get_result.is_err(), "Getting deleted object should return an error"); // Check that the error is NoSuchKey, not a networking error let err = get_result.unwrap_err(); // Print the error for debugging info!("Error received: {:?}", err); // Check if it's a service error match err { SdkError::ServiceError(service_err) => { let s3_err = service_err.into_err(); info!("Service error code: {:?}", s3_err.meta().code()); // The error should be NoSuchKey assert!(s3_err.is_no_such_key(), "Error should be NoSuchKey, got: {s3_err:?}"); info!("βœ… Test passed: GetObject on deleted object correctly returns NoSuchKey"); } other_err => { panic!("Expected ServiceError with NoSuchKey, but got: {other_err:?}"); } } // Cleanup let _ = client.delete_object().bucket(BUCKET).key(key).send().await; Ok(()) } /// Test that HeadObject on a deleted object also returns NoSuchKey #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_head_deleted_object_returns_nosuchkey() -> Result<(), Box<dyn std::error::Error>> { let _ = tracing_subscriber::fmt() .with_max_level(tracing::Level::INFO) .with_test_writer() .try_init(); info!("πŸ§ͺ Starting test_head_deleted_object_returns_nosuchkey"); let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; let key = "test-head-deleted.txt"; let content = b"Test content for HeadObject"; // Upload and verify client .put_object() .bucket(BUCKET) .key(key) .body(Bytes::from_static(content).into()) .send() .await?; // Delete the object client.delete_object().bucket(BUCKET).key(key).send().await?; // Try to head the deleted object let head_result = client.head_object().bucket(BUCKET).key(key).send().await; assert!(head_result.is_err(), "HeadObject on deleted object should return an error"); match head_result.unwrap_err() { SdkError::ServiceError(service_err) => { let s3_err = service_err.into_err(); assert!( s3_err.meta().code() == Some("NoSuchKey") || s3_err.meta().code() == Some("NotFound"), "Error should be NoSuchKey or NotFound, got: {s3_err:?}" ); info!("βœ… HeadObject correctly returns NoSuchKey/NotFound"); } other_err => { panic!("Expected ServiceError but got: {other_err:?}"); } } Ok(()) } /// Test GetObject with non-existent key (never existed) #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_get_nonexistent_object_returns_nosuchkey() -> Result<(), Box<dyn std::error::Error>> { let _ = tracing_subscriber::fmt() .with_max_level(tracing::Level::INFO) .with_test_writer() .try_init(); info!("πŸ§ͺ Starting test_get_nonexistent_object_returns_nosuchkey"); let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; // Try to get an object that never existed let key = "this-key-never-existed.txt"; let get_result = client.get_object().bucket(BUCKET).key(key).send().await; assert!(get_result.is_err(), "Getting non-existent object should return an error"); match get_result.unwrap_err() { SdkError::ServiceError(service_err) => { let s3_err = service_err.into_err(); assert!(s3_err.is_no_such_key(), "Error should be NoSuchKey, got: {s3_err:?}"); info!("βœ… GetObject correctly returns NoSuchKey for non-existent object"); } other_err => { panic!("Expected ServiceError with NoSuchKey, but got: {other_err:?}"); } } Ok(()) } /// Test multiple consecutive GetObject calls on deleted object /// This ensures the fix is stable and doesn't have race conditions #[tokio::test] #[serial] #[ignore = "requires running RustFS server at localhost:9000"] async fn test_multiple_gets_deleted_object() -> Result<(), Box<dyn std::error::Error>> { let _ = tracing_subscriber::fmt() .with_max_level(tracing::Level::INFO) .with_test_writer() .try_init(); info!("πŸ§ͺ Starting test_multiple_gets_deleted_object"); let client = create_aws_s3_client().await?; setup_test_bucket(&client).await?; let key = "test-multiple-gets.txt"; let content = b"Test content"; // Upload and delete client .put_object() .bucket(BUCKET) .key(key) .body(Bytes::from_static(content).into()) .send() .await?; client.delete_object().bucket(BUCKET).key(key).send().await?; // Try multiple consecutive GetObject calls for i in 1..=5 { info!("Attempt {} to get deleted object", i); let get_result = client.get_object().bucket(BUCKET).key(key).send().await; assert!(get_result.is_err(), "Attempt {i}: should return error"); match get_result.unwrap_err() { SdkError::ServiceError(service_err) => { let s3_err = service_err.into_err(); assert!(s3_err.is_no_such_key(), "Attempt {i}: Error should be NoSuchKey, got: {s3_err:?}"); } other_err => { panic!("Attempt {i}: Expected ServiceError but got: {other_err:?}"); } } } info!("βœ… All 5 attempts correctly returned NoSuchKey"); Ok(()) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/policy/test_env.rs
crates/e2e_test/src/policy/test_env.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Custom test environment for policy variables tests //! //! This module provides a custom test environment that doesn't automatically //! stop servers when destroyed, addressing the server stopping issue. use aws_sdk_s3::Client; use aws_sdk_s3::config::{Config, Credentials, Region}; use std::net::TcpStream; use std::time::Duration; use tokio::time::sleep; use tracing::{info, warn}; // Default credentials const DEFAULT_ACCESS_KEY: &str = "rustfsadmin"; const DEFAULT_SECRET_KEY: &str = "rustfsadmin"; /// Custom test environment that doesn't automatically stop servers pub struct PolicyTestEnvironment { pub temp_dir: String, pub address: String, pub url: String, pub access_key: String, pub secret_key: String, } impl PolicyTestEnvironment { /// Create a new test environment with specific address /// This environment won't stop any server when dropped pub async fn with_address(address: &str) -> Result<Self, Box<dyn std::error::Error + Send + Sync>> { let temp_dir = format!("/tmp/rustfs_policy_test_{}", uuid::Uuid::new_v4()); tokio::fs::create_dir_all(&temp_dir).await?; let url = format!("http://{address}"); Ok(Self { temp_dir, address: address.to_string(), url, access_key: DEFAULT_ACCESS_KEY.to_string(), secret_key: DEFAULT_SECRET_KEY.to_string(), }) } /// Create an AWS S3 client configured for this RustFS instance pub fn create_s3_client(&self, access_key: &str, secret_key: &str) -> Client { let credentials = Credentials::new(access_key, secret_key, None, None, "policy-test"); let config = Config::builder() .credentials_provider(credentials) .region(Region::new("us-east-1")) .endpoint_url(&self.url) .force_path_style(true) .behavior_version_latest() .build(); Client::from_conf(config) } /// Wait for RustFS server to be ready by checking TCP connectivity pub async fn wait_for_server_ready(&self) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { info!("Waiting for RustFS server to be ready on {}", self.address); for i in 0..30 { if TcpStream::connect(&self.address).is_ok() { info!("βœ… RustFS server is ready after {} attempts", i + 1); return Ok(()); } if i == 29 { return Err("RustFS server failed to become ready within 30 seconds".into()); } sleep(Duration::from_secs(1)).await; } Ok(()) } } // Implement Drop trait that doesn't stop servers impl Drop for PolicyTestEnvironment { fn drop(&mut self) { // Clean up temp directory only, don't stop any server if let Err(e) = std::fs::remove_dir_all(&self.temp_dir) { warn!("Failed to clean up temp directory {}: {}", self.temp_dir, e); } } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/policy/policy_variables_test.rs
crates/e2e_test/src/policy/policy_variables_test.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Tests for AWS IAM policy variables with single-value, multi-value, and nested scenarios use crate::common::{awscurl_put, init_logging}; use crate::policy::test_env::PolicyTestEnvironment; use aws_sdk_s3::primitives::ByteStream; use serial_test::serial; use tracing::info; /// Helper function to create a regular user with given credentials async fn create_user( env: &PolicyTestEnvironment, username: &str, password: &str, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { let create_user_body = serde_json::json!({ "secretKey": password, "status": "enabled" }) .to_string(); let create_user_url = format!("{}/rustfs/admin/v3/add-user?accessKey={}", env.url, username); awscurl_put(&create_user_url, &create_user_body, &env.access_key, &env.secret_key).await?; Ok(()) } /// Helper function to create an STS user with given credentials async fn create_sts_user( env: &PolicyTestEnvironment, username: &str, password: &str, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { // For STS, we create a regular user first, then use it to assume roles create_user(env, username, password).await?; Ok(()) } /// Helper function to create and attach a policy async fn create_and_attach_policy( env: &PolicyTestEnvironment, policy_name: &str, username: &str, policy_document: serde_json::Value, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { let policy_string = policy_document.to_string(); // Create policy let add_policy_url = format!("{}/rustfs/admin/v3/add-canned-policy?name={}", env.url, policy_name); awscurl_put(&add_policy_url, &policy_string, &env.access_key, &env.secret_key).await?; // Attach policy to user let attach_policy_url = format!( "{}/rustfs/admin/v3/set-user-or-group-policy?policyName={}&userOrGroup={}&isGroup=false", env.url, policy_name, username ); awscurl_put(&attach_policy_url, "", &env.access_key, &env.secret_key).await?; Ok(()) } /// Helper function to clean up test resources async fn cleanup_user_and_policy(env: &PolicyTestEnvironment, username: &str, policy_name: &str) { // Create admin client for cleanup let admin_client = env.create_s3_client(&env.access_key, &env.secret_key); // Delete buckets that might have been created by this user let bucket_patterns = [ format!("{username}-test-bucket"), format!("{username}-bucket1"), format!("{username}-bucket2"), format!("{username}-bucket3"), format!("prefix-{username}-suffix"), format!("{username}-test"), format!("{username}-sts-bucket"), format!("{username}-service-bucket"), "private-test-bucket".to_string(), // For deny test ]; // Try to delete objects and buckets for bucket_name in &bucket_patterns { let _ = admin_client .delete_object() .bucket(bucket_name) .key("test-object.txt") .send() .await; let _ = admin_client .delete_object() .bucket(bucket_name) .key("test-sts-object.txt") .send() .await; let _ = admin_client .delete_object() .bucket(bucket_name) .key("test-service-object.txt") .send() .await; let _ = admin_client.delete_bucket().bucket(bucket_name).send().await; } // Remove user let remove_user_url = format!("{}/rustfs/admin/v3/remove-user?accessKey={}", env.url, username); let _ = awscurl_put(&remove_user_url, "", &env.access_key, &env.secret_key).await; // Remove policy let remove_policy_url = format!("{}/rustfs/admin/v3/remove-canned-policy?name={}", env.url, policy_name); let _ = awscurl_put(&remove_policy_url, "", &env.access_key, &env.secret_key).await; } /// Test AWS policy variables with single-value scenarios #[tokio::test(flavor = "multi_thread")] #[serial] #[ignore = "Starts a rustfs server; enable when running full E2E"] pub async fn test_aws_policy_variables_single_value() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { test_aws_policy_variables_single_value_impl().await } /// Implementation function for single-value policy variables test pub async fn test_aws_policy_variables_single_value_impl() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { init_logging(); info!("Starting AWS policy variables single-value test"); let env = PolicyTestEnvironment::with_address("127.0.0.1:9000").await?; test_aws_policy_variables_single_value_impl_with_env(&env).await } /// Implementation function for single-value policy variables test with shared environment pub async fn test_aws_policy_variables_single_value_impl_with_env( env: &PolicyTestEnvironment, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { // Create test user let test_user = "testuser1"; let test_password = "testpassword123"; let policy_name = "test-single-value-policy"; // Create cleanup function let cleanup = || async { cleanup_user_and_policy(env, test_user, policy_name).await; }; let create_user_body = serde_json::json!({ "secretKey": test_password, "status": "enabled" }) .to_string(); let create_user_url = format!("{}/rustfs/admin/v3/add-user?accessKey={}", env.url, test_user); awscurl_put(&create_user_url, &create_user_body, &env.access_key, &env.secret_key).await?; // Create policy with single-value AWS variables let policy_document = serde_json::json!({ "Version": "2012-10-17", "Statement": [ { "Effect": "Allow", "Action": ["s3:ListAllMyBuckets"], "Resource": ["arn:aws:s3:::*"] }, { "Effect": "Allow", "Action": ["s3:CreateBucket"], "Resource": [format!("arn:aws:s3:::{}-*", "${aws:username}")] }, { "Effect": "Allow", "Action": ["s3:ListBucket"], "Resource": [format!("arn:aws:s3:::{}-*", "${aws:username}")] }, { "Effect": "Allow", "Action": ["s3:PutObject", "s3:GetObject"], "Resource": [format!("arn:aws:s3:::{}-*/*", "${aws:username}")] } ] }) .to_string(); let add_policy_url = format!("{}/rustfs/admin/v3/add-canned-policy?name={}", env.url, policy_name); awscurl_put(&add_policy_url, &policy_document, &env.access_key, &env.secret_key).await?; // Attach policy to user let attach_policy_url = format!( "{}/rustfs/admin/v3/set-user-or-group-policy?policyName={}&userOrGroup={}&isGroup=false", env.url, policy_name, test_user ); awscurl_put(&attach_policy_url, "", &env.access_key, &env.secret_key).await?; // Create S3 client for test user let test_client = env.create_s3_client(test_user, test_password); tokio::time::sleep(std::time::Duration::from_millis(500)).await; // Test 1: User should be able to list buckets (allowed by policy) info!("Test 1: User listing buckets"); let list_result = test_client.list_buckets().send().await; if let Err(e) = list_result { cleanup().await; return Err(format!("User should be able to list buckets: {e}").into()); } // Test 2: User should be able to create bucket matching username pattern info!("Test 2: User creating bucket matching pattern"); let bucket_name = format!("{test_user}-test-bucket"); let create_result = test_client.create_bucket().bucket(&bucket_name).send().await; if let Err(e) = create_result { cleanup().await; return Err(format!("User should be able to create bucket matching username pattern: {e}").into()); } // Test 3: User should be able to list objects in their own bucket info!("Test 3: User listing objects in their bucket"); let list_objects_result = test_client.list_objects_v2().bucket(&bucket_name).send().await; if let Err(e) = list_objects_result { cleanup().await; return Err(format!("User should be able to list objects in their own bucket: {e}").into()); } // Test 4: User should be able to put object in their own bucket info!("Test 4: User putting object in their bucket"); let put_result = test_client .put_object() .bucket(&bucket_name) .key("test-object.txt") .body(ByteStream::from_static(b"Hello, Policy Variables!")) .send() .await; if let Err(e) = put_result { cleanup().await; return Err(format!("User should be able to put object in their own bucket: {e}").into()); } // Test 5: User should be able to get object from their own bucket info!("Test 5: User getting object from their bucket"); let get_result = test_client .get_object() .bucket(&bucket_name) .key("test-object.txt") .send() .await; if let Err(e) = get_result { cleanup().await; return Err(format!("User should be able to get object from their own bucket: {e}").into()); } // Test 6: User should NOT be able to create bucket NOT matching username pattern info!("Test 6: User attempting to create bucket NOT matching pattern"); let other_bucket_name = "other-user-bucket"; let create_other_result = test_client.create_bucket().bucket(other_bucket_name).send().await; if create_other_result.is_ok() { cleanup().await; return Err("User should NOT be able to create bucket NOT matching username pattern".into()); } // Cleanup info!("Cleaning up test resources"); cleanup().await; info!("AWS policy variables single-value test completed successfully"); Ok(()) } /// Test AWS policy variables with multi-value scenarios #[tokio::test(flavor = "multi_thread")] #[serial] #[ignore = "Starts a rustfs server; enable when running full E2E"] pub async fn test_aws_policy_variables_multi_value() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { test_aws_policy_variables_multi_value_impl().await } /// Implementation function for multi-value policy variables test pub async fn test_aws_policy_variables_multi_value_impl() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { init_logging(); info!("Starting AWS policy variables multi-value test"); let env = PolicyTestEnvironment::with_address("127.0.0.1:9000").await?; test_aws_policy_variables_multi_value_impl_with_env(&env).await } /// Implementation function for multi-value policy variables test with shared environment pub async fn test_aws_policy_variables_multi_value_impl_with_env( env: &PolicyTestEnvironment, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { // Create test user let test_user = "testuser2"; let test_password = "testpassword123"; let policy_name = "test-multi-value-policy"; // Create cleanup function let cleanup = || async { cleanup_user_and_policy(env, test_user, policy_name).await; }; // Create user create_user(env, test_user, test_password).await?; // Create policy with multi-value AWS variables let policy_document = serde_json::json!({ "Version": "2012-10-17", "Statement": [ { "Effect": "Allow", "Action": ["s3:ListAllMyBuckets"], "Resource": ["arn:aws:s3:::*"] }, { "Effect": "Allow", "Action": ["s3:CreateBucket"], "Resource": [ format!("arn:aws:s3:::{}-bucket1", "${aws:username}"), format!("arn:aws:s3:::{}-bucket2", "${aws:username}"), format!("arn:aws:s3:::{}-bucket3", "${aws:username}") ] }, { "Effect": "Allow", "Action": ["s3:ListBucket"], "Resource": [ format!("arn:aws:s3:::{}-bucket1", "${aws:username}"), format!("arn:aws:s3:::{}-bucket2", "${aws:username}"), format!("arn:aws:s3:::{}-bucket3", "${aws:username}") ] } ] }); create_and_attach_policy(env, policy_name, test_user, policy_document).await?; // Create S3 client for test user let test_client = env.create_s3_client(test_user, test_password); // Test 1: User should be able to create buckets matching any of the multi-value patterns info!("Test 1: User creating first bucket matching multi-value pattern"); let bucket1_name = format!("{test_user}-bucket1"); let create_result1 = test_client.create_bucket().bucket(&bucket1_name).send().await; if let Err(e) = create_result1 { cleanup().await; return Err(format!("User should be able to create first bucket matching multi-value pattern: {e}").into()); } info!("Test 2: User creating second bucket matching multi-value pattern"); let bucket2_name = format!("{test_user}-bucket2"); let create_result2 = test_client.create_bucket().bucket(&bucket2_name).send().await; if let Err(e) = create_result2 { cleanup().await; return Err(format!("User should be able to create second bucket matching multi-value pattern: {e}").into()); } info!("Test 3: User creating third bucket matching multi-value pattern"); let bucket3_name = format!("{test_user}-bucket3"); let create_result3 = test_client.create_bucket().bucket(&bucket3_name).send().await; if let Err(e) = create_result3 { cleanup().await; return Err(format!("User should be able to create third bucket matching multi-value pattern: {e}").into()); } // Test 4: User should NOT be able to create bucket NOT matching any multi-value pattern info!("Test 4: User attempting to create bucket NOT matching any pattern"); let other_bucket_name = format!("{test_user}-other-bucket"); let create_other_result = test_client.create_bucket().bucket(&other_bucket_name).send().await; if create_other_result.is_ok() { cleanup().await; return Err("User should NOT be able to create bucket NOT matching any multi-value pattern".into()); } // Test 5: User should be able to list objects in their allowed buckets info!("Test 5: User listing objects in allowed buckets"); let list_objects_result1 = test_client.list_objects_v2().bucket(&bucket1_name).send().await; if let Err(e) = list_objects_result1 { cleanup().await; return Err(format!("User should be able to list objects in first allowed bucket: {e}").into()); } let list_objects_result2 = test_client.list_objects_v2().bucket(&bucket2_name).send().await; if let Err(e) = list_objects_result2 { cleanup().await; return Err(format!("User should be able to list objects in second allowed bucket: {e}").into()); } // Cleanup info!("Cleaning up test resources"); cleanup().await; info!("AWS policy variables multi-value test completed successfully"); Ok(()) } /// Test AWS policy variables with variable concatenation #[tokio::test(flavor = "multi_thread")] #[serial] #[ignore = "Starts a rustfs server; enable when running full E2E"] pub async fn test_aws_policy_variables_concatenation() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { test_aws_policy_variables_concatenation_impl().await } /// Implementation function for concatenation policy variables test pub async fn test_aws_policy_variables_concatenation_impl() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { init_logging(); info!("Starting AWS policy variables concatenation test"); let env = PolicyTestEnvironment::with_address("127.0.0.1:9000").await?; test_aws_policy_variables_concatenation_impl_with_env(&env).await } /// Implementation function for concatenation policy variables test with shared environment pub async fn test_aws_policy_variables_concatenation_impl_with_env( env: &PolicyTestEnvironment, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { // Create test user let test_user = "testuser3"; let test_password = "testpassword123"; let policy_name = "test-concatenation-policy"; // Create cleanup function let cleanup = || async { cleanup_user_and_policy(env, test_user, policy_name).await; }; // Create user create_user(env, test_user, test_password).await?; // Create policy with variable concatenation let policy_document = serde_json::json!({ "Version": "2012-10-17", "Statement": [ { "Effect": "Allow", "Action": ["s3:ListAllMyBuckets"], "Resource": ["arn:aws:s3:::*"] }, { "Effect": "Allow", "Action": ["s3:CreateBucket"], "Resource": [format!("arn:aws:s3:::prefix-{}-suffix", "${aws:username}")] }, { "Effect": "Allow", "Action": ["s3:ListBucket"], "Resource": [format!("arn:aws:s3:::prefix-{}-suffix", "${aws:username}")] } ] }); create_and_attach_policy(env, policy_name, test_user, policy_document).await?; // Create S3 client for test user let test_client = env.create_s3_client(test_user, test_password); // Add a small delay to allow policy to propagate tokio::time::sleep(std::time::Duration::from_millis(500)).await; // Test: User should be able to create bucket matching concatenated pattern info!("Test: User creating bucket matching concatenated pattern"); let bucket_name = format!("prefix-{test_user}-suffix"); let create_result = test_client.create_bucket().bucket(&bucket_name).send().await; if let Err(e) = create_result { cleanup().await; return Err(format!("User should be able to create bucket matching concatenated pattern: {e}").into()); } // Test: User should be able to list objects in the concatenated pattern bucket info!("Test: User listing objects in concatenated pattern bucket"); let list_objects_result = test_client.list_objects_v2().bucket(&bucket_name).send().await; if let Err(e) = list_objects_result { cleanup().await; return Err(format!("User should be able to list objects in concatenated pattern bucket: {e}").into()); } // Cleanup info!("Cleaning up test resources"); cleanup().await; info!("AWS policy variables concatenation test completed successfully"); Ok(()) } /// Test AWS policy variables with nested scenarios #[tokio::test(flavor = "multi_thread")] #[serial] #[ignore = "Starts a rustfs server; enable when running full E2E"] pub async fn test_aws_policy_variables_nested() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { test_aws_policy_variables_nested_impl().await } /// Implementation function for nested policy variables test pub async fn test_aws_policy_variables_nested_impl() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { init_logging(); info!("Starting AWS policy variables nested test"); let env = PolicyTestEnvironment::with_address("127.0.0.1:9000").await?; test_aws_policy_variables_nested_impl_with_env(&env).await } /// Test AWS policy variables with STS temporary credentials #[tokio::test(flavor = "multi_thread")] #[serial] #[ignore = "Starts a rustfs server; enable when running full E2E"] pub async fn test_aws_policy_variables_sts() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { test_aws_policy_variables_sts_impl().await } /// Implementation function for STS policy variables test pub async fn test_aws_policy_variables_sts_impl() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { init_logging(); info!("Starting AWS policy variables STS test"); let env = PolicyTestEnvironment::with_address("127.0.0.1:9000").await?; test_aws_policy_variables_sts_impl_with_env(&env).await } /// Implementation function for nested policy variables test with shared environment pub async fn test_aws_policy_variables_nested_impl_with_env( env: &PolicyTestEnvironment, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { // Create test user let test_user = "testuser4"; let test_password = "testpassword123"; let policy_name = "test-nested-policy"; // Create cleanup function let cleanup = || async { cleanup_user_and_policy(env, test_user, policy_name).await; }; // Create user create_user(env, test_user, test_password).await?; // Create policy with nested variables - this tests complex variable resolution let policy_document = serde_json::json!({ "Version": "2012-10-17", "Statement": [ { "Effect": "Allow", "Action": ["s3:ListAllMyBuckets"], "Resource": ["arn:aws:s3:::*"] }, { "Effect": "Allow", "Action": ["s3:CreateBucket"], "Resource": ["arn:aws:s3:::${${aws:username}-test}"] }, { "Effect": "Allow", "Action": ["s3:ListBucket"], "Resource": ["arn:aws:s3:::${${aws:username}-test}"] } ] }); create_and_attach_policy(env, policy_name, test_user, policy_document).await?; // Create S3 client for test user let test_client = env.create_s3_client(test_user, test_password); // Add a small delay to allow policy to propagate tokio::time::sleep(std::time::Duration::from_millis(500)).await; // Test nested variable resolution info!("Test: Nested variable resolution"); // Create bucket with expected resolved name let expected_bucket = format!("{test_user}-test"); // Attempt to create bucket with resolved name let create_result = test_client.create_bucket().bucket(&expected_bucket).send().await; // Verify bucket creation succeeds (nested variable resolved correctly) if let Err(e) = create_result { cleanup().await; return Err(format!("User should be able to create bucket with nested variable: {e}").into()); } // Verify bucket creation fails with unresolved variable let unresolved_bucket = format!("${{}}-test {test_user}"); let create_unresolved = test_client.create_bucket().bucket(&unresolved_bucket).send().await; if create_unresolved.is_ok() { cleanup().await; return Err("User should NOT be able to create bucket with unresolved variable".into()); } // Cleanup info!("Cleaning up test resources"); cleanup().await; info!("AWS policy variables nested test completed successfully"); Ok(()) } /// Implementation function for STS policy variables test with shared environment pub async fn test_aws_policy_variables_sts_impl_with_env( env: &PolicyTestEnvironment, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { // Create test user for STS let test_user = "testuser-sts"; let test_password = "testpassword123"; let policy_name = "test-sts-policy"; // Create cleanup function let cleanup = || async { cleanup_user_and_policy(env, test_user, policy_name).await; }; // Create STS user create_sts_user(env, test_user, test_password).await?; // Create policy with STS-compatible variables let policy_document = serde_json::json!({ "Version": "2012-10-17", "Statement": [ { "Effect": "Allow", "Action": ["s3:ListAllMyBuckets"], "Resource": ["arn:aws:s3:::*"] }, { "Effect": "Allow", "Action": ["s3:CreateBucket"], "Resource": [format!("arn:aws:s3:::{}-sts-bucket", "${aws:username}")] }, { "Effect": "Allow", "Action": ["s3:ListBucket", "s3:PutObject", "s3:GetObject"], "Resource": [format!("arn:aws:s3:::{}-sts-bucket/*", "${aws:username}")] } ] }); create_and_attach_policy(env, policy_name, test_user, policy_document).await?; // Create S3 client for test user let test_client = env.create_s3_client(test_user, test_password); // Add a small delay to allow policy to propagate tokio::time::sleep(std::time::Duration::from_millis(500)).await; // Test: User should be able to create bucket matching STS pattern info!("Test: User creating bucket matching STS pattern"); let bucket_name = format!("{test_user}-sts-bucket"); let create_result = test_client.create_bucket().bucket(&bucket_name).send().await; if let Err(e) = create_result { cleanup().await; return Err(format!("User should be able to create STS bucket: {e}").into()); } // Test: User should be able to put object in STS bucket info!("Test: User putting object in STS bucket"); let put_result = test_client .put_object() .bucket(&bucket_name) .key("test-sts-object.txt") .body(ByteStream::from_static(b"STS Test Object")) .send() .await; if let Err(e) = put_result { cleanup().await; return Err(format!("User should be able to put object in STS bucket: {e}").into()); } // Test: User should be able to get object from STS bucket info!("Test: User getting object from STS bucket"); let get_result = test_client .get_object() .bucket(&bucket_name) .key("test-sts-object.txt") .send() .await; if let Err(e) = get_result { cleanup().await; return Err(format!("User should be able to get object from STS bucket: {e}").into()); } // Test: User should be able to list objects in STS bucket info!("Test: User listing objects in STS bucket"); let list_result = test_client.list_objects_v2().bucket(&bucket_name).send().await; if let Err(e) = list_result { cleanup().await; return Err(format!("User should be able to list objects in STS bucket: {e}").into()); } // Cleanup info!("Cleaning up test resources"); cleanup().await; info!("AWS policy variables STS test completed successfully"); Ok(()) } /// Test AWS policy variables with deny scenarios #[tokio::test(flavor = "multi_thread")] #[serial] #[ignore = "Starts a rustfs server; enable when running full E2E"] pub async fn test_aws_policy_variables_deny() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { test_aws_policy_variables_deny_impl().await } /// Implementation function for deny policy variables test pub async fn test_aws_policy_variables_deny_impl() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { init_logging(); info!("Starting AWS policy variables deny test"); let env = PolicyTestEnvironment::with_address("127.0.0.1:9000").await?; test_aws_policy_variables_deny_impl_with_env(&env).await } /// Implementation function for deny policy variables test with shared environment pub async fn test_aws_policy_variables_deny_impl_with_env( env: &PolicyTestEnvironment, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { // Create test user let test_user = "testuser5"; let test_password = "testpassword123"; let policy_name = "test-deny-policy"; // Create cleanup function let cleanup = || async { cleanup_user_and_policy(env, test_user, policy_name).await; }; // Create user create_user(env, test_user, test_password).await?; // Create policy with both allow and deny statements let policy_document = serde_json::json!({ "Version": "2012-10-17", "Statement": [ // Allow general access { "Effect": "Allow", "Action": ["s3:ListAllMyBuckets"], "Resource": ["arn:aws:s3:::*"] }, // Allow creating buckets matching username pattern { "Effect": "Allow", "Action": ["s3:CreateBucket"], "Resource": [format!("arn:aws:s3:::{}-*", "${aws:username}")] }, // Deny creating buckets with "private" in the name { "Effect": "Deny", "Action": ["s3:CreateBucket"], "Resource": ["arn:aws:s3:::*private*"] } ] }); create_and_attach_policy(env, policy_name, test_user, policy_document).await?; // Create S3 client for test user let test_client = env.create_s3_client(test_user, test_password); // Add a small delay to allow policy to propagate tokio::time::sleep(std::time::Duration::from_millis(500)).await; // Test 1: User should be able to create bucket matching username pattern info!("Test 1: User creating bucket matching username pattern"); let bucket_name = format!("{test_user}-test-bucket"); let create_result = test_client.create_bucket().bucket(&bucket_name).send().await; if let Err(e) = create_result { cleanup().await; return Err(format!("User should be able to create bucket matching username pattern: {e}").into()); } // Test 2: User should NOT be able to create bucket with "private" in the name (deny rule) info!("Test 2: User attempting to create bucket with 'private' in name (should be denied)"); let private_bucket_name = "private-test-bucket"; let create_private_result = test_client.create_bucket().bucket(private_bucket_name).send().await; if create_private_result.is_ok() { cleanup().await; return Err("User should NOT be able to create bucket with 'private' in name due to deny rule".into()); } // Cleanup info!("Cleaning up test resources"); cleanup().await; info!("AWS policy variables deny test completed successfully"); Ok(()) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/policy/mod.rs
crates/e2e_test/src/policy/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Policy-specific tests for RustFS //! //! This module provides comprehensive tests for AWS IAM policy variables //! including single-value, multi-value, and nested variable scenarios. mod policy_variables_test; mod test_env; mod test_runner;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/e2e_test/src/policy/test_runner.rs
crates/e2e_test/src/policy/test_runner.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use crate::common::init_logging; use crate::policy::test_env::PolicyTestEnvironment; use serial_test::serial; use std::time::Instant; use tokio::time::{Duration, sleep}; use tracing::{error, info}; /// Core test categories #[derive(Debug, Clone, PartialEq, Eq)] pub enum TestCategory { SingleValue, MultiValue, Concatenation, Nested, DenyScenarios, } impl TestCategory {} /// Test case definition #[derive(Debug, Clone)] pub struct TestDefinition { pub name: String, #[allow(dead_code)] pub category: TestCategory, pub is_critical: bool, } impl TestDefinition { pub fn new(name: impl Into<String>, category: TestCategory, is_critical: bool) -> Self { Self { name: name.into(), category, is_critical, } } } /// Test result #[derive(Debug, Clone)] pub struct TestResult { pub test_name: String, pub success: bool, pub error_message: Option<String>, } impl TestResult { pub fn success(test_name: String) -> Self { Self { test_name, success: true, error_message: None, } } pub fn failure(test_name: String, error: String) -> Self { Self { test_name, success: false, error_message: Some(error), } } } /// Test suite configuration #[derive(Debug, Clone, Default)] pub struct TestSuiteConfig { pub include_critical_only: bool, } /// Policy test suite pub struct PolicyTestSuite { tests: Vec<TestDefinition>, config: TestSuiteConfig, } impl PolicyTestSuite { /// Create default test suite pub fn new() -> Self { let tests = vec![ TestDefinition::new("test_aws_policy_variables_single_value", TestCategory::SingleValue, true), TestDefinition::new("test_aws_policy_variables_multi_value", TestCategory::MultiValue, true), TestDefinition::new("test_aws_policy_variables_concatenation", TestCategory::Concatenation, true), TestDefinition::new("test_aws_policy_variables_nested", TestCategory::Nested, true), TestDefinition::new("test_aws_policy_variables_deny", TestCategory::DenyScenarios, true), TestDefinition::new("test_aws_policy_variables_sts", TestCategory::SingleValue, true), ]; Self { tests, config: TestSuiteConfig::default(), } } /// Configure test suite pub fn with_config(mut self, config: TestSuiteConfig) -> Self { self.config = config; self } /// Run test suite pub async fn run_test_suite(&self) -> Vec<TestResult> { init_logging(); info!("Starting Policy Variables test suite"); let start_time = Instant::now(); let mut results = Vec::new(); // Create test environment let env = match PolicyTestEnvironment::with_address("127.0.0.1:9000").await { Ok(env) => env, Err(e) => { error!("Failed to create test environment: {}", e); return vec![TestResult::failure("env_creation".into(), e.to_string())]; } }; // Wait for server to be ready if env.wait_for_server_ready().await.is_err() { error!("Server is not ready"); return vec![TestResult::failure("server_check".into(), "Server not ready".into())]; } // Filter tests let tests_to_run: Vec<&TestDefinition> = self .tests .iter() .filter(|test| !self.config.include_critical_only || test.is_critical) .collect(); info!("Scheduled {} tests", tests_to_run.len()); // Run tests for (i, test_def) in tests_to_run.iter().enumerate() { info!("Running test {}/{}: {}", i + 1, tests_to_run.len(), test_def.name); let test_start = Instant::now(); let result = self.run_single_test(test_def, &env).await; let test_duration = test_start.elapsed(); match result { Ok(_) => { info!("Test passed: {} ({:.2}s)", test_def.name, test_duration.as_secs_f64()); results.push(TestResult::success(test_def.name.clone())); } Err(e) => { error!("Test failed: {} ({:.2}s): {}", test_def.name, test_duration.as_secs_f64(), e); results.push(TestResult::failure(test_def.name.clone(), e.to_string())); } } // Delay between tests to avoid resource conflicts if i < tests_to_run.len() - 1 { sleep(Duration::from_secs(2)).await; } } // Print summary self.print_summary(&results, start_time.elapsed()); results } /// Run a single test async fn run_single_test( &self, test_def: &TestDefinition, env: &PolicyTestEnvironment, ) -> Result<(), Box<dyn std::error::Error + Send + Sync>> { match test_def.name.as_str() { "test_aws_policy_variables_single_value" => { super::policy_variables_test::test_aws_policy_variables_single_value_impl_with_env(env).await } "test_aws_policy_variables_multi_value" => { super::policy_variables_test::test_aws_policy_variables_multi_value_impl_with_env(env).await } "test_aws_policy_variables_concatenation" => { super::policy_variables_test::test_aws_policy_variables_concatenation_impl_with_env(env).await } "test_aws_policy_variables_nested" => { super::policy_variables_test::test_aws_policy_variables_nested_impl_with_env(env).await } "test_aws_policy_variables_deny" => { super::policy_variables_test::test_aws_policy_variables_deny_impl_with_env(env).await } "test_aws_policy_variables_sts" => { super::policy_variables_test::test_aws_policy_variables_sts_impl_with_env(env).await } _ => Err(format!("Test {} not implemented", test_def.name).into()), } } /// Print test summary fn print_summary(&self, results: &[TestResult], total_duration: Duration) { info!("=== Test Suite Summary ==="); info!("Total duration: {:.2}s", total_duration.as_secs_f64()); info!("Total tests: {}", results.len()); let passed = results.iter().filter(|r| r.success).count(); let failed = results.len() - passed; let success_rate = (passed as f64 / results.len() as f64) * 100.0; info!("Passed: {} | Failed: {}", passed, failed); info!("Success rate: {:.1}%", success_rate); if failed > 0 { error!("Failed tests:"); for result in results.iter().filter(|r| !r.success) { error!(" - {}: {}", result.test_name, result.error_message.as_ref().unwrap()); } } } } /// Test suite #[tokio::test] #[serial] #[ignore = "Connects to existing rustfs server"] async fn test_policy_critical_suite() -> Result<(), Box<dyn std::error::Error + Send + Sync>> { let config = TestSuiteConfig { include_critical_only: true, }; let suite = PolicyTestSuite::new().with_config(config); let results = suite.run_test_suite().await; let failed = results.iter().filter(|r| !r.success).count(); if failed > 0 { return Err(format!("Critical tests failed: {failed} failures").into()); } info!("All critical tests passed"); Ok(()) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/iam/src/lib.rs
crates/iam/src/lib.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use crate::error::{Error, Result}; use manager::IamCache; use rustfs_ecstore::store::ECStore; use std::sync::{Arc, OnceLock}; use store::object::ObjectStore; use sys::IamSys; use tracing::{error, info, instrument}; pub mod cache; pub mod error; pub mod manager; pub mod store; pub mod sys; pub mod utils; static IAM_SYS: OnceLock<Arc<IamSys<ObjectStore>>> = OnceLock::new(); #[instrument(skip(ecstore))] pub async fn init_iam_sys(ecstore: Arc<ECStore>) -> Result<()> { if IAM_SYS.get().is_some() { info!("IAM system already initialized, skipping."); return Ok(()); } info!("Starting IAM system initialization sequence..."); // 1. Create the persistent storage adapter let storage_adapter = ObjectStore::new(ecstore); // 2. Create the cache manager. // The `new` method now performs a blocking initial load from disk. let cache_manager = IamCache::new(storage_adapter).await; // 3. Construct the system interface let iam_instance = Arc::new(IamSys::new(cache_manager)); // 4. Securely set the global singleton if IAM_SYS.set(iam_instance).is_err() { error!("Critical: Race condition detected during IAM initialization!"); return Err(Error::IamSysAlreadyInitialized); } info!("IAM system initialization completed successfully."); Ok(()) } #[inline] pub fn get() -> Result<Arc<IamSys<ObjectStore>>> { let sys = IAM_SYS.get().map(Arc::clone).ok_or(Error::IamSysNotInitialized)?; // Double-check the internal readiness state. The OnceLock is only set // after initialization and data loading complete, so this is a defensive // guard to ensure callers never operate on a partially initialized system. if !sys.is_ready() { return Err(Error::IamSysNotInitialized); } Ok(sys) } pub fn get_global_iam_sys() -> Option<Arc<IamSys<ObjectStore>>> { IAM_SYS.get().cloned() }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/iam/src/store.rs
crates/iam/src/store.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. pub mod object; use crate::cache::Cache; use crate::error::Result; use rustfs_policy::{auth::UserIdentity, policy::PolicyDoc}; use serde::{Deserialize, Serialize, de::DeserializeOwned}; use std::collections::{HashMap, HashSet}; use time::OffsetDateTime; #[async_trait::async_trait] pub trait Store: Clone + Send + Sync + 'static { fn has_watcher(&self) -> bool; async fn save_iam_config<Item: Serialize + Send>(&self, item: Item, path: impl AsRef<str> + Send) -> Result<()>; async fn load_iam_config<Item: DeserializeOwned>(&self, path: impl AsRef<str> + Send) -> Result<Item>; async fn delete_iam_config(&self, path: impl AsRef<str> + Send) -> Result<()>; async fn save_user_identity(&self, name: &str, user_type: UserType, item: UserIdentity, ttl: Option<usize>) -> Result<()>; async fn delete_user_identity(&self, name: &str, user_type: UserType) -> Result<()>; async fn load_user_identity(&self, name: &str, user_type: UserType) -> Result<UserIdentity>; async fn load_user(&self, name: &str, user_type: UserType, m: &mut HashMap<String, UserIdentity>) -> Result<()>; async fn load_users(&self, user_type: UserType, m: &mut HashMap<String, UserIdentity>) -> Result<()>; async fn load_secret_key(&self, name: &str, user_type: UserType) -> Result<String>; async fn save_group_info(&self, name: &str, item: GroupInfo) -> Result<()>; async fn delete_group_info(&self, name: &str) -> Result<()>; async fn load_group(&self, name: &str, m: &mut HashMap<String, GroupInfo>) -> Result<()>; async fn load_groups(&self, m: &mut HashMap<String, GroupInfo>) -> Result<()>; async fn save_policy_doc(&self, name: &str, item: PolicyDoc) -> Result<()>; async fn delete_policy_doc(&self, name: &str) -> Result<()>; async fn load_policy(&self, name: &str) -> Result<PolicyDoc>; async fn load_policy_doc(&self, name: &str, m: &mut HashMap<String, PolicyDoc>) -> Result<()>; async fn load_policy_docs(&self, m: &mut HashMap<String, PolicyDoc>) -> Result<()>; async fn save_mapped_policy( &self, name: &str, user_type: UserType, is_group: bool, item: MappedPolicy, ttl: Option<usize>, ) -> Result<()>; async fn delete_mapped_policy(&self, name: &str, user_type: UserType, is_group: bool) -> Result<()>; async fn load_mapped_policy( &self, name: &str, user_type: UserType, is_group: bool, m: &mut HashMap<String, MappedPolicy>, ) -> Result<()>; async fn load_mapped_policies( &self, user_type: UserType, is_group: bool, m: &mut HashMap<String, MappedPolicy>, ) -> Result<()>; async fn load_all(&self, cache: &Cache) -> Result<()>; } #[derive(Debug, PartialEq, Eq, Clone, Copy)] pub enum UserType { Svc, Sts, Reg, None, } impl UserType { pub fn prefix(&self) -> &'static str { match self { UserType::Svc => "service-accounts/", UserType::Sts => "sts/", UserType::Reg => "users/", UserType::None => "", } } pub fn to_u64(&self) -> u64 { match self { UserType::Svc => 1, UserType::Sts => 2, UserType::Reg => 3, UserType::None => 0, } } pub fn from_u64(u64: u64) -> Option<Self> { match u64 { 1 => Some(UserType::Svc), 2 => Some(UserType::Sts), 3 => Some(UserType::Reg), 0 => Some(UserType::None), _ => None, } } } #[derive(Serialize, Deserialize, Clone)] pub struct MappedPolicy { pub version: i64, pub policies: String, pub update_at: OffsetDateTime, } impl Default for MappedPolicy { fn default() -> Self { Self { version: 0, policies: "".to_owned(), update_at: OffsetDateTime::now_utc(), } } } impl MappedPolicy { pub fn new(policy: &str) -> Self { Self { version: 1, policies: policy.to_owned(), update_at: OffsetDateTime::now_utc(), } } pub fn to_slice(&self) -> Vec<String> { self.policies .split(",") .filter(|v| !v.trim().is_empty()) .map(|v| v.to_string()) .collect() } pub fn policy_set(&self) -> HashSet<String> { self.policies .split(",") .filter(|v| !v.trim().is_empty()) .map(|v| v.to_string()) .collect() } } #[derive(Serialize, Deserialize, Clone, Debug, Default)] pub struct GroupInfo { pub version: i64, pub status: String, pub members: Vec<String>, pub update_at: Option<OffsetDateTime>, } impl GroupInfo { pub fn new(members: Vec<String>) -> Self { Self { version: 1, status: "enabled".to_owned(), members, update_at: Some(OffsetDateTime::now_utc()), } } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/iam/src/manager.rs
crates/iam/src/manager.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use crate::error::{Error, Result, is_err_config_not_found}; use crate::sys::get_claims_from_token_with_secret; use crate::{ cache::{Cache, CacheEntity}, error::{Error as IamError, is_err_no_such_group, is_err_no_such_policy, is_err_no_such_user}, store::{GroupInfo, MappedPolicy, Store, UserType, object::IAM_CONFIG_PREFIX}, sys::{ MAX_SVCSESSION_POLICY_SIZE, SESSION_POLICY_NAME, SESSION_POLICY_NAME_EXTRACTED, STATUS_DISABLED, STATUS_ENABLED, UpdateServiceAccountOpts, }, }; use futures::future::join_all; use rustfs_credentials::{Credentials, EMBEDDED_POLICY_TYPE, INHERITED_POLICY_TYPE, get_global_action_cred}; use rustfs_madmin::{AccountStatus, AddOrUpdateUserReq, GroupDesc}; use rustfs_policy::{ arn::ARN, auth::{self, UserIdentity, is_secret_key_valid, jwt_sign}, format::Format, policy::{Policy, PolicyDoc, default::DEFAULT_POLICIES, iam_policy_claim_name_sa}, }; use rustfs_utils::path::path_join_buf; use serde::{Deserialize, Serialize}; use serde_json::Value; use std::sync::atomic::AtomicU8; use std::{ collections::{HashMap, HashSet}, sync::{ Arc, atomic::{AtomicBool, AtomicI64, Ordering}, }, time::Duration, }; use time::OffsetDateTime; use tokio::{ select, sync::{ mpsc, mpsc::{Receiver, Sender}, }, }; use tracing::warn; use tracing::{error, info}; const IAM_FORMAT_FILE: &str = "format.json"; const IAM_FORMAT_VERSION_1: i32 = 1; #[derive(Serialize, Deserialize)] struct IAMFormat { version: i32, } impl IAMFormat { fn new_version_1() -> Self { IAMFormat { version: IAM_FORMAT_VERSION_1, } } } fn get_iam_format_file_path() -> String { path_join_buf(&[&IAM_CONFIG_PREFIX, IAM_FORMAT_FILE]) } #[repr(u8)] #[derive(Debug, PartialEq)] pub enum IamState { Uninitialized = 0, Loading = 1, Ready = 2, Error = 3, } pub struct IamCache<T> { pub cache: Cache, pub api: T, pub state: Arc<AtomicU8>, pub loading: Arc<AtomicBool>, pub roles: HashMap<ARN, Vec<String>>, pub send_chan: Sender<i64>, pub last_timestamp: AtomicI64, } impl<T> IamCache<T> where T: Store, { /// Create a new IAM system instance /// # Arguments /// * `api` - The storage backend implementing the Store trait /// /// # Returns /// An Arc-wrapped instance of IamSystem pub(crate) async fn new(api: T) -> Arc<Self> { let (sender, receiver) = mpsc::channel::<i64>(100); let sys = Arc::new(Self { api, cache: Cache::default(), state: Arc::new(AtomicU8::new(IamState::Uninitialized as u8)), loading: Arc::new(AtomicBool::new(false)), send_chan: sender, roles: HashMap::new(), last_timestamp: AtomicI64::new(0), }); sys.clone().init(receiver).await.unwrap(); sys } /// Initialize the IAM system async fn init(self: Arc<Self>, receiver: Receiver<i64>) -> Result<()> { self.state.store(IamState::Loading as u8, Ordering::SeqCst); // Ensure the IAM format file is persisted first self.clone().save_iam_formatter().await?; // Critical: Load all existing users/policies into memory cache const MAX_RETRIES: usize = 3; for attempt in 0..MAX_RETRIES { if let Err(e) = self.clone().load().await { if attempt == MAX_RETRIES - 1 { self.state.store(IamState::Error as u8, Ordering::SeqCst); error!("IAM fail to load initial data after {} attempts: {:?}", MAX_RETRIES, e); return Err(e); } else { warn!("IAM load failed, retrying... attempt {}", attempt + 1); tokio::time::sleep(Duration::from_secs(1)).await; } } else { break; } } self.state.store(IamState::Ready as u8, Ordering::SeqCst); info!("IAM System successfully initialized and marked as READY"); // Background ticker for synchronization // Check if environment variable is set let skip_background_task = std::env::var("RUSTFS_SKIP_BACKGROUND_TASK").is_ok(); if !skip_background_task { // Background thread starts periodic updates or receives signal updates tokio::spawn({ let s = Arc::clone(&self); async move { let ticker = tokio::time::interval(Duration::from_secs(120)); tokio::pin!(ticker, receiver); loop { select! { _ = ticker.tick() => { info!("iam load ticker"); if let Err(err) =s.clone().load().await{ error!("iam load err {:?}", err); } }, i = receiver.recv() => { info!("iam load receiver"); match i { Some(t) => { let last = s.last_timestamp.load(Ordering::Relaxed); if last <= t { info!("iam load receiver load"); if let Err(err) =s.clone().load().await{ error!("iam load err {:?}", err); } ticker.reset(); } }, None => return, } } } } } }); } Ok(()) } /// Check if IAM system is ready pub fn is_ready(&self) -> bool { self.state.load(Ordering::SeqCst) == IamState::Ready as u8 } async fn _notify(&self) { self.send_chan.send(OffsetDateTime::now_utc().unix_timestamp()).await.unwrap(); } async fn load(self: Arc<Self>) -> Result<()> { // debug!("load iam to cache"); self.api.load_all(&self.cache).await?; self.last_timestamp .store(OffsetDateTime::now_utc().unix_timestamp(), Ordering::Relaxed); Ok(()) } pub async fn load_user(&self, access_key: &str) -> Result<()> { let mut users_map: HashMap<String, UserIdentity> = HashMap::new(); let mut user_policy_map = HashMap::new(); let mut sts_users_map = HashMap::new(); let mut sts_policy_map = HashMap::new(); let mut policy_docs_map = HashMap::new(); let _ = self.api.load_user(access_key, UserType::Svc, &mut users_map).await; let parent_user = users_map.get(access_key).map(|svc| svc.credentials.parent_user.clone()); if let Some(parent_user) = parent_user { let _ = self.api.load_user(&parent_user, UserType::Reg, &mut users_map).await; let _ = self .api .load_mapped_policy(&parent_user, UserType::Reg, false, &mut user_policy_map) .await; } else { let _ = self.api.load_user(access_key, UserType::Reg, &mut users_map).await; if users_map.contains_key(access_key) { let _ = self .api .load_mapped_policy(access_key, UserType::Reg, false, &mut user_policy_map) .await; } let _ = self.api.load_user(access_key, UserType::Sts, &mut sts_users_map).await; let has_sts_user = sts_users_map.get(access_key); let sts_parent = has_sts_user.map(|sts| sts.credentials.parent_user.clone()); if let Some(parent) = sts_parent { let _ = self .api .load_mapped_policy(&parent, UserType::Sts, false, &mut sts_policy_map) .await; } let sts_user = has_sts_user.map(|sts| sts.credentials.access_key.clone()); if let Some(ref sts) = sts_user && let Some(plc) = sts_policy_map.get(sts) { for p in plc.to_slice().iter() { if !policy_docs_map.contains_key(p) { let _ = self.api.load_policy_doc(p, &mut policy_docs_map).await; } } } } if let Some(plc) = user_policy_map.get(access_key) { for p in plc.to_slice().iter() { if !policy_docs_map.contains_key(p) { let _ = self.api.load_policy_doc(p, &mut policy_docs_map).await; } } } if let Some(user) = users_map.get(access_key) { Cache::add_or_update(&self.cache.users, access_key, user, OffsetDateTime::now_utc()); } if let Some(user_policy) = user_policy_map.get(access_key) { Cache::add_or_update(&self.cache.user_policies, access_key, user_policy, OffsetDateTime::now_utc()); } if let Some(sts_user) = sts_users_map.get(access_key) { Cache::add_or_update(&self.cache.sts_accounts, access_key, sts_user, OffsetDateTime::now_utc()); } if let Some(sts_policy) = sts_policy_map.get(access_key) { Cache::add_or_update(&self.cache.sts_policies, access_key, sts_policy, OffsetDateTime::now_utc()); } if let Some(policy_doc) = policy_docs_map.get(access_key) { Cache::add_or_update(&self.cache.policy_docs, access_key, policy_doc, OffsetDateTime::now_utc()); } Ok(()) } // TODO: Check if exists, whether retry is possible #[tracing::instrument(level = "debug", skip(self))] async fn save_iam_formatter(self: Arc<Self>) -> Result<()> { let path = get_iam_format_file_path(); let iam_fmt: Format = match self.api.load_iam_config(&path).await { Ok(v) => v, Err(err) => { if !is_err_config_not_found(&err) { return Err(err); } Format::default() } }; if iam_fmt.version >= IAM_FORMAT_VERSION_1 { return Ok(()); } self.api.save_iam_config(IAMFormat::new_version_1(), path).await } pub async fn get_user(&self, access_key: &str) -> Option<UserIdentity> { self.cache .users .load() .get(access_key) .cloned() .or_else(|| self.cache.sts_accounts.load().get(access_key).cloned()) } pub async fn get_mapped_policy(&self, name: &str, is_group: bool) -> Option<MappedPolicy> { if is_group { self.cache.group_policies.load().get(name).cloned() } else { self.cache.user_policies.load().get(name).cloned() } } pub async fn get_policy(&self, name: &str) -> Result<Policy> { if name.is_empty() { return Err(Error::InvalidArgument); } let policies = MappedPolicy::new(name).to_slice(); let mut to_merge = Vec::new(); for policy in policies { if policy.is_empty() { continue; } let v = self .cache .policy_docs .load() .get(&policy) .cloned() .ok_or(Error::NoSuchPolicy)?; to_merge.push(v.policy); } if to_merge.is_empty() { return Err(Error::NoSuchPolicy); } Ok(Policy::merge_policies(to_merge)) } pub async fn get_policy_doc(&self, name: &str) -> Result<PolicyDoc> { if name.is_empty() { return Err(Error::InvalidArgument); } self.cache.policy_docs.load().get(name).cloned().ok_or(Error::NoSuchPolicy) } pub async fn delete_policy(&self, name: &str, is_from_notify: bool) -> Result<()> { if name.is_empty() { return Err(Error::InvalidArgument); } if is_from_notify { let user_policy_cache = self.cache.user_policies.load(); let group_policy_cache = self.cache.group_policies.load(); let users_cache = self.cache.users.load(); let mut users = Vec::new(); user_policy_cache.iter().for_each(|(k, v)| { if !users_cache.contains_key(k) { Cache::delete(&self.cache.user_policies, k, OffsetDateTime::now_utc()); return; } if v.policy_set().contains(name) { users.push(k.to_owned()); } }); let mut groups = Vec::new(); group_policy_cache.iter().for_each(|(k, v)| { if v.policy_set().contains(name) { groups.push(k.to_owned()); } }); if !users.is_empty() || !groups.is_empty() { return Err(Error::PolicyInUse); } if let Err(err) = self.api.delete_policy_doc(name).await { if !is_err_no_such_policy(&err) { Cache::delete(&self.cache.policy_docs, name, OffsetDateTime::now_utc()); return Ok(()); } return Err(err); } } Cache::delete(&self.cache.policy_docs, name, OffsetDateTime::now_utc()); Ok(()) } pub async fn set_policy(&self, name: &str, policy: Policy) -> Result<OffsetDateTime> { if name.is_empty() || policy.is_empty() { return Err(Error::InvalidArgument); } let policy_doc = self .cache .policy_docs .load() .get(name) .map(|v| { let mut p = v.clone(); p.update(policy.clone()); p }) .unwrap_or(PolicyDoc::new(policy)); self.api.save_policy_doc(name, policy_doc.clone()).await?; let now = OffsetDateTime::now_utc(); Cache::add_or_update(&self.cache.policy_docs, name, &policy_doc, now); Ok(now) } pub async fn list_polices(&self, bucket_name: &str) -> Result<HashMap<String, Policy>> { let mut m = HashMap::new(); self.api.load_policy_docs(&mut m).await?; set_default_canned_policies(&mut m); let cache = CacheEntity::new(m.clone()).update_load_time(); self.cache.policy_docs.store(Arc::new(cache)); let items: Vec<_> = m.into_iter().map(|(k, v)| (k, v.policy.clone())).collect(); let futures: Vec<_> = items.iter().map(|(_, policy)| policy.match_resource(bucket_name)).collect(); let results = join_all(futures).await; let filtered = items .into_iter() .zip(results) .filter_map(|((k, policy), matches)| { if bucket_name.is_empty() || matches { Some((k, policy)) } else { None } }) .collect(); Ok(filtered) } pub async fn merge_policies(&self, name: &str) -> (String, Policy) { let mut policies = Vec::new(); let mut to_merge = Vec::new(); let mut miss_policies = Vec::new(); for policy in MappedPolicy::new(name).to_slice() { if policy.is_empty() { continue; } if let Some(v) = self.cache.policy_docs.load().get(&policy).cloned() { policies.push(policy); to_merge.push(v.policy); } else { miss_policies.push(policy); } } if !miss_policies.is_empty() { let mut m = HashMap::new(); for policy in miss_policies { let _ = self.api.load_policy_doc(&policy, &mut m).await; } for (k, v) in m.iter() { Cache::add_or_update(&self.cache.policy_docs, k, v, OffsetDateTime::now_utc()); policies.push(k.clone()); to_merge.push(v.policy.clone()); } } (policies.join(","), Policy::merge_policies(to_merge)) } pub async fn list_policy_docs(&self, bucket_name: &str) -> Result<HashMap<String, PolicyDoc>> { let mut m = HashMap::new(); self.api.load_policy_docs(&mut m).await?; set_default_canned_policies(&mut m); let cache = CacheEntity::new(m.clone()).update_load_time(); self.cache.policy_docs.store(Arc::new(cache)); let items: Vec<_> = m.into_iter().map(|(k, v)| (k, v.clone())).collect(); let futures: Vec<_> = items .iter() .map(|(_, policy_doc)| policy_doc.policy.match_resource(bucket_name)) .collect(); let results = join_all(futures).await; let filtered = items .into_iter() .zip(results) .filter_map(|((k, policy_doc), matches)| { if bucket_name.is_empty() || matches { Some((k, policy_doc)) } else { None } }) .collect(); Ok(filtered) } pub async fn list_policy_docs_internal(&self, bucket_name: &str) -> Result<HashMap<String, PolicyDoc>> { let cache = self.cache.policy_docs.load(); let items: Vec<_> = cache.iter().map(|(k, v)| (k.clone(), v.clone())).collect(); let futures: Vec<_> = items .iter() .map(|(_, policy_doc)| policy_doc.policy.match_resource(bucket_name)) .collect(); let results = join_all(futures).await; let ret = items .into_iter() .zip(results) .filter_map(|((k, policy_doc), matches)| { if bucket_name.is_empty() || matches { Some((k, policy_doc)) } else { None } }) .collect(); Ok(ret) } pub async fn list_temp_accounts(&self, access_key: &str) -> Result<Vec<UserIdentity>> { let users = self.cache.users.load(); let mut user_exists = false; let mut ret = Vec::new(); for (_, v) in users.iter() { let is_derived = v.credentials.is_service_account() || v.credentials.is_temp(); if !is_derived && v.credentials.access_key.as_str() == access_key { user_exists = true; } else if is_derived && v.credentials.parent_user == access_key { user_exists = true; if v.credentials.is_temp() { let mut u = v.clone(); u.credentials.secret_key = String::new(); u.credentials.session_token = String::new(); ret.push(u); } } } if !user_exists { return Err(Error::NoSuchUser(access_key.to_string())); } Ok(ret) } pub async fn list_sts_accounts(&self, access_key: &str) -> Result<Vec<Credentials>> { let users = self.cache.users.load(); Ok(users .values() .filter_map(|x| { if !access_key.is_empty() && x.credentials.parent_user.as_str() == access_key && x.credentials.is_temp() { let mut c = x.credentials.clone(); c.secret_key = String::new(); c.session_token = String::new(); return Some(c); } None }) .collect()) } pub async fn list_service_accounts(&self, access_key: &str) -> Result<Vec<Credentials>> { let users = self.cache.users.load(); Ok(users .values() .filter_map(|x| { if !access_key.is_empty() && x.credentials.parent_user.as_str() == access_key && x.credentials.is_service_account() { let mut c = x.credentials.clone(); c.secret_key = String::new(); c.session_token = String::new(); return Some(c); } None }) .collect()) } /// create a service account and update cache pub async fn add_service_account(&self, cred: Credentials) -> Result<OffsetDateTime> { if cred.access_key.is_empty() || cred.parent_user.is_empty() { return Err(Error::InvalidArgument); } let users = self.cache.users.load(); if let Some(x) = users.get(&cred.access_key) && x.credentials.is_service_account() { return Err(Error::IAMActionNotAllowed); } let u = UserIdentity::new(cred); self.api .save_user_identity(&u.credentials.access_key, UserType::Svc, u.clone(), None) .await?; self.update_user_with_claims(&u.credentials.access_key, u.clone())?; Ok(OffsetDateTime::now_utc()) } pub async fn update_service_account(&self, name: &str, opts: UpdateServiceAccountOpts) -> Result<OffsetDateTime> { let Some(ui) = self.cache.users.load().get(name).cloned() else { return Err(Error::NoSuchServiceAccount(name.to_string())); }; if !ui.credentials.is_service_account() { return Err(Error::NoSuchServiceAccount(name.to_string())); } let mut cr = ui.credentials.clone(); let current_secret_key = cr.secret_key.clone(); if let Some(secret) = opts.secret_key { if !is_secret_key_valid(&secret) { return Err(Error::InvalidSecretKeyLength); } cr.secret_key = secret; } if opts.name.is_some() { cr.name = opts.name; } if opts.description.is_some() { cr.description = opts.description; } if opts.expiration.is_some() { // TODO: check expiration cr.expiration = opts.expiration; } if let Some(status) = opts.status { match status.as_str() { val if val == AccountStatus::Enabled.as_ref() => cr.status = auth::ACCOUNT_ON.to_owned(), val if val == AccountStatus::Disabled.as_ref() => cr.status = auth::ACCOUNT_OFF.to_owned(), auth::ACCOUNT_ON => cr.status = auth::ACCOUNT_ON.to_owned(), auth::ACCOUNT_OFF => cr.status = auth::ACCOUNT_OFF.to_owned(), _ => cr.status = auth::ACCOUNT_OFF.to_owned(), } } let mut m: HashMap<String, Value> = get_claims_from_token_with_secret(&cr.session_token, &current_secret_key)?; m.remove(SESSION_POLICY_NAME_EXTRACTED); let nosp = if let Some(policy) = &opts.session_policy { policy.version.is_empty() && policy.statements.is_empty() } else { false }; if m.contains_key(SESSION_POLICY_NAME) && nosp { m.remove(SESSION_POLICY_NAME); m.insert(iam_policy_claim_name_sa(), Value::String(INHERITED_POLICY_TYPE.to_owned())); } if let Some(session_policy) = &opts.session_policy { session_policy.validate()?; if !session_policy.version.is_empty() && !session_policy.statements.is_empty() { let policy_buf = serde_json::to_vec(&session_policy)?; if policy_buf.len() > MAX_SVCSESSION_POLICY_SIZE { return Err(Error::PolicyTooLarge); } m.insert( SESSION_POLICY_NAME.to_owned(), Value::String(base64_simd::URL_SAFE_NO_PAD.encode_to_string(&policy_buf)), ); m.insert(iam_policy_claim_name_sa(), Value::String(EMBEDDED_POLICY_TYPE.to_owned())); } } m.insert("accessKey".to_owned(), Value::String(name.to_owned())); cr.session_token = jwt_sign(&m, &cr.secret_key)?; let u = UserIdentity::new(cr); self.api .save_user_identity(&u.credentials.access_key, UserType::Svc, u.clone(), None) .await?; self.update_user_with_claims(&u.credentials.access_key, u.clone())?; Ok(OffsetDateTime::now_utc()) } pub async fn policy_db_get(&self, name: &str, groups: &Option<Vec<String>>) -> Result<Vec<String>> { if name.is_empty() { return Err(Error::InvalidArgument); } let (mut policies, _) = self.policy_db_get_internal(name, false, false).await?; let present = !policies.is_empty(); if let Some(groups) = groups { for group in groups.iter() { let (gp, _) = self.policy_db_get_internal(group, true, present).await?; gp.iter().for_each(|v| { policies.push(v.clone()); }); } } Ok(policies) } async fn policy_db_get_internal( &self, name: &str, is_group: bool, policy_present: bool, ) -> Result<(Vec<String>, OffsetDateTime)> { if is_group { let groups = self.cache.groups.load(); let g = match groups.get(name) { Some(p) => p.clone(), None => { let mut m = HashMap::new(); self.api.load_group(name, &mut m).await?; if let Some(p) = m.get(name) { Cache::add_or_update(&self.cache.groups, name, p, OffsetDateTime::now_utc()); } m.get(name).cloned().ok_or(Error::NoSuchGroup(name.to_string()))? } }; if g.status == STATUS_DISABLED { return Ok((Vec::new(), OffsetDateTime::now_utc())); } if let Some(policy) = self.cache.group_policies.load().get(name) { return Ok((policy.to_slice(), policy.update_at)); } if !policy_present { let mut m = HashMap::new(); if let Err(err) = self.api.load_mapped_policy(name, UserType::Reg, true, &mut m).await && !is_err_no_such_policy(&err) { return Err(err); } if let Some(p) = m.get(name) { Cache::add_or_update(&self.cache.group_policies, name, p, OffsetDateTime::now_utc()); return Ok((p.to_slice(), p.update_at)); } return Ok((Vec::new(), OffsetDateTime::now_utc())); } return Ok((Vec::new(), OffsetDateTime::now_utc())); } let users = self.cache.users.load(); let u = users.get(name).cloned().unwrap_or_default(); if !u.credentials.is_valid() { return Ok((Vec::new(), OffsetDateTime::now_utc())); } let mp = match self.cache.user_policies.load().get(name) { Some(p) => p.clone(), None => { let mut m = HashMap::new(); if let Err(err) = self.api.load_mapped_policy(name, UserType::Reg, false, &mut m).await && !is_err_no_such_policy(&err) { return Err(err); } if let Some(p) = m.get(name) { Cache::add_or_update(&self.cache.user_policies, name, p, OffsetDateTime::now_utc()); p.clone() } else { match self.cache.sts_policies.load().get(name) { Some(p) => p.clone(), None => { let mut m = HashMap::new(); if let Err(err) = self.api.load_mapped_policy(name, UserType::Sts, false, &mut m).await && !is_err_no_such_policy(&err) { return Err(err); } if let Some(p) = m.get(name) { Cache::add_or_update(&self.cache.sts_policies, name, p, OffsetDateTime::now_utc()); p.clone() } else { MappedPolicy::default() } } } } } }; let mut policies: HashSet<String> = mp.to_slice().into_iter().collect(); if let Some(groups) = u.credentials.groups.as_ref() { for group in groups.iter() { if self .cache .groups .load() .get(group) .filter(|v| v.status == STATUS_DISABLED) .is_some() { return Ok((Vec::new(), OffsetDateTime::now_utc())); } let mp = match self.cache.group_policies.load().get(group) { Some(p) => p.clone(), None => { let mut m = HashMap::new(); if let Err(err) = self.api.load_mapped_policy(group, UserType::Reg, true, &mut m).await && !is_err_no_such_policy(&err) { return Err(err); } if let Some(p) = m.get(group) { Cache::add_or_update(&self.cache.group_policies, group, p, OffsetDateTime::now_utc()); p.clone() } else { MappedPolicy::default() } } }; mp.to_slice().iter().for_each(|v| { policies.insert(v.clone()); }); } } let update_at = mp.update_at; for group in self .cache .user_group_memberships .load() .get(name) .cloned() .unwrap_or_default() .iter() { if self .cache .groups .load() .get(group) .filter(|v| v.status == STATUS_DISABLED) .is_some() { return Ok((Vec::new(), OffsetDateTime::now_utc())); } let mp = match self.cache.group_policies.load().get(group) { Some(p) => p.clone(), None => { let mut m = HashMap::new(); if let Err(err) = self.api.load_mapped_policy(group, UserType::Reg, true, &mut m).await && !is_err_no_such_policy(&err) { return Err(err); } if let Some(p) = m.get(group) { Cache::add_or_update(&self.cache.group_policies, group, p, OffsetDateTime::now_utc()); p.clone() } else { MappedPolicy::default() } } }; mp.to_slice().iter().for_each(|v| { policies.insert(v.clone()); }); } Ok((policies.into_iter().collect(), update_at)) } pub async fn policy_db_set(&self, name: &str, user_type: UserType, is_group: bool, policy: &str) -> Result<OffsetDateTime> { if name.is_empty() { return Err(Error::InvalidArgument); } if policy.is_empty() {
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
true
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/iam/src/error.rs
crates/iam/src/error.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use rustfs_policy::policy::Error as PolicyError; pub type Result<T> = core::result::Result<T, Error>; #[derive(thiserror::Error, Debug)] pub enum Error { #[error(transparent)] PolicyError(#[from] PolicyError), #[error("{0}")] StringError(String), #[error("crypto: {0}")] CryptoError(#[from] rustfs_crypto::Error), #[error("user '{0}' does not exist")] NoSuchUser(String), #[error("account '{0}' does not exist")] NoSuchAccount(String), #[error("service account '{0}' does not exist")] NoSuchServiceAccount(String), #[error("temp account '{0}' does not exist")] NoSuchTempAccount(String), #[error("group '{0}' does not exist")] NoSuchGroup(String), #[error("policy does not exist")] NoSuchPolicy, #[error("policy in use")] PolicyInUse, #[error("group not empty")] GroupNotEmpty, #[error("invalid arguments specified")] InvalidArgument, #[error("not initialized")] IamSysNotInitialized, #[error("invalid service type: {0}")] InvalidServiceType(String), #[error("malformed credential")] ErrCredMalformed, #[error("CredNotInitialized")] CredNotInitialized, #[error("invalid access key length")] InvalidAccessKeyLength, #[error("invalid secret key length")] InvalidSecretKeyLength, #[error("access key contains reserved characters =,")] ContainsReservedChars, #[error("group name contains reserved characters =,")] GroupNameContainsReservedChars, #[error("jwt err {0}")] JWTError(jsonwebtoken::errors::Error), #[error("no access key")] NoAccessKey, #[error("invalid token")] InvalidToken, #[error("invalid access_key")] InvalidAccessKey, #[error("action not allowed")] IAMActionNotAllowed, #[error("invalid expiration")] InvalidExpiration, #[error("no secret key with access key")] NoSecretKeyWithAccessKey, #[error("no access key with secret key")] NoAccessKeyWithSecretKey, #[error("policy too large")] PolicyTooLarge, #[error("config not found")] ConfigNotFound, #[error("io error: {0}")] Io(std::io::Error), #[error("system already initialized")] IamSysAlreadyInitialized, } impl PartialEq for Error { fn eq(&self, other: &Self) -> bool { match (self, other) { (Error::StringError(a), Error::StringError(b)) => a == b, (Error::NoSuchUser(a), Error::NoSuchUser(b)) => a == b, (Error::NoSuchAccount(a), Error::NoSuchAccount(b)) => a == b, (Error::NoSuchServiceAccount(a), Error::NoSuchServiceAccount(b)) => a == b, (Error::NoSuchTempAccount(a), Error::NoSuchTempAccount(b)) => a == b, (Error::NoSuchGroup(a), Error::NoSuchGroup(b)) => a == b, (Error::InvalidServiceType(a), Error::InvalidServiceType(b)) => a == b, (Error::Io(a), Error::Io(b)) => a.kind() == b.kind() && a.to_string() == b.to_string(), // For complex types like PolicyError, CryptoError, JWTError, compare string representations (a, b) => std::mem::discriminant(a) == std::mem::discriminant(b) && a.to_string() == b.to_string(), } } } impl Clone for Error { fn clone(&self) -> Self { match self { Error::PolicyError(e) => Error::StringError(e.to_string()), // Convert to string since PolicyError may not be cloneable Error::StringError(s) => Error::StringError(s.clone()), Error::CryptoError(e) => Error::StringError(format!("crypto: {e}")), // Convert to string Error::NoSuchUser(s) => Error::NoSuchUser(s.clone()), Error::NoSuchAccount(s) => Error::NoSuchAccount(s.clone()), Error::NoSuchServiceAccount(s) => Error::NoSuchServiceAccount(s.clone()), Error::NoSuchTempAccount(s) => Error::NoSuchTempAccount(s.clone()), Error::NoSuchGroup(s) => Error::NoSuchGroup(s.clone()), Error::NoSuchPolicy => Error::NoSuchPolicy, Error::PolicyInUse => Error::PolicyInUse, Error::GroupNotEmpty => Error::GroupNotEmpty, Error::InvalidArgument => Error::InvalidArgument, Error::IamSysNotInitialized => Error::IamSysNotInitialized, Error::InvalidServiceType(s) => Error::InvalidServiceType(s.clone()), Error::ErrCredMalformed => Error::ErrCredMalformed, Error::CredNotInitialized => Error::CredNotInitialized, Error::InvalidAccessKeyLength => Error::InvalidAccessKeyLength, Error::InvalidSecretKeyLength => Error::InvalidSecretKeyLength, Error::ContainsReservedChars => Error::ContainsReservedChars, Error::GroupNameContainsReservedChars => Error::GroupNameContainsReservedChars, Error::JWTError(e) => Error::StringError(format!("jwt err {e}")), // Convert to string Error::NoAccessKey => Error::NoAccessKey, Error::InvalidToken => Error::InvalidToken, Error::InvalidAccessKey => Error::InvalidAccessKey, Error::IAMActionNotAllowed => Error::IAMActionNotAllowed, Error::InvalidExpiration => Error::InvalidExpiration, Error::NoSecretKeyWithAccessKey => Error::NoSecretKeyWithAccessKey, Error::NoAccessKeyWithSecretKey => Error::NoAccessKeyWithSecretKey, Error::PolicyTooLarge => Error::PolicyTooLarge, Error::ConfigNotFound => Error::ConfigNotFound, Error::Io(e) => Error::Io(std::io::Error::new(e.kind(), e.to_string())), Error::IamSysAlreadyInitialized => Error::IamSysAlreadyInitialized, } } } impl Error { pub fn other<E>(error: E) -> Self where E: Into<Box<dyn std::error::Error + Send + Sync>>, { Error::Io(std::io::Error::other(error)) } } impl From<rustfs_ecstore::error::StorageError> for Error { fn from(e: rustfs_ecstore::error::StorageError) -> Self { match e { rustfs_ecstore::error::StorageError::ConfigNotFound => Error::ConfigNotFound, _ => Error::other(e), } } } impl From<Error> for rustfs_ecstore::error::StorageError { fn from(e: Error) -> Self { match e { Error::ConfigNotFound => rustfs_ecstore::error::StorageError::ConfigNotFound, _ => rustfs_ecstore::error::StorageError::other(e), } } } impl From<rustfs_policy::error::Error> for Error { fn from(e: rustfs_policy::error::Error) -> Self { match e { rustfs_policy::error::Error::PolicyTooLarge => Error::PolicyTooLarge, rustfs_policy::error::Error::InvalidArgument => Error::InvalidArgument, rustfs_policy::error::Error::InvalidServiceType(s) => Error::InvalidServiceType(s), rustfs_policy::error::Error::IAMActionNotAllowed => Error::IAMActionNotAllowed, rustfs_policy::error::Error::InvalidExpiration => Error::InvalidExpiration, rustfs_policy::error::Error::NoAccessKey => Error::NoAccessKey, rustfs_policy::error::Error::InvalidToken => Error::InvalidToken, rustfs_policy::error::Error::InvalidAccessKey => Error::InvalidAccessKey, rustfs_policy::error::Error::NoSecretKeyWithAccessKey => Error::NoSecretKeyWithAccessKey, rustfs_policy::error::Error::NoAccessKeyWithSecretKey => Error::NoAccessKeyWithSecretKey, rustfs_policy::error::Error::Io(e) => Error::Io(e), rustfs_policy::error::Error::JWTError(e) => Error::JWTError(e), rustfs_policy::error::Error::NoSuchUser(s) => Error::NoSuchUser(s), rustfs_policy::error::Error::NoSuchAccount(s) => Error::NoSuchAccount(s), rustfs_policy::error::Error::NoSuchServiceAccount(s) => Error::NoSuchServiceAccount(s), rustfs_policy::error::Error::NoSuchTempAccount(s) => Error::NoSuchTempAccount(s), rustfs_policy::error::Error::NoSuchGroup(s) => Error::NoSuchGroup(s), rustfs_policy::error::Error::NoSuchPolicy => Error::NoSuchPolicy, rustfs_policy::error::Error::PolicyInUse => Error::PolicyInUse, rustfs_policy::error::Error::GroupNotEmpty => Error::GroupNotEmpty, rustfs_policy::error::Error::InvalidAccessKeyLength => Error::InvalidAccessKeyLength, rustfs_policy::error::Error::InvalidSecretKeyLength => Error::InvalidSecretKeyLength, rustfs_policy::error::Error::ContainsReservedChars => Error::ContainsReservedChars, rustfs_policy::error::Error::GroupNameContainsReservedChars => Error::GroupNameContainsReservedChars, rustfs_policy::error::Error::CredNotInitialized => Error::CredNotInitialized, rustfs_policy::error::Error::IamSysNotInitialized => Error::IamSysNotInitialized, rustfs_policy::error::Error::PolicyError(e) => Error::PolicyError(e), rustfs_policy::error::Error::StringError(s) => Error::StringError(s), rustfs_policy::error::Error::CryptoError(e) => Error::CryptoError(e), rustfs_policy::error::Error::ErrCredMalformed => Error::ErrCredMalformed, rustfs_policy::error::Error::IamSysAlreadyInitialized => Error::IamSysAlreadyInitialized, } } } impl From<Error> for std::io::Error { fn from(e: Error) -> Self { std::io::Error::other(e) } } impl From<serde_json::Error> for Error { fn from(e: serde_json::Error) -> Self { Error::other(e) } } impl From<base64_simd::Error> for Error { fn from(e: base64_simd::Error) -> Self { Error::other(e) } } pub fn is_err_config_not_found(err: &Error) -> bool { matches!(err, Error::ConfigNotFound) } // pub fn is_err_no_such_user(e: &Error) -> bool { // matches!(e, Error::NoSuchUser(_)) // } pub fn is_err_no_such_policy(err: &Error) -> bool { matches!(err, Error::NoSuchPolicy) } pub fn is_err_no_such_user(err: &Error) -> bool { matches!(err, Error::NoSuchUser(_)) } pub fn is_err_no_such_account(err: &Error) -> bool { matches!(err, Error::NoSuchAccount(_)) } pub fn is_err_no_such_temp_account(err: &Error) -> bool { matches!(err, Error::NoSuchTempAccount(_)) } pub fn is_err_no_such_group(err: &Error) -> bool { matches!(err, Error::NoSuchGroup(_)) } pub fn is_err_no_such_service_account(err: &Error) -> bool { matches!(err, Error::NoSuchServiceAccount(_)) } // pub fn clone_err(e: &Error) -> Error { // if let Some(e) = e.downcast_ref::<DiskError>() { // clone_disk_err(e) // } else if let Some(e) = e.downcast_ref::<std::io::Error>() { // if let Some(code) = e.raw_os_error() { // Error::new(std::io::Error::from_raw_os_error(code)) // } else { // Error::new(std::io::Error::new(e.kind(), e.to_string())) // } // } else { // //TODO: Optimize other types // Error::msg(e.to_string()) // } // } #[cfg(test)] mod tests { use super::*; use std::io::{Error as IoError, ErrorKind}; #[test] fn test_iam_error_to_io_error_conversion() { let iam_errors = vec![ Error::NoSuchUser("testuser".to_string()), Error::NoSuchAccount("testaccount".to_string()), Error::InvalidArgument, Error::IAMActionNotAllowed, Error::PolicyTooLarge, Error::ConfigNotFound, ]; for iam_error in iam_errors { let io_error: std::io::Error = iam_error.clone().into(); // Check that conversion creates an io::Error assert_eq!(io_error.kind(), ErrorKind::Other); // Check that the error message is preserved assert!(io_error.to_string().contains(&iam_error.to_string())); } } #[test] fn test_iam_error_from_storage_error() { // Test conversion from StorageError let storage_error = rustfs_ecstore::error::StorageError::ConfigNotFound; let iam_error: Error = storage_error.into(); assert_eq!(iam_error, Error::ConfigNotFound); // Test reverse conversion let back_to_storage: rustfs_ecstore::error::StorageError = iam_error.into(); assert_eq!(back_to_storage, rustfs_ecstore::error::StorageError::ConfigNotFound); } #[test] fn test_iam_error_from_policy_error() { use rustfs_policy::error::Error as PolicyError; let policy_errors = vec![ (PolicyError::NoSuchUser("user1".to_string()), Error::NoSuchUser("user1".to_string())), (PolicyError::NoSuchPolicy, Error::NoSuchPolicy), (PolicyError::InvalidArgument, Error::InvalidArgument), (PolicyError::PolicyTooLarge, Error::PolicyTooLarge), ]; for (policy_error, expected_iam_error) in policy_errors { let converted_iam_error: Error = policy_error.into(); assert_eq!(converted_iam_error, expected_iam_error); } } #[test] fn test_iam_error_other_function() { let custom_error = "Custom IAM error"; let iam_error = Error::other(custom_error); match iam_error { Error::Io(io_error) => { assert!(io_error.to_string().contains(custom_error)); assert_eq!(io_error.kind(), ErrorKind::Other); } _ => panic!("Expected Io variant"), } } #[test] fn test_iam_error_from_serde_json() { // Test conversion from serde_json::Error let invalid_json = r#"{"invalid": json}"#; let json_error = serde_json::from_str::<serde_json::Value>(invalid_json).unwrap_err(); let iam_error: Error = json_error.into(); match iam_error { Error::Io(io_error) => { assert_eq!(io_error.kind(), ErrorKind::Other); } _ => panic!("Expected Io variant"), } } #[test] fn test_helper_functions() { // Test helper functions for error type checking assert!(is_err_config_not_found(&Error::ConfigNotFound)); assert!(!is_err_config_not_found(&Error::NoSuchPolicy)); assert!(is_err_no_such_policy(&Error::NoSuchPolicy)); assert!(!is_err_no_such_policy(&Error::ConfigNotFound)); assert!(is_err_no_such_user(&Error::NoSuchUser("test".to_string()))); assert!(!is_err_no_such_user(&Error::NoSuchAccount("test".to_string()))); assert!(is_err_no_such_account(&Error::NoSuchAccount("test".to_string()))); assert!(!is_err_no_such_account(&Error::NoSuchUser("test".to_string()))); assert!(is_err_no_such_temp_account(&Error::NoSuchTempAccount("test".to_string()))); assert!(!is_err_no_such_temp_account(&Error::NoSuchAccount("test".to_string()))); assert!(is_err_no_such_group(&Error::NoSuchGroup("test".to_string()))); assert!(!is_err_no_such_group(&Error::NoSuchUser("test".to_string()))); assert!(is_err_no_such_service_account(&Error::NoSuchServiceAccount("test".to_string()))); assert!(!is_err_no_such_service_account(&Error::NoSuchAccount("test".to_string()))); } #[test] fn test_iam_error_io_preservation() { // Test that Io variant preserves original io::Error let original_io = IoError::new(ErrorKind::PermissionDenied, "access denied"); let iam_error = Error::Io(original_io); let converted_io: std::io::Error = iam_error.into(); // Note: Our clone implementation creates a new io::Error with the same kind and message // but it becomes ErrorKind::Other when cloned assert_eq!(converted_io.kind(), ErrorKind::Other); assert!(converted_io.to_string().contains("access denied")); } #[test] fn test_error_display_format() { let test_cases = vec![ (Error::NoSuchUser("testuser".to_string()), "user 'testuser' does not exist"), (Error::NoSuchAccount("testaccount".to_string()), "account 'testaccount' does not exist"), (Error::InvalidArgument, "invalid arguments specified"), (Error::IAMActionNotAllowed, "action not allowed"), (Error::ConfigNotFound, "config not found"), ]; for (error, expected_message) in test_cases { assert_eq!(error.to_string(), expected_message); } } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/iam/src/utils.rs
crates/iam/src/utils.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use jsonwebtoken::{Algorithm, DecodingKey, EncodingKey, Header}; use rand::{Rng, RngCore}; use serde::{Serialize, de::DeserializeOwned}; use std::io::{Error, Result}; /// Generates a random access key of the specified length. /// /// # Arguments /// /// * `length` - The length of the access key to be generated. /// /// # Returns /// /// * `Result<String>` - A result containing the generated access key or an error if the length is invalid. /// /// # Errors /// /// * Returns an error if the length is less than 3. /// pub fn gen_access_key(length: usize) -> Result<String> { const ALPHA_NUMERIC_TABLE: [char; 36] = [ '0', '1', '2', '3', '4', '5', '6', '7', '8', '9', 'A', 'B', 'C', 'D', 'E', 'F', 'G', 'H', 'I', 'J', 'K', 'L', 'M', 'N', 'O', 'P', 'Q', 'R', 'S', 'T', 'U', 'V', 'W', 'X', 'Y', 'Z', ]; if length < 3 { return Err(Error::other("access key length is too short")); } let mut result = String::with_capacity(length); let mut rng = rand::rng(); for _ in 0..length { result.push(ALPHA_NUMERIC_TABLE[rng.random_range(0..ALPHA_NUMERIC_TABLE.len())]); } Ok(result) } /// Generates a random secret key of the specified length. /// /// # Arguments /// /// * `length` - The length of the secret key to be generated. /// /// # Returns /// /// * `Result<String>` - A result containing the generated secret key or an error if the length is invalid. /// /// # Errors /// /// * Returns an error if the length is less than 8. /// pub fn gen_secret_key(length: usize) -> Result<String> { use base64_simd::URL_SAFE_NO_PAD; if length < 8 { return Err(Error::other("secret key length is too short")); } let mut rng = rand::rng(); let mut key = vec![0u8; URL_SAFE_NO_PAD.estimated_decoded_length(length)]; rng.fill_bytes(&mut key); let encoded = URL_SAFE_NO_PAD.encode_to_string(&key); let key_str = encoded.replace("/", "+"); Ok(key_str) } pub fn generate_jwt<T: Serialize>(claims: &T, secret: &str) -> std::result::Result<String, jsonwebtoken::errors::Error> { let header = Header::new(Algorithm::HS512); jsonwebtoken::encode(&header, &claims, &EncodingKey::from_secret(secret.as_bytes())) } pub fn extract_claims<T: DeserializeOwned + Clone>( token: &str, secret: &str, ) -> std::result::Result<jsonwebtoken::TokenData<T>, jsonwebtoken::errors::Error> { jsonwebtoken::decode::<T>( token, &DecodingKey::from_secret(secret.as_bytes()), &jsonwebtoken::Validation::new(Algorithm::HS512), ) } #[cfg(test)] mod tests { use super::{extract_claims, gen_access_key, gen_secret_key, generate_jwt}; use serde::{Deserialize, Serialize}; #[test] fn test_gen_access_key_valid_length() { // Test valid access key generation let key = gen_access_key(10).unwrap(); assert_eq!(key.len(), 10); // Test different lengths let key_20 = gen_access_key(20).unwrap(); assert_eq!(key_20.len(), 20); let key_3 = gen_access_key(3).unwrap(); assert_eq!(key_3.len(), 3); } #[test] fn test_gen_access_key_uniqueness() { // Test that generated keys are unique let key1 = gen_access_key(16).unwrap(); let key2 = gen_access_key(16).unwrap(); assert_ne!(key1, key2, "Generated access keys should be unique"); } #[test] fn test_gen_access_key_character_set() { // Test that generated keys only contain valid characters let key = gen_access_key(100).unwrap(); for ch in key.chars() { assert!(ch.is_ascii_alphanumeric(), "Access key should only contain alphanumeric characters"); assert!( ch.is_ascii_uppercase() || ch.is_ascii_digit(), "Access key should only contain uppercase letters and digits" ); } } #[test] fn test_gen_access_key_invalid_length() { // Test error cases for invalid lengths assert!(gen_access_key(0).is_err(), "Should fail for length 0"); assert!(gen_access_key(1).is_err(), "Should fail for length 1"); assert!(gen_access_key(2).is_err(), "Should fail for length 2"); // Verify error message let error = gen_access_key(2).unwrap_err(); assert_eq!(error.to_string(), "access key length is too short"); } #[test] fn test_gen_secret_key_valid_length() { // Test valid secret key generation let key = gen_secret_key(10).unwrap(); assert!(!key.is_empty(), "Secret key should not be empty"); let key_20 = gen_secret_key(20).unwrap(); assert!(!key_20.is_empty(), "Secret key should not be empty"); } #[test] fn test_gen_secret_key_uniqueness() { // Test that generated secret keys are unique let key1 = gen_secret_key(16).unwrap(); let key2 = gen_secret_key(16).unwrap(); assert_ne!(key1, key2, "Generated secret keys should be unique"); } #[test] fn test_gen_secret_key_base64_format() { // Test that secret key is valid base64-like format let key = gen_secret_key(32).unwrap(); // Should not contain invalid characters for URL-safe base64 for ch in key.chars() { assert!( ch.is_ascii_alphanumeric() || ch == '+' || ch == '-' || ch == '_', "Secret key should be URL-safe base64 compatible" ); } } #[test] fn test_gen_secret_key_invalid_length() { // Test error cases for invalid lengths assert!(gen_secret_key(0).is_err(), "Should fail for length 0"); assert!(gen_secret_key(7).is_err(), "Should fail for length 7"); // Verify error message let error = gen_secret_key(5).unwrap_err(); assert_eq!(error.to_string(), "secret key length is too short"); } #[derive(Debug, Serialize, Deserialize, PartialEq, Clone)] struct Claims { sub: String, company: String, exp: usize, // Expiration time (as UTC timestamp) } #[test] fn test_generate_jwt_valid_token() { // Test JWT generation with valid claims let claims = Claims { sub: "user1".to_string(), company: "example".to_string(), exp: 9999999999, // Far future timestamp for testing }; let secret = "my_secret"; let token = generate_jwt(&claims, secret).unwrap(); assert!(!token.is_empty(), "JWT token should not be empty"); // JWT should have 3 parts separated by dots let parts: Vec<&str> = token.split('.').collect(); assert_eq!(parts.len(), 3, "JWT should have 3 parts (header.payload.signature)"); // Each part should be non-empty for part in parts { assert!(!part.is_empty(), "JWT parts should not be empty"); } } #[test] fn test_generate_jwt_different_secrets() { // Test that different secrets produce different tokens let claims = Claims { sub: "user1".to_string(), company: "example".to_string(), exp: 9999999999, // Far future timestamp for testing }; let token1 = generate_jwt(&claims, "secret1").unwrap(); let token2 = generate_jwt(&claims, "secret2").unwrap(); assert_ne!(token1, token2, "Different secrets should produce different tokens"); } #[test] fn test_generate_jwt_different_claims() { // Test that different claims produce different tokens let claims1 = Claims { sub: "user1".to_string(), company: "example".to_string(), exp: 9999999999, // Far future timestamp for testing }; let claims2 = Claims { sub: "user2".to_string(), company: "example".to_string(), exp: 9999999999, // Far future timestamp for testing }; let secret = "my_secret"; let token1 = generate_jwt(&claims1, secret).unwrap(); let token2 = generate_jwt(&claims2, secret).unwrap(); assert_ne!(token1, token2, "Different claims should produce different tokens"); } #[test] fn test_extract_claims_valid_token() { // Test JWT claims extraction with valid token let original_claims = Claims { sub: "user1".to_string(), company: "example".to_string(), exp: 9999999999, // Far future timestamp for testing }; let secret = "my_secret"; let token = generate_jwt(&original_claims, secret).unwrap(); let decoded = extract_claims::<Claims>(&token, secret).unwrap(); assert_eq!(decoded.claims, original_claims, "Decoded claims should match original claims"); } #[test] fn test_extract_claims_invalid_secret() { // Test JWT claims extraction with wrong secret let claims = Claims { sub: "user1".to_string(), company: "example".to_string(), exp: 9999999999, // Far future timestamp for testing }; let token = generate_jwt(&claims, "correct_secret").unwrap(); let result = extract_claims::<Claims>(&token, "wrong_secret"); assert!(result.is_err(), "Should fail with wrong secret"); } #[test] fn test_extract_claims_invalid_token() { // Test JWT claims extraction with invalid token format let invalid_tokens = [ "invalid.token", "not.a.jwt.token", "", "header.payload", // Missing signature "invalid_base64.invalid_base64.invalid_base64", ]; for invalid_token in &invalid_tokens { let result = extract_claims::<Claims>(invalid_token, "secret"); assert!(result.is_err(), "Should fail with invalid token: {invalid_token}"); } } #[test] fn test_jwt_round_trip_consistency() { // Test complete round-trip: generate -> extract -> verify let original_claims = Claims { sub: "test_user".to_string(), company: "test_company".to_string(), exp: 9999999999, // Far future timestamp for testing }; let secret = "test_secret_key"; // Generate token let token = generate_jwt(&original_claims, secret).unwrap(); // Extract claims let decoded = extract_claims::<Claims>(&token, secret).unwrap(); // Verify claims match assert_eq!(decoded.claims, original_claims); // Verify token data structure assert!(matches!(decoded.header.alg, jsonwebtoken::Algorithm::HS512)); } #[test] fn test_jwt_with_empty_claims() { // Test JWT with minimal claims let empty_claims = Claims { sub: String::new(), company: String::new(), exp: 9999999999, // Far future timestamp for testing }; let secret = "secret"; let token = generate_jwt(&empty_claims, secret).unwrap(); let decoded = extract_claims::<Claims>(&token, secret).unwrap(); assert_eq!(decoded.claims, empty_claims); } #[test] fn test_jwt_with_special_characters() { // Test JWT with special characters in claims let special_claims = Claims { sub: "user@example.com".to_string(), company: "Company & Co. (Ltd.)".to_string(), exp: 9999999999, // Far future timestamp for testing }; let secret = "secret_with_special_chars!@#$%"; let token = generate_jwt(&special_claims, secret).unwrap(); let decoded = extract_claims::<Claims>(&token, secret).unwrap(); assert_eq!(decoded.claims, special_claims); } #[test] fn test_access_key_length_boundaries() { // Test boundary conditions for access key length assert!(gen_access_key(3).is_ok(), "Length 3 should be valid (minimum)"); assert!(gen_access_key(1000).is_ok(), "Large length should be valid"); // Test that minimum length is enforced let min_key = gen_access_key(3).unwrap(); assert_eq!(min_key.len(), 3); } #[test] fn test_secret_key_length_boundaries() { // Test boundary conditions for secret key length assert!(gen_secret_key(8).is_ok(), "Length 8 should be valid (minimum)"); assert!(gen_secret_key(1000).is_ok(), "Large length should be valid"); // Test that minimum length is enforced let result = gen_secret_key(8); assert!(result.is_ok(), "Minimum valid length should work"); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/iam/src/cache.rs
crates/iam/src/cache.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::{ collections::{HashMap, HashSet}, ops::{Deref, DerefMut}, ptr, sync::Arc, }; use arc_swap::{ArcSwap, AsRaw, Guard}; use rustfs_policy::{ auth::UserIdentity, policy::{Args, PolicyDoc}, }; use time::OffsetDateTime; use tracing::warn; use crate::store::{GroupInfo, MappedPolicy}; pub struct Cache { pub policy_docs: ArcSwap<CacheEntity<PolicyDoc>>, pub users: ArcSwap<CacheEntity<UserIdentity>>, pub user_policies: ArcSwap<CacheEntity<MappedPolicy>>, pub sts_accounts: ArcSwap<CacheEntity<UserIdentity>>, pub sts_policies: ArcSwap<CacheEntity<MappedPolicy>>, pub groups: ArcSwap<CacheEntity<GroupInfo>>, pub user_group_memberships: ArcSwap<CacheEntity<HashSet<String>>>, pub group_policies: ArcSwap<CacheEntity<MappedPolicy>>, } impl Default for Cache { fn default() -> Self { Self { policy_docs: ArcSwap::new(Arc::new(CacheEntity::default())), users: ArcSwap::new(Arc::new(CacheEntity::default())), user_policies: ArcSwap::new(Arc::new(CacheEntity::default())), sts_accounts: ArcSwap::new(Arc::new(CacheEntity::default())), sts_policies: ArcSwap::new(Arc::new(CacheEntity::default())), groups: ArcSwap::new(Arc::new(CacheEntity::default())), user_group_memberships: ArcSwap::new(Arc::new(CacheEntity::default())), group_policies: ArcSwap::new(Arc::new(CacheEntity::default())), } } } impl Cache { pub fn ptr_eq<Base, A, B>(a: A, b: B) -> bool where A: AsRaw<Base>, B: AsRaw<Base>, { let a = a.as_raw(); let b = b.as_raw(); ptr::eq(a, b) } fn exec<T: Clone>(target: &ArcSwap<CacheEntity<T>>, t: OffsetDateTime, mut op: impl FnMut(&mut CacheEntity<T>)) { let mut cur = target.load(); loop { // If the current update time is later than the execution time, // the background task is loaded and the current operation does not need to be performed. if cur.load_time >= t { return; } let mut new = CacheEntity::clone(&cur); op(&mut new); // Replace content with CAS atoms let prev = target.compare_and_swap(&*cur, Arc::new(new)); let swapped = Self::ptr_eq(&*cur, &*prev); if swapped { return; } else { cur = prev; } } } pub fn add_or_update<T: Clone>(target: &ArcSwap<CacheEntity<T>>, key: &str, value: &T, t: OffsetDateTime) { Self::exec(target, t, |map: &mut CacheEntity<T>| { map.insert(key.to_string(), value.clone()); }) } pub fn delete<T: Clone>(target: &ArcSwap<CacheEntity<T>>, key: &str, t: OffsetDateTime) { Self::exec(target, t, |map: &mut CacheEntity<T>| { map.remove(key); }) } pub fn build_user_group_memberships(&self) { let groups = self.groups.load(); let mut user_group_memberships = HashMap::new(); for (group_name, group) in groups.iter() { for user_name in &group.members { user_group_memberships .entry(user_name.clone()) .or_insert_with(HashSet::new) .insert(group_name.clone()); } } self.user_group_memberships .store(Arc::new(CacheEntity::new(user_group_memberships))); } } impl CacheInner { #[inline] pub fn get_user(&self, user_name: &str) -> Option<&UserIdentity> { self.users.get(user_name).or_else(|| self.sts_accounts.get(user_name)) } // fn get_policy(&self, _name: &str, _groups: &[String]) -> crate::Result<Vec<Policy>> { // todo!() // } // /// Return Ok(Some(parent_name)) when the user is temporary. // /// Return Ok(None) for non-temporary users. // fn is_temp_user(&self, user_name: &str) -> crate::Result<Option<&str>> { // let user = self // .get_user(user_name) // .ok_or_else(|| Error::NoSuchUser(user_name.to_owned()))?; // if user.credentials.is_temp() { // Ok(Some(&user.credentials.parent_user)) // } else { // Ok(None) // } // } // /// Return Ok(Some(parent_name)) when the user is a temporary identity. // /// Return Ok(None) when the user is not temporary. // fn is_service_account(&self, user_name: &str) -> crate::Result<Option<&str>> { // let user = self // .get_user(user_name) // .ok_or_else(|| Error::NoSuchUser(user_name.to_owned()))?; // if user.credentials.is_service_account() { // Ok(Some(&user.credentials.parent_user)) // } else { // Ok(None) // } // } // todo pub fn is_allowed_sts(&self, _args: &Args, _parent: &str) -> bool { warn!("unimplement is_allowed_sts"); false } // todo pub fn is_allowed_service_account(&self, _args: &Args, _parent: &str) -> bool { warn!("unimplement is_allowed_sts"); false } pub fn is_allowed(&self, _args: Args) -> bool { todo!() } pub fn policy_db_get(&self, _name: &str, _groups: &[String]) -> Vec<String> { todo!() } } #[derive(Clone)] pub struct CacheEntity<T> { map: HashMap<String, T>, /// The time of the reload load_time: OffsetDateTime, } impl<T> Deref for CacheEntity<T> { type Target = HashMap<String, T>; fn deref(&self) -> &Self::Target { &self.map } } impl<T> DerefMut for CacheEntity<T> { fn deref_mut(&mut self) -> &mut Self::Target { &mut self.map } } impl<T> CacheEntity<T> { pub fn new(map: HashMap<String, T>) -> Self { Self { map, load_time: OffsetDateTime::UNIX_EPOCH, } } } impl<T> Default for CacheEntity<T> { fn default() -> Self { Self { map: HashMap::new(), load_time: OffsetDateTime::UNIX_EPOCH, } } } impl<T> CacheEntity<T> { pub fn update_load_time(mut self) -> Self { self.load_time = OffsetDateTime::now_utc(); self } } pub type G<T> = Guard<Arc<CacheEntity<T>>>; pub struct CacheInner { pub policy_docs: G<PolicyDoc>, pub users: G<UserIdentity>, pub user_policies: G<MappedPolicy>, pub sts_accounts: G<UserIdentity>, pub sts_policies: G<MappedPolicy>, pub groups: G<GroupInfo>, pub user_group_memberships: G<HashSet<String>>, pub group_policies: G<MappedPolicy>, } impl From<&Cache> for CacheInner { fn from(value: &Cache) -> Self { Self { policy_docs: value.policy_docs.load(), users: value.users.load(), user_policies: value.user_policies.load(), sts_accounts: value.sts_accounts.load(), sts_policies: value.sts_policies.load(), groups: value.groups.load(), user_group_memberships: value.user_group_memberships.load(), group_policies: value.group_policies.load(), } } } #[cfg(test)] mod tests { use std::sync::Arc; use arc_swap::ArcSwap; use futures::future::join_all; use time::OffsetDateTime; use super::CacheEntity; use crate::cache::Cache; #[tokio::test] async fn test_cache_entity_add() { let cache = ArcSwap::new(Arc::new(CacheEntity::<usize>::default())); let mut f = vec![]; for (index, key) in (0..100).map(|x| x.to_string()).enumerate() { let c = &cache; f.push(async move { Cache::add_or_update(c, &key, &index, OffsetDateTime::now_utc()); }); } join_all(f).await; let cache = cache.load(); for (index, key) in (0..100).map(|x| x.to_string()).enumerate() { assert_eq!(cache.get(&key), Some(&index)); } } #[tokio::test] async fn test_cache_entity_update() { let cache = ArcSwap::new(Arc::new(CacheEntity::<usize>::default())); let mut f = vec![]; for (index, key) in (0..100).map(|x| x.to_string()).enumerate() { let c = &cache; f.push(async move { Cache::add_or_update(c, &key, &index, OffsetDateTime::now_utc()); }); } join_all(f).await; let cache_load = cache.load(); for (index, key) in (0..100).map(|x| x.to_string()).enumerate() { assert_eq!(cache_load.get(&key), Some(&index)); } let mut f = vec![]; for (index, key) in (0..100).map(|x| x.to_string()).enumerate() { let c = &cache; f.push(async move { Cache::add_or_update(c, &key, &(index * 1000), OffsetDateTime::now_utc()); }); } join_all(f).await; let cache_load = cache.load(); for (index, key) in (0..100).map(|x| x.to_string()).enumerate() { assert_eq!(cache_load.get(&key), Some(&(index * 1000))); } } #[tokio::test] async fn test_cache_entity_delete() { let cache = ArcSwap::new(Arc::new(CacheEntity::<usize>::default())); let mut f = vec![]; for (index, key) in (0..100).map(|x| x.to_string()).enumerate() { let c = &cache; f.push(async move { Cache::add_or_update(c, &key, &index, OffsetDateTime::now_utc()); }); } join_all(f).await; let cache_load = cache.load(); for (index, key) in (0..100).map(|x| x.to_string()).enumerate() { assert_eq!(cache_load.get(&key), Some(&index)); } let mut f = vec![]; for key in (0..100).map(|x| x.to_string()) { let c = &cache; f.push(async move { Cache::delete(c, &key, OffsetDateTime::now_utc()); }); } join_all(f).await; let cache_load = cache.load(); assert!(cache_load.is_empty()); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/iam/src/sys.rs
crates/iam/src/sys.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use crate::error::Error as IamError; use crate::error::is_err_no_such_account; use crate::error::is_err_no_such_temp_account; use crate::error::{Error, Result}; use crate::manager::IamCache; use crate::manager::extract_jwt_claims; use crate::manager::get_default_policyes; use crate::store::GroupInfo; use crate::store::MappedPolicy; use crate::store::Store; use crate::store::UserType; use crate::utils::extract_claims; use rustfs_credentials::{Credentials, EMBEDDED_POLICY_TYPE, INHERITED_POLICY_TYPE, get_global_action_cred}; use rustfs_ecstore::notification_sys::get_global_notification_sys; use rustfs_madmin::AddOrUpdateUserReq; use rustfs_madmin::GroupDesc; use rustfs_policy::arn::ARN; use rustfs_policy::auth::{ ACCOUNT_ON, UserIdentity, contains_reserved_chars, create_new_credentials_with_metadata, generate_credentials, is_access_key_valid, is_secret_key_valid, }; use rustfs_policy::policy::Args; use rustfs_policy::policy::opa; use rustfs_policy::policy::{Policy, PolicyDoc, iam_policy_claim_name_sa}; use serde_json::Value; use serde_json::json; use std::collections::HashMap; use std::sync::Arc; use std::sync::OnceLock; use time::OffsetDateTime; use tokio::sync::RwLock; use tracing::{error, info, warn}; pub const MAX_SVCSESSION_POLICY_SIZE: usize = 4096; pub const STATUS_ENABLED: &str = "enabled"; pub const STATUS_DISABLED: &str = "disabled"; pub const POLICYNAME: &str = "policy"; pub const SESSION_POLICY_NAME: &str = "sessionPolicy"; pub const SESSION_POLICY_NAME_EXTRACTED: &str = "sessionPolicy-extracted"; static POLICY_PLUGIN_CLIENT: OnceLock<Arc<RwLock<Option<rustfs_policy::policy::opa::AuthZPlugin>>>> = OnceLock::new(); fn get_policy_plugin_client() -> Arc<RwLock<Option<rustfs_policy::policy::opa::AuthZPlugin>>> { POLICY_PLUGIN_CLIENT.get_or_init(|| Arc::new(RwLock::new(None))).clone() } pub struct IamSys<T> { store: Arc<IamCache<T>>, roles_map: HashMap<ARN, String>, } impl<T: Store> IamSys<T> { /// Create a new IamSys instance with the given IamCache store /// /// # Arguments /// * `store` - An Arc to the IamCache instance /// /// # Returns /// A new instance of IamSys pub fn new(store: Arc<IamCache<T>>) -> Self { tokio::spawn(async move { match opa::lookup_config().await { Ok(conf) => { if conf.enable() { Self::set_policy_plugin_client(opa::AuthZPlugin::new(conf)).await; info!("OPA plugin enabled"); } } Err(e) => { error!("Error loading OPA configuration err:{}", e); } }; }); Self { store, roles_map: HashMap::new(), } } /// Check if the IamSys has a watcher configured /// /// # Returns /// `true` if a watcher is configured, `false` otherwise pub fn has_watcher(&self) -> bool { self.store.api.has_watcher() } pub async fn set_policy_plugin_client(client: rustfs_policy::policy::opa::AuthZPlugin) { let policy_plugin_client = get_policy_plugin_client(); let mut guard = policy_plugin_client.write().await; *guard = Some(client); } pub async fn get_policy_plugin_client() -> Option<rustfs_policy::policy::opa::AuthZPlugin> { let policy_plugin_client = get_policy_plugin_client(); let guard = policy_plugin_client.read().await; guard.clone() } pub async fn load_group(&self, name: &str) -> Result<()> { self.store.group_notification_handler(name).await } pub async fn load_groups(&self, m: &mut HashMap<String, GroupInfo>) -> Result<()> { self.store.api.load_groups(m).await } pub async fn load_policy(&self, name: &str) -> Result<()> { self.store.policy_notification_handler(name).await } pub async fn load_policy_mapping(&self, name: &str, user_type: UserType, is_group: bool) -> Result<()> { self.store .policy_mapping_notification_handler(name, user_type, is_group) .await } pub async fn load_user(&self, name: &str, user_type: UserType) -> Result<()> { self.store.user_notification_handler(name, user_type).await } pub async fn load_users(&self, user_type: UserType, m: &mut HashMap<String, UserIdentity>) -> Result<()> { self.store.api.load_users(user_type, m).await?; Ok(()) } pub async fn load_service_account(&self, name: &str) -> Result<()> { self.store.user_notification_handler(name, UserType::Svc).await } pub async fn delete_policy(&self, name: &str, notify: bool) -> Result<()> { for k in get_default_policyes().keys() { if k == name { return Err(Error::other("system policy can not be deleted")); } } self.store.delete_policy(name, notify).await?; if !notify || self.has_watcher() { return Ok(()); } if let Some(notification_sys) = get_global_notification_sys() { let resp = notification_sys.delete_policy(name).await; for r in resp { if let Some(err) = r.err { warn!("notify delete_policy failed: {}", err); } } } Ok(()) } pub async fn info_policy(&self, name: &str) -> Result<rustfs_madmin::PolicyInfo> { let d = self.store.get_policy_doc(name).await?; let pdata = serde_json::to_string(&d.policy)?; Ok(rustfs_madmin::PolicyInfo { policy_name: name.to_string(), policy: json!(pdata), create_date: d.create_date, update_date: d.update_date, }) } pub async fn load_mapped_policies( &self, user_type: UserType, is_group: bool, m: &mut HashMap<String, MappedPolicy>, ) -> Result<()> { self.store.api.load_mapped_policies(user_type, is_group, m).await } pub async fn list_polices(&self, bucket_name: &str) -> Result<HashMap<String, Policy>> { self.store.list_polices(bucket_name).await } pub async fn list_policy_docs(&self, bucket_name: &str) -> Result<HashMap<String, PolicyDoc>> { self.store.list_policy_docs(bucket_name).await } pub async fn set_policy(&self, name: &str, policy: Policy) -> Result<OffsetDateTime> { let updated_at = self.store.set_policy(name, policy).await?; if !self.has_watcher() && let Some(notification_sys) = get_global_notification_sys() { let resp = notification_sys.load_policy(name).await; for r in resp { if let Some(err) = r.err { warn!("notify load_policy failed: {}", err); } } } Ok(updated_at) } pub async fn get_role_policy(&self, arn_str: &str) -> Result<(ARN, String)> { let Some(arn) = ARN::parse(arn_str).ok() else { return Err(Error::other("Invalid ARN")); }; let Some(policy) = self.roles_map.get(&arn) else { return Err(Error::other("No such role")); }; Ok((arn, policy.clone())) } pub async fn delete_user(&self, name: &str, notify: bool) -> Result<()> { self.store.delete_user(name, UserType::Reg).await?; if notify && !self.has_watcher() && let Some(notification_sys) = get_global_notification_sys() { let resp = notification_sys.delete_user(name).await; for r in resp { if let Some(err) = r.err { warn!("notify delete_user failed: {}", err); } } } Ok(()) } async fn notify_for_user(&self, name: &str, is_temp: bool) { if self.has_watcher() { return; } // Fire-and-forget notification to peers - don't block auth operations // This is critical for cluster recovery: login should not wait for dead peers let name = name.to_string(); tokio::spawn(async move { if let Some(notification_sys) = get_global_notification_sys() { let resp = notification_sys.load_user(&name, is_temp).await; for r in resp { if let Some(err) = r.err { warn!("notify load_user failed (non-blocking): {}", err); } } } }); } async fn notify_for_service_account(&self, name: &str) { if self.has_watcher() { return; } // Fire-and-forget notification to peers - don't block service account operations let name = name.to_string(); tokio::spawn(async move { if let Some(notification_sys) = get_global_notification_sys() { let resp = notification_sys.load_service_account(&name).await; for r in resp { if let Some(err) = r.err { warn!("notify load_service_account failed (non-blocking): {}", err); } } } }); } pub async fn current_policies(&self, name: &str) -> String { self.store.merge_policies(name).await.0 } pub async fn list_bucket_users(&self, bucket_name: &str) -> Result<HashMap<String, rustfs_madmin::UserInfo>> { self.store.get_bucket_users(bucket_name).await } pub async fn list_users(&self) -> Result<HashMap<String, rustfs_madmin::UserInfo>> { self.store.get_users().await } pub async fn set_temp_user(&self, name: &str, cred: &Credentials, policy_name: Option<&str>) -> Result<OffsetDateTime> { let updated_at = self.store.set_temp_user(name, cred, policy_name).await?; self.notify_for_user(&cred.access_key, true).await; Ok(updated_at) } pub async fn is_temp_user(&self, name: &str) -> Result<(bool, String)> { let Some(u) = self.store.get_user(name).await else { return Err(IamError::NoSuchUser(name.to_string())); }; if u.credentials.is_temp() { Ok((true, u.credentials.parent_user)) } else { Ok((false, "".to_string())) } } pub async fn is_service_account(&self, name: &str) -> Result<(bool, String)> { let Some(u) = self.store.get_user(name).await else { return Err(IamError::NoSuchUser(name.to_string())); }; if u.credentials.is_service_account() { Ok((true, u.credentials.parent_user)) } else { Ok((false, "".to_string())) } } pub async fn get_user_info(&self, name: &str) -> Result<rustfs_madmin::UserInfo> { self.store.get_user_info(name).await } pub async fn set_user_status(&self, name: &str, status: rustfs_madmin::AccountStatus) -> Result<OffsetDateTime> { let updated_at = self.store.set_user_status(name, status).await?; self.notify_for_user(name, false).await; Ok(updated_at) } pub async fn new_service_account( &self, parent_user: &str, groups: Option<Vec<String>>, opts: NewServiceAccountOpts, ) -> Result<(Credentials, OffsetDateTime)> { if parent_user.is_empty() { return Err(IamError::InvalidArgument); } if !opts.access_key.is_empty() && opts.secret_key.is_empty() { return Err(IamError::NoSecretKeyWithAccessKey); } if !opts.secret_key.is_empty() && opts.access_key.is_empty() { return Err(IamError::NoAccessKeyWithSecretKey); } if parent_user == opts.access_key { return Err(IamError::IAMActionNotAllowed); } if opts.expiration.is_none() { return Err(IamError::InvalidExpiration); } // TODO: check allow_site_replicator_account let policy_buf = if let Some(policy) = opts.session_policy { policy.validate()?; let buf = serde_json::to_vec(&policy)?; if buf.len() > MAX_SVCSESSION_POLICY_SIZE { return Err(IamError::PolicyTooLarge); } buf } else { Vec::new() }; let mut m: HashMap<String, Value> = HashMap::new(); m.insert("parent".to_owned(), Value::String(parent_user.to_owned())); if !policy_buf.is_empty() { m.insert( SESSION_POLICY_NAME.to_owned(), Value::String(base64_simd::URL_SAFE_NO_PAD.encode_to_string(&policy_buf)), ); m.insert(iam_policy_claim_name_sa(), Value::String(EMBEDDED_POLICY_TYPE.to_owned())); } else { m.insert(iam_policy_claim_name_sa(), Value::String(INHERITED_POLICY_TYPE.to_owned())); } if let Some(claims) = opts.claims { for (k, v) in claims.iter() { if !m.contains_key(k) { m.insert(k.to_owned(), v.to_owned()); } } } // set expiration time default to 1 hour m.insert( "exp".to_string(), Value::Number(serde_json::Number::from( opts.expiration .map_or(OffsetDateTime::now_utc().unix_timestamp() + 3600, |t| t.unix_timestamp()), )), ); let (access_key, secret_key) = if !opts.access_key.is_empty() || !opts.secret_key.is_empty() { (opts.access_key, opts.secret_key) } else { generate_credentials()? }; let mut cred = create_new_credentials_with_metadata(&access_key, &secret_key, &m, &secret_key)?; cred.parent_user = parent_user.to_owned(); cred.groups = groups; cred.status = ACCOUNT_ON.to_owned(); cred.name = opts.name; cred.description = opts.description; cred.expiration = opts.expiration; let create_at = self.store.add_service_account(cred.clone()).await?; self.notify_for_service_account(&cred.access_key).await; Ok((cred, create_at)) } pub async fn update_service_account(&self, name: &str, opts: UpdateServiceAccountOpts) -> Result<OffsetDateTime> { let updated_at = self.store.update_service_account(name, opts).await?; self.notify_for_service_account(name).await; Ok(updated_at) } pub async fn list_service_accounts(&self, access_key: &str) -> Result<Vec<Credentials>> { self.store.list_service_accounts(access_key).await } pub async fn list_temp_accounts(&self, access_key: &str) -> Result<Vec<UserIdentity>> { self.store.list_temp_accounts(access_key).await } pub async fn list_sts_accounts(&self, access_key: &str) -> Result<Vec<Credentials>> { self.store.list_sts_accounts(access_key).await } pub async fn get_service_account(&self, access_key: &str) -> Result<(Credentials, Option<Policy>)> { let (mut da, policy) = self.get_service_account_internal(access_key).await?; da.credentials.secret_key.clear(); da.credentials.session_token.clear(); Ok((da.credentials, policy)) } async fn get_service_account_internal(&self, access_key: &str) -> Result<(UserIdentity, Option<Policy>)> { let (sa, claims) = match self.get_account_with_claims(access_key).await { Ok(res) => res, Err(err) => { if is_err_no_such_account(&err) { return Err(IamError::NoSuchServiceAccount(access_key.to_string())); } return Err(err); } }; if !sa.credentials.is_service_account() { return Err(IamError::NoSuchServiceAccount(access_key.to_string())); } let op_pt = claims.get(&iam_policy_claim_name_sa()); let op_sp = claims.get(SESSION_POLICY_NAME); if let (Some(pt), Some(sp)) = (op_pt, op_sp) && pt == EMBEDDED_POLICY_TYPE { let policy = serde_json::from_slice(&base64_simd::URL_SAFE_NO_PAD.decode_to_vec(sp.as_str().unwrap_or_default().as_bytes())?)?; return Ok((sa, Some(policy))); } Ok((sa, None)) } async fn get_account_with_claims(&self, access_key: &str) -> Result<(UserIdentity, HashMap<String, Value>)> { let Some(acc) = self.store.get_user(access_key).await else { return Err(IamError::NoSuchAccount(access_key.to_string())); }; let m = extract_jwt_claims(&acc)?; Ok((acc, m)) } pub async fn get_temporary_account(&self, access_key: &str) -> Result<(Credentials, Option<Policy>)> { let (mut sa, policy) = match self.get_temp_account(access_key).await { Ok(res) => res, Err(err) => { if is_err_no_such_temp_account(&err) { // TODO: load_user match self.get_temp_account(access_key).await { Ok(res) => res, Err(err) => return Err(err), }; } return Err(err); } }; sa.credentials.secret_key.clear(); sa.credentials.session_token.clear(); Ok((sa.credentials, policy)) } async fn get_temp_account(&self, access_key: &str) -> Result<(UserIdentity, Option<Policy>)> { let (sa, claims) = match self.get_account_with_claims(access_key).await { Ok(res) => res, Err(err) => { if is_err_no_such_account(&err) { return Err(IamError::NoSuchTempAccount(access_key.to_string())); } return Err(err); } }; if !sa.credentials.is_temp() { return Err(IamError::NoSuchTempAccount(access_key.to_string())); } let op_pt = claims.get(&iam_policy_claim_name_sa()); let op_sp = claims.get(SESSION_POLICY_NAME); if let (Some(pt), Some(sp)) = (op_pt, op_sp) && pt == EMBEDDED_POLICY_TYPE { let policy = serde_json::from_slice(&base64_simd::URL_SAFE_NO_PAD.decode_to_vec(sp.as_str().unwrap_or_default().as_bytes())?)?; return Ok((sa, Some(policy))); } Ok((sa, None)) } pub async fn get_claims_for_svc_acc(&self, access_key: &str) -> Result<HashMap<String, Value>> { let Some(u) = self.store.get_user(access_key).await else { return Err(IamError::NoSuchServiceAccount(access_key.to_string())); }; if !u.credentials.is_service_account() { return Err(IamError::NoSuchServiceAccount(access_key.to_string())); } extract_jwt_claims(&u) } pub async fn delete_service_account(&self, access_key: &str, notify: bool) -> Result<()> { let Some(u) = self.store.get_user(access_key).await else { return Ok(()); }; if !u.credentials.is_service_account() { return Ok(()); } self.store.delete_user(access_key, UserType::Svc).await?; if notify && !self.has_watcher() && let Some(notification_sys) = get_global_notification_sys() { let resp = notification_sys.delete_service_account(access_key).await; for r in resp { if let Some(err) = r.err { warn!("notify delete_service_account failed: {}", err); } } } Ok(()) } async fn notify_for_group(&self, group: &str) { if self.has_watcher() { return; } // Fire-and-forget notification to peers - don't block group operations let group = group.to_string(); tokio::spawn(async move { if let Some(notification_sys) = get_global_notification_sys() { let resp = notification_sys.load_group(&group).await; for r in resp { if let Some(err) = r.err { warn!("notify load_group failed (non-blocking): {}", err); } } } }); } pub async fn create_user(&self, access_key: &str, args: &AddOrUpdateUserReq) -> Result<OffsetDateTime> { if !is_access_key_valid(access_key) { return Err(IamError::InvalidAccessKeyLength); } if contains_reserved_chars(access_key) { return Err(IamError::ContainsReservedChars); } if !is_secret_key_valid(&args.secret_key) { return Err(IamError::InvalidSecretKeyLength); } let updated_at = self.store.add_user(access_key, args).await?; self.notify_for_user(access_key, false).await; Ok(updated_at) } pub async fn set_user_secret_key(&self, access_key: &str, secret_key: &str) -> Result<()> { if !is_access_key_valid(access_key) { return Err(IamError::InvalidAccessKeyLength); } if !is_secret_key_valid(secret_key) { return Err(IamError::InvalidSecretKeyLength); } self.store.update_user_secret_key(access_key, secret_key).await } /// Add SSH public key for a user (for SFTP authentication) pub async fn add_user_ssh_public_key(&self, access_key: &str, public_key: &str) -> Result<()> { if !is_access_key_valid(access_key) { return Err(IamError::InvalidAccessKeyLength); } if public_key.is_empty() { return Err(IamError::InvalidArgument); } self.store.add_user_ssh_public_key(access_key, public_key).await } pub async fn check_key(&self, access_key: &str) -> Result<(Option<UserIdentity>, bool)> { if let Some(sys_cred) = get_global_action_cred() && sys_cred.access_key == access_key { return Ok((Some(UserIdentity::new(sys_cred)), true)); } match self.store.get_user(access_key).await { Some(res) => { let ok = res.credentials.is_valid(); Ok((Some(res), ok)) } None => { let _ = self.store.load_user(access_key).await; if let Some(res) = self.store.get_user(access_key).await { let ok = res.credentials.is_valid(); Ok((Some(res), ok)) } else { Ok((None, false)) } } } } pub async fn get_user(&self, access_key: &str) -> Option<UserIdentity> { match self.check_key(access_key).await { Ok((u, _)) => u, _ => None, } } pub async fn add_users_to_group(&self, group: &str, users: Vec<String>) -> Result<OffsetDateTime> { if contains_reserved_chars(group) { return Err(IamError::GroupNameContainsReservedChars); } let updated_at = self.store.add_users_to_group(group, users).await?; self.notify_for_group(group).await; Ok(updated_at) } pub async fn remove_users_from_group(&self, group: &str, users: Vec<String>) -> Result<OffsetDateTime> { let updated_at = self.store.remove_users_from_group(group, users).await?; self.notify_for_group(group).await; Ok(updated_at) } pub async fn set_group_status(&self, group: &str, enable: bool) -> Result<OffsetDateTime> { let updated_at = self.store.set_group_status(group, enable).await?; self.notify_for_group(group).await; Ok(updated_at) } pub async fn get_group_description(&self, group: &str) -> Result<GroupDesc> { self.store.get_group_description(group).await } pub async fn list_groups_load(&self) -> Result<Vec<String>> { self.store.update_groups().await } pub async fn list_groups(&self) -> Result<Vec<String>> { self.store.list_groups().await } pub async fn policy_db_set(&self, name: &str, user_type: UserType, is_group: bool, policy: &str) -> Result<OffsetDateTime> { let updated_at = self.store.policy_db_set(name, user_type, is_group, policy).await?; if !self.has_watcher() && let Some(notification_sys) = get_global_notification_sys() { let resp = notification_sys.load_policy_mapping(name, user_type.to_u64(), is_group).await; for r in resp { if let Some(err) = r.err { warn!("notify load_policy failed: {}", err); } } } Ok(updated_at) } pub async fn policy_db_get(&self, name: &str, groups: &Option<Vec<String>>) -> Result<Vec<String>> { self.store.policy_db_get(name, groups).await } pub async fn is_allowed_sts(&self, args: &Args<'_>, parent_user: &str) -> bool { let is_owner = parent_user == get_global_action_cred().unwrap().access_key; let role_arn = args.get_role_arn(); let policies = { if is_owner { Vec::new() } else if role_arn.is_some() { let Ok(arn) = ARN::parse(role_arn.unwrap_or_default()) else { return false }; MappedPolicy::new(self.roles_map.get(&arn).map_or_else(String::default, |v| v.clone()).as_str()).to_slice() } else { let Ok(p) = self.policy_db_get(parent_user, args.groups).await else { return false }; p //TODO: FROM JWT } }; if !is_owner && policies.is_empty() { return false; } let combined_policy = { if is_owner { Policy::default() } else { let (a, c) = self.store.merge_policies(&policies.join(",")).await; if a.is_empty() { return false; } c } }; let (has_session_policy, is_allowed_sp) = is_allowed_by_session_policy(args); if has_session_policy { return is_allowed_sp && (is_owner || combined_policy.is_allowed(args).await); } is_owner || combined_policy.is_allowed(args).await } pub async fn is_allowed_service_account(&self, args: &Args<'_>, parent_user: &str) -> bool { let Some(p) = args.claims.get("parent") else { return false; }; if p.as_str() != Some(parent_user) { return false; } let is_owner = parent_user == get_global_action_cred().unwrap().access_key; let role_arn = args.get_role_arn(); let svc_policies = { if is_owner { Vec::new() } else if role_arn.is_some() { let Ok(arn) = ARN::parse(role_arn.unwrap_or_default()) else { return false }; MappedPolicy::new(self.roles_map.get(&arn).map_or_else(String::default, |v| v.clone()).as_str()).to_slice() } else { let Ok(p) = self.policy_db_get(parent_user, args.groups).await else { return false }; p } }; if !is_owner && svc_policies.is_empty() { return false; } let combined_policy = { if is_owner { Policy::default() } else { let (a, c) = self.store.merge_policies(&svc_policies.join(",")).await; if a.is_empty() { return false; } c } }; let mut parent_args = args.clone(); parent_args.account = parent_user; let Some(sa) = args.claims.get(&iam_policy_claim_name_sa()) else { return false; }; let Some(sa_str) = sa.as_str() else { return false; }; if sa_str == INHERITED_POLICY_TYPE { return is_owner || combined_policy.is_allowed(&parent_args).await; } let (has_session_policy, is_allowed_sp) = is_allowed_by_session_policy_for_service_account(args); if has_session_policy { return is_allowed_sp && (is_owner || combined_policy.is_allowed(&parent_args).await); } is_owner || combined_policy.is_allowed(&parent_args).await } pub async fn get_combined_policy(&self, policies: &[String]) -> Policy { self.store.merge_policies(&policies.join(",")).await.1 } pub async fn is_allowed(&self, args: &Args<'_>) -> bool { if args.is_owner { return true; } let opa_enable = Self::get_policy_plugin_client().await; if let Some(opa_enable) = opa_enable { return opa_enable.is_allowed(args).await; } let Ok((is_temp, parent_user)) = self.is_temp_user(args.account).await else { return false }; if is_temp { return self.is_allowed_sts(args, &parent_user).await; } let Ok((is_svc, parent_user)) = self.is_service_account(args.account).await else { return false }; if is_svc { return self.is_allowed_service_account(args, &parent_user).await; } let Ok(policies) = self.policy_db_get(args.account, args.groups).await else { return false }; if policies.is_empty() { return false; } self.get_combined_policy(&policies).await.is_allowed(args).await } /// Check if the underlying store is ready pub fn is_ready(&self) -> bool { self.store.is_ready() } } fn is_allowed_by_session_policy(args: &Args<'_>) -> (bool, bool) { let Some(policy) = args.claims.get(SESSION_POLICY_NAME_EXTRACTED) else { return (false, false); }; let has_session_policy = true; let Some(policy_str) = policy.as_str() else { return (has_session_policy, false); }; let Ok(sub_policy) = Policy::parse_config(policy_str.as_bytes()) else { return (has_session_policy, false); }; if sub_policy.version.is_empty() { return (has_session_policy, false); } let mut session_policy_args = args.clone(); session_policy_args.is_owner = false; (has_session_policy, pollster::block_on(sub_policy.is_allowed(&session_policy_args))) } fn is_allowed_by_session_policy_for_service_account(args: &Args<'_>) -> (bool, bool) { let Some(policy) = args.claims.get(SESSION_POLICY_NAME_EXTRACTED) else { return (false, false); }; let mut has_session_policy = true; let Some(policy_str) = policy.as_str() else { return (has_session_policy, false); }; let Ok(sub_policy) = Policy::parse_config(policy_str.as_bytes()) else { return (has_session_policy, false); }; if sub_policy.version.is_empty() && sub_policy.statements.is_empty() && sub_policy.id.is_empty() { has_session_policy = false; return (has_session_policy, false); } let mut session_policy_args = args.clone(); session_policy_args.is_owner = false; (has_session_policy, pollster::block_on(sub_policy.is_allowed(&session_policy_args))) } #[derive(Debug, Clone, Default)] pub struct NewServiceAccountOpts { pub session_policy: Option<Policy>, pub access_key: String, pub secret_key: String, pub name: Option<String>, pub description: Option<String>, pub expiration: Option<OffsetDateTime>, pub allow_site_replicator_account: bool, pub claims: Option<HashMap<String, Value>>, } pub struct UpdateServiceAccountOpts { pub session_policy: Option<Policy>, pub secret_key: Option<String>, pub name: Option<String>, pub description: Option<String>, pub expiration: Option<OffsetDateTime>, pub status: Option<String>, } pub fn get_claims_from_token_with_secret(token: &str, secret: &str) -> Result<HashMap<String, Value>> { let mut ms = extract_claims::<HashMap<String, Value>>(token, secret).map_err(|e| Error::other(format!("extract claims err {e}")))?;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
true
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/iam/src/store/object.rs
crates/iam/src/store/object.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use super::{GroupInfo, MappedPolicy, Store, UserType}; use crate::error::{Error, Result, is_err_config_not_found, is_err_no_such_group}; use crate::{ cache::{Cache, CacheEntity}, error::{is_err_no_such_policy, is_err_no_such_user}, manager::{extract_jwt_claims, get_default_policyes}, }; use futures::future::join_all; use rustfs_credentials::get_global_action_cred; use rustfs_ecstore::StorageAPI as _; use rustfs_ecstore::store_api::{ObjectInfoOrErr, WalkOptions}; use rustfs_ecstore::{ config::{ RUSTFS_CONFIG_PREFIX, com::{delete_config, read_config, read_config_with_metadata, save_config}, }, store::ECStore, store_api::{ObjectInfo, ObjectOptions}, }; use rustfs_policy::{auth::UserIdentity, policy::PolicyDoc}; use rustfs_utils::path::{SLASH_SEPARATOR, path_join_buf}; use serde::{Serialize, de::DeserializeOwned}; use std::sync::LazyLock; use std::{collections::HashMap, sync::Arc}; use tokio::sync::mpsc::{self, Sender}; use tokio_util::sync::CancellationToken; use tracing::{debug, error, info, warn}; pub static IAM_CONFIG_PREFIX: LazyLock<String> = LazyLock::new(|| format!("{RUSTFS_CONFIG_PREFIX}/iam")); pub static IAM_CONFIG_USERS_PREFIX: LazyLock<String> = LazyLock::new(|| format!("{RUSTFS_CONFIG_PREFIX}/iam/users/")); pub static IAM_CONFIG_SERVICE_ACCOUNTS_PREFIX: LazyLock<String> = LazyLock::new(|| format!("{RUSTFS_CONFIG_PREFIX}/iam/service-accounts/")); pub static IAM_CONFIG_GROUPS_PREFIX: LazyLock<String> = LazyLock::new(|| format!("{RUSTFS_CONFIG_PREFIX}/iam/groups/")); pub static IAM_CONFIG_POLICIES_PREFIX: LazyLock<String> = LazyLock::new(|| format!("{RUSTFS_CONFIG_PREFIX}/iam/policies/")); pub static IAM_CONFIG_STS_PREFIX: LazyLock<String> = LazyLock::new(|| format!("{RUSTFS_CONFIG_PREFIX}/iam/sts/")); pub static IAM_CONFIG_POLICY_DB_PREFIX: LazyLock<String> = LazyLock::new(|| format!("{RUSTFS_CONFIG_PREFIX}/iam/policydb/")); pub static IAM_CONFIG_POLICY_DB_USERS_PREFIX: LazyLock<String> = LazyLock::new(|| format!("{RUSTFS_CONFIG_PREFIX}/iam/policydb/users/")); pub static IAM_CONFIG_POLICY_DB_STS_USERS_PREFIX: LazyLock<String> = LazyLock::new(|| format!("{RUSTFS_CONFIG_PREFIX}/iam/policydb/sts-users/")); pub static IAM_CONFIG_POLICY_DB_SERVICE_ACCOUNTS_PREFIX: LazyLock<String> = LazyLock::new(|| format!("{RUSTFS_CONFIG_PREFIX}/iam/policydb/service-accounts/")); pub static IAM_CONFIG_POLICY_DB_GROUPS_PREFIX: LazyLock<String> = LazyLock::new(|| format!("{RUSTFS_CONFIG_PREFIX}/iam/policydb/groups/")); const IAM_IDENTITY_FILE: &str = "identity.json"; const IAM_POLICY_FILE: &str = "policy.json"; const IAM_GROUP_MEMBERS_FILE: &str = "members.json"; fn get_user_identity_path(user: &str, user_type: UserType) -> String { let base_path: &str = match user_type { UserType::Svc => &IAM_CONFIG_SERVICE_ACCOUNTS_PREFIX, UserType::Sts => &IAM_CONFIG_STS_PREFIX, _ => &IAM_CONFIG_USERS_PREFIX, }; path_join_buf(&[base_path, user, IAM_IDENTITY_FILE]) } fn get_group_info_path(group: &str) -> String { path_join_buf(&[&IAM_CONFIG_GROUPS_PREFIX, group, IAM_GROUP_MEMBERS_FILE]) } fn get_policy_doc_path(name: &str) -> String { path_join_buf(&[&IAM_CONFIG_POLICIES_PREFIX, name, IAM_POLICY_FILE]) } fn get_mapped_policy_path(name: &str, user_type: UserType, is_group: bool) -> String { if is_group { return path_join_buf(&[&IAM_CONFIG_POLICY_DB_GROUPS_PREFIX, format!("{name}.json").as_str()]); } match user_type { UserType::Svc => path_join_buf(&[&IAM_CONFIG_POLICY_DB_SERVICE_ACCOUNTS_PREFIX, format!("{name}.json").as_str()]), UserType::Sts => path_join_buf(&[&IAM_CONFIG_POLICY_DB_STS_USERS_PREFIX, format!("{name}.json").as_str()]), _ => path_join_buf(&[&IAM_CONFIG_POLICY_DB_USERS_PREFIX, format!("{name}.json").as_str()]), } } #[derive(Debug)] pub struct StringOrErr { pub item: Option<String>, pub err: Option<Error>, } const USERS_LIST_KEY: &str = "users/"; const SVC_ACC_LIST_KEY: &str = "service-accounts/"; const GROUPS_LIST_KEY: &str = "groups/"; const POLICIES_LIST_KEY: &str = "policies/"; const STS_LIST_KEY: &str = "sts/"; const POLICY_DB_PREFIX: &str = "policydb/"; const POLICY_DB_USERS_LIST_KEY: &str = "policydb/users/"; const POLICY_DB_STS_USERS_LIST_KEY: &str = "policydb/sts-users/"; const POLICY_DB_GROUPS_LIST_KEY: &str = "policydb/groups/"; // split_path splits a path into a top-level directory and a child item. The // parent directory retains the trailing slash. fn split_path(s: &str, last_index: bool) -> (&str, &str) { let i = if last_index { s.rfind('/') } else { s.find('/') }; match i { Some(index) => (&s[..index + 1], &s[index + 1..]), None => (s, ""), } } #[derive(Clone)] pub struct ObjectStore { object_api: Arc<ECStore>, } impl ObjectStore { const BUCKET_NAME: &'static str = ".rustfs.sys"; pub fn new(object_api: Arc<ECStore>) -> Self { Self { object_api } } fn decrypt_data(data: &[u8]) -> Result<Vec<u8>> { let de = rustfs_crypto::decrypt_data(get_global_action_cred().unwrap_or_default().secret_key.as_bytes(), data)?; Ok(de) } fn encrypt_data(data: &[u8]) -> Result<Vec<u8>> { let en = rustfs_crypto::encrypt_data(get_global_action_cred().unwrap_or_default().secret_key.as_bytes(), data)?; Ok(en) } async fn load_iamconfig_bytes_with_metadata(&self, path: impl AsRef<str> + Send) -> Result<(Vec<u8>, ObjectInfo)> { let (data, obj) = read_config_with_metadata(self.object_api.clone(), path.as_ref(), &ObjectOptions::default()).await?; Ok((Self::decrypt_data(&data)?, obj)) } async fn list_iam_config_items(&self, prefix: &str, ctx: CancellationToken, sender: Sender<StringOrErr>) { // debug!("list iam config items, prefix: {}", &prefix); // TODO: Implement walk, use walk // let prefix = format!("{}{}", prefix, item); let store = self.object_api.clone(); let (tx, mut rx) = mpsc::channel::<ObjectInfoOrErr>(100); let path = prefix.to_owned(); tokio::spawn(async move { store .walk(ctx.clone(), Self::BUCKET_NAME, &path, tx, WalkOptions::default()) .await }); let prefix = prefix.to_owned(); tokio::spawn(async move { while let Some(v) = rx.recv().await { if let Some(err) = v.err { let _ = sender .send(StringOrErr { item: None, err: Some(err.into()), }) .await; return; } if let Some(info) = v.item { let object_name = if cfg!(target_os = "windows") { info.name.replace('\\', "/") } else { info.name }; let name = object_name.trim_start_matches(&prefix).trim_end_matches(SLASH_SEPARATOR); let _ = sender .send(StringOrErr { item: Some(name.to_owned()), err: None, }) .await; } } }); } async fn list_all_iamconfig_items(&self) -> Result<HashMap<String, Vec<String>>> { let (tx, mut rx) = mpsc::channel::<StringOrErr>(100); let ctx = CancellationToken::new(); self.list_iam_config_items(format!("{}/", *IAM_CONFIG_PREFIX).as_str(), ctx.clone(), tx) .await; let mut res = HashMap::new(); while let Some(v) = rx.recv().await { if let Some(err) = v.err { warn!("list_iam_config_items {:?}", err); ctx.cancel(); return Err(err); } if let Some(item) = v.item { let last_index = item.starts_with(POLICY_DB_PREFIX); let (list_key, trimmed_item) = split_path(&item, last_index); res.entry(list_key.to_owned()) .or_insert_with(Vec::new) .push(trimmed_item.to_owned()); } } ctx.cancel(); Ok(res) } async fn load_policy_doc_concurrent(&self, names: &[String]) -> Result<Vec<PolicyDoc>> { let mut futures = Vec::with_capacity(names.len()); for name in names { let policy_name = rustfs_utils::path::dir(name); futures.push(async move { match self.load_policy(&policy_name).await { Ok(p) => Ok(p), Err(err) => { if !is_err_no_such_policy(&err) { Err(Error::other(format!("load policy doc failed: {err}"))) } else { Ok(PolicyDoc::default()) } } } }); } let results = join_all(futures).await; let mut policies = Vec::with_capacity(results.len()); for r in results { match r { Ok(p) => policies.push(p), Err(e) => return Err(e), } } Ok(policies) } async fn load_user_concurrent(&self, names: &[String], user_type: UserType) -> Result<Vec<UserIdentity>> { let mut futures = Vec::with_capacity(names.len()); for name in names { let user_name = rustfs_utils::path::dir(name); futures.push(async move { match self.load_user_identity(&user_name, user_type).await { Ok(res) => Ok(res), Err(err) => { if !is_err_no_such_user(&err) { Err(Error::other(format!("load user failed: {err}"))) } else { Ok(UserIdentity::default()) } } } }); } let results = join_all(futures).await; let mut users = Vec::with_capacity(results.len()); for r in results { match r { Ok(u) => users.push(u), Err(e) => return Err(e), } } Ok(users) } async fn load_mapped_policy_internal(&self, name: &str, user_type: UserType, is_group: bool) -> Result<MappedPolicy> { let info: MappedPolicy = self .load_iam_config(get_mapped_policy_path(name, user_type, is_group)) .await .map_err(|err| { if is_err_config_not_found(&err) { Error::NoSuchPolicy } else { err } })?; Ok(info) } async fn load_mapped_policy_concurrent( &self, names: &[String], user_type: UserType, is_group: bool, ) -> Result<Vec<MappedPolicy>> { let mut futures = Vec::with_capacity(names.len()); for name in names { let policy_name = name.trim_end_matches(".json"); futures.push(async move { match self.load_mapped_policy_internal(policy_name, user_type, is_group).await { Ok(p) => Ok(p), Err(err) => { if !is_err_no_such_policy(&err) { Err(Error::other(format!("load mapped policy failed: {err}"))) } else { Ok(MappedPolicy::default()) } } } }); } let results = join_all(futures).await; let mut policies = Vec::with_capacity(results.len()); for r in results { match r { Ok(p) => policies.push(p), Err(e) => return Err(e), } } Ok(policies) } /// Checks if the underlying ECStore is ready for metadata operations. /// This prevents silent failures during the storage boot-up phase. /// /// Performs a lightweight probe by attempting to read a known configuration object. /// If the object is not found, it indicates the storage metadata is not ready. /// The upper-level caller should handle retries if needed. async fn check_storage_readiness(&self) -> Result<()> { // Probe path for a fixed object under the IAM root prefix. // If it doesn't exist, the system bucket or metadata is not ready. let probe_path = format!("{}/format.json", *IAM_CONFIG_PREFIX); match read_config(self.object_api.clone(), &probe_path).await { Ok(_) => Ok(()), Err(rustfs_ecstore::error::StorageError::ConfigNotFound) => Err(Error::other(format!( "Storage metadata not ready: probe object '{}' not found (expected IAM config to be initialized)", probe_path ))), Err(e) => Err(e.into()), } } // async fn load_policy(&self, name: &str) -> Result<PolicyDoc> { // let mut policy = self // .load_iam_config::<PolicyDoc>(&format!("config/iam/policies/{name}/policy.json")) // .await?; // // FIXME: // // if policy.version == 0 { // // policy.create_date = object.mod_time; // // policy.update_date = object.mod_time; // // } // Ok(policy) // } // async fn load_user_identity(&self, user_type: UserType, name: &str) -> Result<Option<UserIdentity>> { // let mut user = self // .load_iam_config::<UserIdentity>(&format!( // "config/iam/{base}{name}/identity.json", // base = user_type.prefix(), // name = name // )) // .await?; // if user.credentials.is_expired() { // return Ok(None); // } // if user.credentials.access_key.is_empty() { // user.credentials.access_key = name.to_owned(); // } // // todo, validate session token // Ok(Some(user)) // } } #[async_trait::async_trait] impl Store for ObjectStore { fn has_watcher(&self) -> bool { false } async fn load_iam_config<Item: DeserializeOwned>(&self, path: impl AsRef<str> + Send) -> Result<Item> { let mut data = read_config(self.object_api.clone(), path.as_ref()).await?; data = match Self::decrypt_data(&data) { Ok(v) => v, Err(err) => { warn!("config decrypt failed, keeping file: {}, path: {}", err, path.as_ref()); // keep the config file when decrypt failed - do not delete return Err(Error::ConfigNotFound); } }; Ok(serde_json::from_slice(&data)?) } /// Saves IAM configuration with a retry mechanism on failure. /// /// Attempts to save the IAM configuration up to 5 times if the storage layer is not ready, /// using exponential backoff between attempts (starting at 200ms, doubling each retry). /// /// # Arguments /// /// * `item` - The IAM configuration item to save, must implement `Serialize` and `Send`. /// * `path` - The path where the configuration will be saved. /// /// # Returns /// /// * `Result<()>` - `Ok(())` on success, or an `Error` if all attempts fail. #[tracing::instrument(level = "debug", skip(self, item, path))] async fn save_iam_config<Item: Serialize + Send>(&self, item: Item, path: impl AsRef<str> + Send) -> Result<()> { let mut data = serde_json::to_vec(&item)?; data = Self::encrypt_data(&data)?; let mut attempts = 0; let max_attempts = 5; let path_ref = path.as_ref(); loop { match save_config(self.object_api.clone(), path_ref, data.clone()).await { Ok(_) => { debug!("Successfully saved IAM config to {}", path_ref); return Ok(()); } Err(e) if attempts < max_attempts => { attempts += 1; // Exponential backoff: 200ms, 400ms, 800ms... let wait_ms = 200 * (1 << attempts); warn!( "Storage layer not ready for IAM write (attempt {}/{}). Retrying in {}ms. Path: {}, Error: {:?}", attempts, max_attempts, wait_ms, path_ref, e ); tokio::time::sleep(std::time::Duration::from_millis(wait_ms)).await; } Err(e) => { error!("Final failure saving IAM config to {}: {:?}", path_ref, e); return Err(e.into()); } } } } async fn delete_iam_config(&self, path: impl AsRef<str> + Send) -> Result<()> { delete_config(self.object_api.clone(), path.as_ref()).await?; Ok(()) } async fn save_user_identity( &self, name: &str, user_type: UserType, user_identity: UserIdentity, _ttl: Option<usize>, ) -> Result<()> { // Pre-check storage health self.check_storage_readiness().await?; let path = get_user_identity_path(name, user_type); debug!("Saving IAM identity to path: {}", path); self.save_iam_config(user_identity, path).await.map_err(|e| { error!("ObjectStore save failure for {}: {:?}", name, e); e }) } async fn delete_user_identity(&self, name: &str, user_type: UserType) -> Result<()> { self.delete_iam_config(get_user_identity_path(name, user_type)) .await .map_err(|err| { if is_err_config_not_found(&err) { Error::NoSuchPolicy } else { err } })?; Ok(()) } async fn load_user_identity(&self, name: &str, user_type: UserType) -> Result<UserIdentity> { let mut u: UserIdentity = self .load_iam_config(get_user_identity_path(name, user_type)) .await .map_err(|err| { if is_err_config_not_found(&err) { warn!("load_user_identity failed: no such user, name: {name}, user_type: {user_type:?}"); Error::NoSuchUser(name.to_owned()) } else { warn!("load_user_identity failed: {err:?}, name: {name}, user_type: {user_type:?}"); err } })?; if u.credentials.is_expired() { let _ = self.delete_iam_config(get_user_identity_path(name, user_type)).await; let _ = self.delete_iam_config(get_mapped_policy_path(name, user_type, false)).await; warn!( "load_user_identity failed: user is expired, delete the user and mapped policy, name: {name}, user_type: {user_type:?}" ); return Err(Error::NoSuchUser(name.to_owned())); } if u.credentials.access_key.is_empty() { u.credentials.access_key = name.to_owned(); } if !u.credentials.session_token.is_empty() { match extract_jwt_claims(&u) { Ok(claims) => { u.credentials.claims = Some(claims); } Err(err) => { if u.credentials.is_temp() { let _ = self.delete_iam_config(get_user_identity_path(name, user_type)).await; let _ = self.delete_iam_config(get_mapped_policy_path(name, user_type, false)).await; } warn!("extract_jwt_claims failed: {err:?}, name: {name}, user_type: {user_type:?}"); return Err(Error::NoSuchUser(name.to_owned())); } } } Ok(u) } async fn load_user(&self, name: &str, user_type: UserType, m: &mut HashMap<String, UserIdentity>) -> Result<()> { self.load_user_identity(name, user_type).await.map(|u| { m.insert(name.to_owned(), u); }) } async fn load_users(&self, user_type: UserType, m: &mut HashMap<String, UserIdentity>) -> Result<()> { let base_prefix = match user_type { UserType::Reg => IAM_CONFIG_USERS_PREFIX.as_str(), UserType::Svc => IAM_CONFIG_SERVICE_ACCOUNTS_PREFIX.as_str(), UserType::Sts => IAM_CONFIG_STS_PREFIX.as_str(), UserType::None => "", }; let ctx = CancellationToken::new(); let (tx, mut rx) = mpsc::channel::<StringOrErr>(100); self.list_iam_config_items(base_prefix, ctx.clone(), tx).await; while let Some(v) = rx.recv().await { if let Some(err) = v.err { warn!("list_iam_config_items {:?}", err); let _ = ctx.cancel(); return Err(err); } if let Some(item) = v.item { let name = rustfs_utils::path::dir(&item); self.load_user(&name, user_type, m).await?; } } let _ = ctx.cancel(); Ok(()) } async fn load_secret_key(&self, name: &str, user_type: UserType) -> Result<String> { let u: UserIdentity = self .load_iam_config(get_user_identity_path(name, user_type)) .await .map_err(|err| { if is_err_config_not_found(&err) { Error::NoSuchUser(name.to_owned()) } else { err } })?; Ok(u.credentials.secret_key) } async fn save_group_info(&self, name: &str, item: GroupInfo) -> Result<()> { self.save_iam_config(item, get_group_info_path(name)).await } async fn delete_group_info(&self, name: &str) -> Result<()> { self.delete_iam_config(get_group_info_path(name)).await.map_err(|err| { if is_err_config_not_found(&err) { Error::NoSuchPolicy } else { err } })?; Ok(()) } async fn load_group(&self, name: &str, m: &mut HashMap<String, GroupInfo>) -> Result<()> { let u: GroupInfo = self.load_iam_config(get_group_info_path(name)).await.map_err(|err| { if is_err_config_not_found(&err) { Error::NoSuchPolicy } else { err } })?; m.insert(name.to_owned(), u); Ok(()) } async fn load_groups(&self, m: &mut HashMap<String, GroupInfo>) -> Result<()> { let ctx = CancellationToken::new(); let (tx, mut rx) = mpsc::channel::<StringOrErr>(100); self.list_iam_config_items(&IAM_CONFIG_GROUPS_PREFIX, ctx.clone(), tx).await; while let Some(v) = rx.recv().await { if let Some(err) = v.err { warn!("list_iam_config_items {:?}", err); let _ = ctx.cancel(); return Err(err); } if let Some(item) = v.item { let name = rustfs_utils::path::dir(&item); if let Err(err) = self.load_group(&name, m).await && !is_err_no_such_group(&err) { return Err(err); } } } let _ = ctx.cancel(); Ok(()) } async fn save_policy_doc(&self, name: &str, policy_doc: PolicyDoc) -> Result<()> { self.save_iam_config(policy_doc, get_policy_doc_path(name)).await } async fn delete_policy_doc(&self, name: &str) -> Result<()> { self.delete_iam_config(get_policy_doc_path(name)).await.map_err(|err| { if is_err_config_not_found(&err) { Error::NoSuchPolicy } else { err } })?; Ok(()) } async fn load_policy(&self, name: &str) -> Result<PolicyDoc> { let (data, obj) = self .load_iamconfig_bytes_with_metadata(get_policy_doc_path(name)) .await .map_err(|err| { if is_err_config_not_found(&err) { Error::NoSuchPolicy } else { err } })?; let mut info = PolicyDoc::try_from(data)?; if info.version == 0 { info.create_date = obj.mod_time; info.update_date = obj.mod_time; } Ok(info) } async fn load_policy_doc(&self, name: &str, m: &mut HashMap<String, PolicyDoc>) -> Result<()> { let info = self.load_policy(name).await?; m.insert(name.to_owned(), info); Ok(()) } async fn load_policy_docs(&self, m: &mut HashMap<String, PolicyDoc>) -> Result<()> { let ctx = CancellationToken::new(); let (tx, mut rx) = mpsc::channel::<StringOrErr>(100); self.list_iam_config_items(&IAM_CONFIG_POLICIES_PREFIX, ctx.clone(), tx).await; while let Some(v) = rx.recv().await { if let Some(err) = v.err { warn!("list_iam_config_items {:?}", err); let _ = ctx.cancel(); return Err(err); } if let Some(item) = v.item { let name = rustfs_utils::path::dir(&item); self.load_policy_doc(&name, m).await?; } } let _ = ctx.cancel(); Ok(()) } async fn save_mapped_policy( &self, name: &str, user_type: UserType, is_group: bool, mapped_policy: MappedPolicy, _ttl: Option<usize>, ) -> Result<()> { self.save_iam_config(mapped_policy, get_mapped_policy_path(name, user_type, is_group)) .await } async fn delete_mapped_policy(&self, name: &str, user_type: UserType, is_group: bool) -> Result<()> { self.delete_iam_config(get_mapped_policy_path(name, user_type, is_group)) .await .map_err(|err| { if is_err_config_not_found(&err) { Error::NoSuchPolicy } else { err } })?; Ok(()) } async fn load_mapped_policy( &self, name: &str, user_type: UserType, is_group: bool, m: &mut HashMap<String, MappedPolicy>, ) -> Result<()> { let info = self.load_mapped_policy_internal(name, user_type, is_group).await?; m.insert(name.to_owned(), info); Ok(()) } async fn load_mapped_policies( &self, user_type: UserType, is_group: bool, m: &mut HashMap<String, MappedPolicy>, ) -> Result<()> { let base_path = { if is_group { IAM_CONFIG_POLICY_DB_GROUPS_PREFIX.as_str() } else { match user_type { UserType::Svc => IAM_CONFIG_POLICY_DB_SERVICE_ACCOUNTS_PREFIX.as_str(), UserType::Sts => IAM_CONFIG_POLICY_DB_STS_USERS_PREFIX.as_str(), _ => IAM_CONFIG_POLICY_DB_USERS_PREFIX.as_str(), } } }; let ctx = CancellationToken::new(); let (tx, mut rx) = mpsc::channel::<StringOrErr>(100); self.list_iam_config_items(base_path, ctx.clone(), tx).await; while let Some(v) = rx.recv().await { if let Some(err) = v.err { warn!("list_iam_config_items {:?}", err); let _ = ctx.cancel(); return Err(err); } if let Some(item) = v.item { let name = item.trim_end_matches(".json"); self.load_mapped_policy(name, user_type, is_group, m).await?; } } let _ = ctx.cancel(); // TODO: check if this is needed Ok(()) } async fn load_all(&self, cache: &Cache) -> Result<()> { let listed_config_items = self.list_all_iamconfig_items().await?; let mut policy_docs_cache = CacheEntity::new(get_default_policyes()); if let Some(policies_list) = listed_config_items.get(POLICIES_LIST_KEY) { let mut policies_list = policies_list.clone(); loop { if policies_list.len() < 32 { let policy_docs = self.load_policy_doc_concurrent(&policies_list).await?; for (idx, p) in policy_docs.into_iter().enumerate() { if p.policy.version.is_empty() { continue; } let policy_name = rustfs_utils::path::dir(&policies_list[idx]); info!("load policy: {}", policy_name); policy_docs_cache.insert(policy_name, p); } break; } let policy_docs = self.load_policy_doc_concurrent(&policies_list).await?; for (idx, p) in policy_docs.into_iter().enumerate() { if p.policy.version.is_empty() { continue; } let policy_name = rustfs_utils::path::dir(&policies_list[idx]); info!("load policy: {}", policy_name); policy_docs_cache.insert(policy_name, p); } policies_list = policies_list.split_off(32); } } cache.policy_docs.store(Arc::new(policy_docs_cache.update_load_time())); let mut user_items_cache = CacheEntity::default(); // users if let Some(item_name_list) = listed_config_items.get(USERS_LIST_KEY) { let mut item_name_list = item_name_list.clone(); // let mut items_cache = CacheEntity::default(); loop { if item_name_list.len() < 32 { let items = self.load_user_concurrent(&item_name_list, UserType::Reg).await?; for (idx, p) in items.into_iter().enumerate() { if p.credentials.access_key.is_empty() { continue; } let name = rustfs_utils::path::dir(&item_name_list[idx]); info!("load reg user: {}", name); user_items_cache.insert(name, p); } break; } let items = self.load_user_concurrent(&item_name_list, UserType::Reg).await?; for (idx, p) in items.into_iter().enumerate() { if p.credentials.access_key.is_empty() { continue; } let name = rustfs_utils::path::dir(&item_name_list[idx]); info!("load reg user: {}", name); user_items_cache.insert(name, p); } item_name_list = item_name_list.split_off(32); } // cache.users.store(Arc::new(items_cache.update_load_time())); } // groups if let Some(item_name_list) = listed_config_items.get(GROUPS_LIST_KEY) { let mut items_cache = CacheEntity::default(); for item in item_name_list.iter() { let name = rustfs_utils::path::dir(item); info!("load group: {}", name); if let Err(err) = self.load_group(&name, &mut items_cache).await { return Err(Error::other(format!("load group failed: {err}"))); }; } cache.groups.store(Arc::new(items_cache.update_load_time())); } // user policies if let Some(item_name_list) = listed_config_items.get(POLICY_DB_USERS_LIST_KEY) { let mut item_name_list = item_name_list.clone(); let mut items_cache = CacheEntity::default(); loop {
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
true
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/kms/src/config.rs
crates/kms/src/config.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! KMS configuration management use crate::error::{KmsError, Result}; use serde::{Deserialize, Serialize}; use std::path::PathBuf; use std::time::Duration; use url::Url; /// KMS backend types #[derive(Debug, Default, Clone, Serialize, Deserialize, PartialEq, Eq)] pub enum KmsBackend { /// Vault backend (recommended for production) Vault, /// Local file-based backend for development and testing only #[default] Local, } /// Main KMS configuration #[derive(Debug, Clone, Serialize, Deserialize)] pub struct KmsConfig { /// Backend type pub backend: KmsBackend, /// Default master key ID for auto-encryption pub default_key_id: Option<String>, /// Backend-specific configuration pub backend_config: BackendConfig, /// Operation timeout pub timeout: Duration, /// Number of retry attempts pub retry_attempts: u32, /// Enable caching pub enable_cache: bool, /// Cache configuration pub cache_config: CacheConfig, } impl Default for KmsConfig { fn default() -> Self { Self { backend: KmsBackend::default(), default_key_id: None, backend_config: BackendConfig::default(), timeout: Duration::from_secs(30), retry_attempts: 3, enable_cache: true, cache_config: CacheConfig::default(), } } } /// Backend-specific configuration #[derive(Debug, Clone, Serialize, Deserialize)] pub enum BackendConfig { /// Local backend configuration Local(LocalConfig), /// Vault backend configuration Vault(VaultConfig), } impl Default for BackendConfig { fn default() -> Self { Self::Local(LocalConfig::default()) } } /// Local KMS backend configuration #[derive(Debug, Clone, Serialize, Deserialize)] pub struct LocalConfig { /// Directory to store key files pub key_dir: PathBuf, /// Master key for encrypting stored keys (if None, keys are stored in plaintext) pub master_key: Option<String>, /// File permissions for key files (octal) pub file_permissions: Option<u32>, } impl Default for LocalConfig { fn default() -> Self { Self { key_dir: std::env::temp_dir().join("rustfs_kms_keys"), master_key: None, file_permissions: Some(0o600), // Owner read/write only } } } /// Vault backend configuration #[derive(Debug, Clone, Serialize, Deserialize)] pub struct VaultConfig { /// Vault server URL pub address: String, /// Authentication method pub auth_method: VaultAuthMethod, /// Vault namespace (Vault Enterprise) pub namespace: Option<String>, /// Transit engine mount path pub mount_path: String, /// KV engine mount path for storing keys pub kv_mount: String, /// Path prefix for keys in KV store pub key_path_prefix: String, /// TLS configuration pub tls: Option<TlsConfig>, } impl Default for VaultConfig { fn default() -> Self { Self { address: "http://localhost:8200".to_string(), auth_method: VaultAuthMethod::Token { token: "dev-token".to_string(), }, namespace: None, mount_path: "transit".to_string(), kv_mount: "secret".to_string(), key_path_prefix: "rustfs/kms/keys".to_string(), tls: None, } } } /// Vault authentication methods #[derive(Debug, Clone, Serialize, Deserialize)] pub enum VaultAuthMethod { /// Token authentication Token { token: String }, /// AppRole authentication AppRole { role_id: String, secret_id: String }, } /// TLS configuration for Vault #[derive(Debug, Clone, Serialize, Deserialize)] pub struct TlsConfig { /// Path to CA certificate file pub ca_cert_path: Option<PathBuf>, /// Path to client certificate file pub client_cert_path: Option<PathBuf>, /// Path to client private key file pub client_key_path: Option<PathBuf>, /// Skip TLS verification (insecure, for development only) pub skip_verify: bool, } /// Cache configuration #[derive(Debug, Clone, Serialize, Deserialize)] pub struct CacheConfig { /// Maximum number of keys to cache pub max_keys: usize, /// TTL for cached keys pub ttl: Duration, /// Enable cache metrics pub enable_metrics: bool, } impl Default for CacheConfig { fn default() -> Self { Self { max_keys: 1000, ttl: Duration::from_secs(3600), // 1 hour enable_metrics: true, } } } impl KmsConfig { /// Create a new KMS configuration for local backend (for development and testing only) pub fn local(key_dir: PathBuf) -> Self { Self { backend: KmsBackend::Local, backend_config: BackendConfig::Local(LocalConfig { key_dir, ..Default::default() }), ..Default::default() } } /// Create a new KMS configuration for Vault backend with token authentication (recommended for production) pub fn vault(address: Url, token: String) -> Self { Self { backend: KmsBackend::Vault, backend_config: BackendConfig::Vault(VaultConfig { address: address.to_string(), auth_method: VaultAuthMethod::Token { token }, ..Default::default() }), ..Default::default() } } /// Create a new KMS configuration for Vault backend with AppRole authentication (recommended for production) pub fn vault_approle(address: Url, role_id: String, secret_id: String) -> Self { Self { backend: KmsBackend::Vault, backend_config: BackendConfig::Vault(VaultConfig { address: address.to_string(), auth_method: VaultAuthMethod::AppRole { role_id, secret_id }, ..Default::default() }), ..Default::default() } } /// Get the local configuration if backend is Local pub fn local_config(&self) -> Option<&LocalConfig> { match &self.backend_config { BackendConfig::Local(config) => Some(config), _ => None, } } /// Get the Vault configuration if backend is Vault pub fn vault_config(&self) -> Option<&VaultConfig> { match &self.backend_config { BackendConfig::Vault(config) => Some(config), _ => None, } } /// Set default key ID pub fn with_default_key(mut self, key_id: String) -> Self { self.default_key_id = Some(key_id); self } /// Set operation timeout pub fn with_timeout(mut self, timeout: Duration) -> Self { self.timeout = timeout; self } /// Enable or disable caching pub fn with_cache(mut self, enable: bool) -> Self { self.enable_cache = enable; self } /// Validate the configuration pub fn validate(&self) -> Result<()> { // Validate timeout if self.timeout.is_zero() { return Err(KmsError::configuration_error("Timeout must be greater than 0")); } // Validate retry attempts if self.retry_attempts == 0 { return Err(KmsError::configuration_error("Retry attempts must be greater than 0")); } // Validate backend-specific configuration match &self.backend_config { BackendConfig::Local(config) => { if !config.key_dir.is_absolute() { return Err(KmsError::configuration_error("Local key directory must be an absolute path")); } } BackendConfig::Vault(config) => { if !config.address.starts_with("http://") && !config.address.starts_with("https://") { return Err(KmsError::configuration_error("Vault address must use http or https scheme")); } if config.mount_path.is_empty() { return Err(KmsError::configuration_error("Vault mount path cannot be empty")); } // Validate TLS configuration if using HTTPS if config.address.starts_with("https://") && let Some(ref tls) = config.tls && !tls.skip_verify { // In production, we should have proper TLS configuration if tls.ca_cert_path.is_none() && tls.client_cert_path.is_none() { tracing::warn!("Using HTTPS without custom TLS configuration - relying on system CA"); } } } } // Validate cache configuration if self.enable_cache && self.cache_config.max_keys == 0 { return Err(KmsError::configuration_error("Cache max_keys must be greater than 0")); } Ok(()) } /// Load configuration from environment variables pub fn from_env() -> Result<Self> { let mut config = Self::default(); // Backend type if let Ok(backend_type) = std::env::var("RUSTFS_KMS_BACKEND") { config.backend = match backend_type.to_lowercase().as_str() { "local" => KmsBackend::Local, "vault" => KmsBackend::Vault, _ => return Err(KmsError::configuration_error(format!("Unknown KMS backend: {backend_type}"))), }; } // Default key ID if let Ok(key_id) = std::env::var("RUSTFS_KMS_DEFAULT_KEY_ID") { config.default_key_id = Some(key_id); } // Timeout if let Ok(timeout_str) = std::env::var("RUSTFS_KMS_TIMEOUT_SECS") { let timeout_secs = timeout_str .parse::<u64>() .map_err(|_| KmsError::configuration_error("Invalid timeout value"))?; config.timeout = Duration::from_secs(timeout_secs); } // Retry attempts if let Ok(retries_str) = std::env::var("RUSTFS_KMS_RETRY_ATTEMPTS") { config.retry_attempts = retries_str .parse() .map_err(|_| KmsError::configuration_error("Invalid retry attempts value"))?; } // Enable cache if let Ok(cache_str) = std::env::var("RUSTFS_KMS_ENABLE_CACHE") { config.enable_cache = cache_str.parse().unwrap_or(true); } // Backend-specific configuration match config.backend { KmsBackend::Local => { let key_dir = std::env::var("RUSTFS_KMS_LOCAL_KEY_DIR").unwrap_or_else(|_| "./kms_keys".to_string()); let master_key = std::env::var("RUSTFS_KMS_LOCAL_MASTER_KEY").ok(); config.backend_config = BackendConfig::Local(LocalConfig { key_dir: PathBuf::from(key_dir), master_key, file_permissions: Some(0o600), }); } KmsBackend::Vault => { let address = std::env::var("RUSTFS_KMS_VAULT_ADDRESS").unwrap_or_else(|_| "http://localhost:8200".to_string()); let token = std::env::var("RUSTFS_KMS_VAULT_TOKEN").unwrap_or_else(|_| "dev-token".to_string()); config.backend_config = BackendConfig::Vault(VaultConfig { address, auth_method: VaultAuthMethod::Token { token }, namespace: std::env::var("RUSTFS_KMS_VAULT_NAMESPACE").ok(), mount_path: std::env::var("RUSTFS_KMS_VAULT_MOUNT_PATH").unwrap_or_else(|_| "transit".to_string()), kv_mount: std::env::var("RUSTFS_KMS_VAULT_KV_MOUNT").unwrap_or_else(|_| "secret".to_string()), key_path_prefix: std::env::var("RUSTFS_KMS_VAULT_KEY_PREFIX") .unwrap_or_else(|_| "rustfs/kms/keys".to_string()), tls: None, }); } } config.validate()?; Ok(config) } } #[cfg(test)] mod tests { use super::*; use tempfile::TempDir; #[test] fn test_default_config() { let config = KmsConfig::default(); assert_eq!(config.backend, KmsBackend::Local); assert!(config.validate().is_ok()); } #[test] fn test_local_config() { let temp_dir = TempDir::new().expect("Failed to create temp dir"); let config = KmsConfig::local(temp_dir.path().to_path_buf()); assert_eq!(config.backend, KmsBackend::Local); assert!(config.validate().is_ok()); let local_config = config.local_config().expect("Should have local config"); assert_eq!(local_config.key_dir, temp_dir.path()); } #[test] fn test_vault_config() { let address = Url::parse("https://vault.example.com:8200").expect("Valid URL"); let config = KmsConfig::vault(address.clone(), "test-token".to_string()); assert_eq!(config.backend, KmsBackend::Vault); assert!(config.validate().is_ok()); let vault_config = config.vault_config().expect("Should have vault config"); assert_eq!(vault_config.address, address.as_str()); } #[test] fn test_config_validation() { let mut config = KmsConfig::default(); // Valid config assert!(config.validate().is_ok()); // Invalid timeout config.timeout = Duration::from_secs(0); assert!(config.validate().is_err()); // Reset timeout and test invalid retry attempts config.timeout = Duration::from_secs(30); config.retry_attempts = 0; assert!(config.validate().is_err()); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/kms/src/lib.rs
crates/kms/src/lib.rs
#![deny(clippy::unwrap_used)] // Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! # RustFS Key Management Service (KMS) //! //! This crate provides a comprehensive Key Management Service (KMS) for RustFS, //! supporting secure key generation, storage, and object encryption capabilities. //! //! ## Features //! //! - **Multiple Backends**: Local file storage and Vault (optional) //! - **Object Encryption**: Transparent S3-compatible object encryption //! - **Streaming Encryption**: Memory-efficient encryption for large files //! - **Key Management**: Full lifecycle management of encryption keys //! - **S3 Compatibility**: SSE-S3, SSE-KMS, and SSE-C encryption modes //! //! ## Architecture //! //! The KMS follows a three-layer key hierarchy: //! - **Master Keys**: Managed by KMS backends (Local/Vault) //! - **Data Encryption Keys (DEK)**: Generated per object, encrypted by master keys //! - **Object Data**: Encrypted using DEKs with AES-256-GCM or ChaCha20-Poly1305 //! //! ## Example //! //! ```rust,no_run //! use rustfs_kms::{KmsConfig, init_global_kms_service_manager}; //! use std::path::PathBuf; //! //! #[tokio::main] //! async fn main() -> Result<(), Box<dyn std::error::Error>> { //! // Initialize global KMS service manager //! let service_manager = init_global_kms_service_manager(); //! //! // Configure with local backend //! let config = KmsConfig::local(PathBuf::from("./kms_keys")); //! service_manager.configure(config).await?; //! //! // Start the KMS service //! service_manager.start().await?; //! //! Ok(()) //! } //! ``` // Core modules pub mod api_types; pub mod backends; mod cache; pub mod config; mod encryption; mod error; pub mod manager; pub mod service_manager; pub mod types; // Re-export public API pub use api_types::{ CacheSummary, ConfigureKmsRequest, ConfigureKmsResponse, ConfigureLocalKmsRequest, ConfigureVaultKmsRequest, KmsConfigSummary, KmsStatusResponse, StartKmsRequest, StartKmsResponse, StopKmsResponse, TagKeyRequest, TagKeyResponse, UntagKeyRequest, UntagKeyResponse, UpdateKeyDescriptionRequest, UpdateKeyDescriptionResponse, }; pub use config::*; pub use encryption::ObjectEncryptionService; pub use encryption::service::DataKey; pub use error::{KmsError, Result}; pub use manager::KmsManager; pub use service_manager::{ KmsServiceManager, KmsServiceStatus, get_global_encryption_service, get_global_kms_service_manager, init_global_kms_service_manager, }; pub use types::*; // For backward compatibility - these functions now delegate to the service manager /// Initialize global encryption service (backward compatibility) /// /// This function is now deprecated. Use `init_global_kms_service_manager` and configure via API instead. #[deprecated(note = "Use dynamic KMS configuration via service manager instead")] pub async fn init_global_services(_service: ObjectEncryptionService) -> Result<()> { // For backward compatibility only - not recommended for new code Ok(()) } /// Check if the global encryption service is initialized and healthy pub async fn is_encryption_service_healthy() -> bool { match get_global_encryption_service().await { Some(service) => service.health_check().await.is_ok(), None => false, } } /// Shutdown the global encryption service (backward compatibility) #[deprecated(note = "Use service manager shutdown instead")] pub fn shutdown_global_services() { // For backward compatibility only - service manager handles shutdown now tracing::info!("KMS global services shutdown requested (deprecated)"); } #[cfg(test)] mod tests { use super::*; use tempfile::TempDir; #[tokio::test] async fn test_global_service_lifecycle() { // Test service manager initialization let manager = init_global_kms_service_manager(); // Test initial status let status = manager.get_status().await; assert_eq!(status, KmsServiceStatus::NotConfigured); // Test configuration and start let temp_dir = TempDir::new().expect("Failed to create temp dir"); let config = KmsConfig::local(temp_dir.path().to_path_buf()); manager.configure(config).await.expect("Configuration should succeed"); manager.start().await.expect("Start should succeed"); // Test that encryption service is now available assert!(get_global_encryption_service().await.is_some()); // Test health check assert!(is_encryption_service_healthy().await); // Test stop manager.stop().await.expect("Stop should succeed"); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/kms/src/manager.rs
crates/kms/src/manager.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! KMS manager for handling key operations and backend coordination use crate::backends::KmsBackend; use crate::cache::KmsCache; use crate::config::KmsConfig; use crate::error::Result; use crate::types::{ CancelKeyDeletionRequest, CancelKeyDeletionResponse, CreateKeyRequest, CreateKeyResponse, DecryptRequest, DecryptResponse, DeleteKeyRequest, DeleteKeyResponse, DescribeKeyRequest, DescribeKeyResponse, EncryptRequest, EncryptResponse, GenerateDataKeyRequest, GenerateDataKeyResponse, ListKeysRequest, ListKeysResponse, }; use std::sync::Arc; use tokio::sync::RwLock; /// KMS Manager coordinates operations between backends and caching #[derive(Clone)] pub struct KmsManager { backend: Arc<dyn KmsBackend>, cache: Arc<RwLock<KmsCache>>, config: KmsConfig, } impl KmsManager { /// Create a new KMS manager with the given backend and config pub fn new(backend: Arc<dyn KmsBackend>, config: KmsConfig) -> Self { let cache = Arc::new(RwLock::new(KmsCache::new(config.cache_config.max_keys as u64))); Self { backend, cache, config } } /// Get the default key ID if configured pub fn get_default_key_id(&self) -> Option<&String> { self.config.default_key_id.as_ref() } /// Create a new master key pub async fn create_key(&self, request: CreateKeyRequest) -> Result<CreateKeyResponse> { let response = self.backend.create_key(request).await?; // Cache the key metadata if enabled if self.config.enable_cache { let mut cache = self.cache.write().await; cache.put_key_metadata(&response.key_id, &response.key_metadata).await; } Ok(response) } /// Encrypt data with a master key pub async fn encrypt(&self, request: EncryptRequest) -> Result<EncryptResponse> { self.backend.encrypt(request).await } /// Decrypt data with a master key pub async fn decrypt(&self, request: DecryptRequest) -> Result<DecryptResponse> { self.backend.decrypt(request).await } /// Generate a data encryption key pub async fn generate_data_key(&self, request: GenerateDataKeyRequest) -> Result<GenerateDataKeyResponse> { // Check cache first if enabled if self.config.enable_cache { let cache = self.cache.read().await; if let Some(cached_key) = cache.get_data_key(&request.key_id).await && cached_key.key_spec == request.key_spec { return Ok(GenerateDataKeyResponse { key_id: request.key_id.clone(), plaintext_key: cached_key.plaintext.clone(), ciphertext_blob: cached_key.ciphertext.clone(), }); } } // Generate new data key from backend let response = self.backend.generate_data_key(request).await?; // Cache the data key if enabled if self.config.enable_cache { let mut cache = self.cache.write().await; cache .put_data_key(&response.key_id, &response.plaintext_key, &response.ciphertext_blob) .await; } Ok(response) } /// Describe a key pub async fn describe_key(&self, request: DescribeKeyRequest) -> Result<DescribeKeyResponse> { // Check cache first if enabled if self.config.enable_cache { let cache = self.cache.read().await; if let Some(cached_metadata) = cache.get_key_metadata(&request.key_id).await { return Ok(DescribeKeyResponse { key_metadata: cached_metadata, }); } } // Get from backend and cache let response = self.backend.describe_key(request).await?; if self.config.enable_cache { let mut cache = self.cache.write().await; cache .put_key_metadata(&response.key_metadata.key_id, &response.key_metadata) .await; } Ok(response) } /// List keys pub async fn list_keys(&self, request: ListKeysRequest) -> Result<ListKeysResponse> { self.backend.list_keys(request).await } /// Get cache statistics pub async fn cache_stats(&self) -> Option<(u64, u64)> { if self.config.enable_cache { let cache = self.cache.read().await; Some(cache.stats()) } else { None } } /// Clear the cache pub async fn clear_cache(&self) -> Result<()> { if self.config.enable_cache { let mut cache = self.cache.write().await; cache.clear().await; } Ok(()) } /// Delete a key pub async fn delete_key(&self, request: DeleteKeyRequest) -> Result<DeleteKeyResponse> { let response = self.backend.delete_key(request).await?; // Remove from cache if enabled and key is being deleted if self.config.enable_cache { let mut cache = self.cache.write().await; cache.remove_key_metadata(&response.key_id).await; cache.remove_data_key(&response.key_id).await; } Ok(response) } /// Cancel key deletion pub async fn cancel_key_deletion(&self, request: CancelKeyDeletionRequest) -> Result<CancelKeyDeletionResponse> { let response = self.backend.cancel_key_deletion(request).await?; // Update cache if enabled if self.config.enable_cache { let mut cache = self.cache.write().await; cache.put_key_metadata(&response.key_id, &response.key_metadata).await; } Ok(response) } /// Perform health check on the KMS backend pub async fn health_check(&self) -> Result<bool> { self.backend.health_check().await } } #[cfg(test)] mod tests { use super::*; use crate::backends::local::LocalKmsBackend; use crate::types::{KeySpec, KeyState, KeyUsage}; use tempfile::tempdir; #[tokio::test] async fn test_manager_operations() { let temp_dir = tempdir().expect("Failed to create temp dir"); let config = KmsConfig::local(temp_dir.path().to_path_buf()); let backend = Arc::new(LocalKmsBackend::new(config.clone()).await.expect("Failed to create backend")); let manager = KmsManager::new(backend, config); // Test key creation let create_request = CreateKeyRequest { key_usage: KeyUsage::EncryptDecrypt, description: Some("Test key".to_string()), ..Default::default() }; let create_response = manager.create_key(create_request).await.expect("Failed to create key"); assert!(!create_response.key_id.is_empty()); assert_eq!(create_response.key_metadata.key_state, KeyState::Enabled); // Test data key generation let data_key_request = GenerateDataKeyRequest { key_id: create_response.key_id.clone(), key_spec: KeySpec::Aes256, encryption_context: Default::default(), }; let data_key_response = manager .generate_data_key(data_key_request) .await .expect("Failed to generate data key"); assert_eq!(data_key_response.plaintext_key.len(), 32); // 256 bits assert!(!data_key_response.ciphertext_blob.is_empty()); // Test describe key let describe_request = DescribeKeyRequest { key_id: create_response.key_id.clone(), }; let describe_response = manager.describe_key(describe_request).await.expect("Failed to describe key"); assert_eq!(describe_response.key_metadata.key_id, create_response.key_id); // Test cache stats let stats = manager.cache_stats().await; assert!(stats.is_some()); // Test health check let health = manager.health_check().await.expect("Health check failed"); assert!(health); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/kms/src/error.rs
crates/kms/src/error.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! KMS error types and result handling use thiserror::Error; /// Result type for KMS operations pub type Result<T> = std::result::Result<T, KmsError>; /// KMS error types covering all possible failure scenarios #[derive(Error, Debug, Clone)] pub enum KmsError { /// Configuration errors #[error("Configuration error: {message}")] ConfigurationError { message: String }, /// Key not found #[error("Key not found: {key_id}")] KeyNotFound { key_id: String }, /// Invalid key format or content #[error("Invalid key: {message}")] InvalidKey { message: String }, /// Cryptographic operation failed #[error("Cryptographic error in {operation}: {message}")] CryptographicError { operation: String, message: String }, /// Backend communication error #[error("Backend error: {message}")] BackendError { message: String }, /// Access denied #[error("Access denied: {message}")] AccessDenied { message: String }, /// Key already exists #[error("Key already exists: {key_id}")] KeyAlreadyExists { key_id: String }, /// Invalid operation state #[error("Invalid operation: {message}")] InvalidOperation { message: String }, /// Internal error #[error("Internal error: {message}")] InternalError { message: String }, /// Serialization/deserialization error #[error("Serialization error: {message}")] SerializationError { message: String }, /// I/O error #[error("I/O error: {message}")] IoError { message: String }, /// Cache error #[error("Cache error: {message}")] CacheError { message: String }, /// Validation error #[error("Validation error: {message}")] ValidationError { message: String }, /// Unsupported algorithm #[error("Unsupported algorithm: {algorithm}")] UnsupportedAlgorithm { algorithm: String }, /// Invalid key size #[error("Invalid key size: expected {expected}, got {actual}")] InvalidKeySize { expected: usize, actual: usize }, /// Encryption context mismatch #[error("Encryption context mismatch: {message}")] ContextMismatch { message: String }, } impl KmsError { /// Create a configuration error pub fn configuration_error<S: Into<String>>(message: S) -> Self { Self::ConfigurationError { message: message.into() } } /// Create a key not found error pub fn key_not_found<S: Into<String>>(key_id: S) -> Self { Self::KeyNotFound { key_id: key_id.into() } } /// Create an invalid key error pub fn invalid_key<S: Into<String>>(message: S) -> Self { Self::InvalidKey { message: message.into() } } /// Create a cryptographic error pub fn cryptographic_error<S1: Into<String>, S2: Into<String>>(operation: S1, message: S2) -> Self { Self::CryptographicError { operation: operation.into(), message: message.into(), } } /// Create a backend error pub fn backend_error<S: Into<String>>(message: S) -> Self { Self::BackendError { message: message.into() } } /// Create access denied error pub fn access_denied<S: Into<String>>(message: S) -> Self { Self::AccessDenied { message: message.into() } } /// Create a key already exists error pub fn key_already_exists<S: Into<String>>(key_id: S) -> Self { Self::KeyAlreadyExists { key_id: key_id.into() } } /// Create an invalid operation error pub fn invalid_operation<S: Into<String>>(message: S) -> Self { Self::InvalidOperation { message: message.into() } } /// Create an internal error pub fn internal_error<S: Into<String>>(message: S) -> Self { Self::InternalError { message: message.into() } } /// Create a serialization error pub fn serialization_error<S: Into<String>>(message: S) -> Self { Self::SerializationError { message: message.into() } } /// Create an I/O error pub fn io_error<S: Into<String>>(message: S) -> Self { Self::IoError { message: message.into() } } /// Create a cache error pub fn cache_error<S: Into<String>>(message: S) -> Self { Self::CacheError { message: message.into() } } /// Create a validation error pub fn validation_error<S: Into<String>>(message: S) -> Self { Self::ValidationError { message: message.into() } } /// Create an invalid parameter error pub fn invalid_parameter<S: Into<String>>(message: S) -> Self { Self::InvalidOperation { message: message.into() } } /// Create an invalid key state error pub fn invalid_key_state<S: Into<String>>(message: S) -> Self { Self::InvalidOperation { message: message.into() } } /// Create an unsupported algorithm error pub fn unsupported_algorithm<S: Into<String>>(algorithm: S) -> Self { Self::UnsupportedAlgorithm { algorithm: algorithm.into(), } } /// Create an invalid key size error pub fn invalid_key_size(expected: usize, actual: usize) -> Self { Self::InvalidKeySize { expected, actual } } /// Create an encryption context mismatch error pub fn context_mismatch<S: Into<String>>(message: S) -> Self { Self::ContextMismatch { message: message.into() } } } /// Convert from standard library errors impl From<std::io::Error> for KmsError { fn from(error: std::io::Error) -> Self { Self::IoError { message: error.to_string(), } } } impl From<serde_json::Error> for KmsError { fn from(error: serde_json::Error) -> Self { Self::SerializationError { message: error.to_string(), } } } // Note: We can't implement From for both aes_gcm::Error and chacha20poly1305::Error // because they might be the same type. Instead, we provide helper functions. impl KmsError { /// Create a KMS error from AES-GCM error /// /// #Arguments /// * `error` - The AES-GCM error to convert /// /// #Returns /// * `KmsError` - The corresponding KMS error /// pub fn from_aes_gcm_error(error: aes_gcm::Error) -> Self { Self::CryptographicError { operation: "AES-GCM".to_string(), message: error.to_string(), } } /// Create a KMS error from ChaCha20-Poly1305 error /// /// #Arguments /// * `error` - The ChaCha20-Poly1305 error to convert /// /// #Returns /// * `KmsError` - The corresponding KMS error /// pub fn from_chacha20_error(error: chacha20poly1305::Error) -> Self { Self::CryptographicError { operation: "ChaCha20-Poly1305".to_string(), message: error.to_string(), } } } impl From<url::ParseError> for KmsError { fn from(error: url::ParseError) -> Self { Self::ConfigurationError { message: format!("Invalid URL: {error}"), } } } impl From<reqwest::Error> for KmsError { fn from(error: reqwest::Error) -> Self { Self::BackendError { message: format!("HTTP request failed: {error}"), } } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/kms/src/types.rs
crates/kms/src/types.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Core type definitions for KMS operations use chrono::{DateTime, Utc}; use serde::{Deserialize, Serialize}; use std::collections::HashMap; use uuid::Uuid; use zeroize::Zeroize; /// Data encryption key (DEK) used for encrypting object data #[derive(Debug, Clone, Serialize, Deserialize)] pub struct DataKey { /// Key identifier pub key_id: String, /// Key version pub version: u32, /// Plaintext key material (only available during generation) /// SECURITY: This field is manually zeroed when dropped pub plaintext: Option<Vec<u8>>, /// Encrypted key material (ciphertext) pub ciphertext: Vec<u8>, /// Key algorithm specification pub key_spec: String, /// Associated metadata pub metadata: HashMap<String, String>, /// Key creation timestamp pub created_at: DateTime<Utc>, } impl DataKey { /// Create a new data key /// /// # Arguments /// * `key_id` - Unique identifier for the key /// * `version` - Key version number /// * `plaintext` - Optional plaintext key material /// * `ciphertext` - Encrypted key material /// * `key_spec` - Key specification (e.g., "AES_256") /// /// # Returns /// A new `DataKey` instance /// pub fn new(key_id: String, version: u32, plaintext: Option<Vec<u8>>, ciphertext: Vec<u8>, key_spec: String) -> Self { Self { key_id, version, plaintext, ciphertext, key_spec, metadata: HashMap::new(), created_at: Utc::now(), } } /// Clear the plaintext key material from memory for security /// /// # Security /// This method zeroes out the plaintext key material before dropping it /// to prevent sensitive data from lingering in memory. /// pub fn clear_plaintext(&mut self) { if let Some(ref mut plaintext) = self.plaintext { // Zero out the memory before dropping plaintext.zeroize(); } self.plaintext = None; } /// Add metadata to the data key /// /// # Arguments /// * `key` - Metadata key /// * `value` - Metadata value /// /// # Returns /// Updated `DataKey` instance with added metadata /// pub fn with_metadata(mut self, key: String, value: String) -> Self { self.metadata.insert(key, value); self } } /// Master key stored in KMS backend #[derive(Debug, Clone, Serialize, Deserialize)] pub struct MasterKey { /// Unique key identifier pub key_id: String, /// Key version pub version: u32, /// Key algorithm (e.g., "AES-256") pub algorithm: String, /// Key usage type pub usage: KeyUsage, /// Key status pub status: KeyStatus, /// Key description pub description: Option<String>, /// Associated metadata pub metadata: HashMap<String, String>, /// Key creation timestamp pub created_at: DateTime<Utc>, /// Key last rotation timestamp pub rotated_at: Option<DateTime<Utc>>, /// Key creator/owner pub created_by: Option<String>, } impl MasterKey { /// Create a new master key /// /// # Arguments /// * `key_id` - Unique identifier for the key /// * `algorithm` - Key algorithm (e.g., "AES-256") /// * `created_by` - Optional creator/owner of the key /// /// # Returns /// A new `MasterKey` instance /// pub fn new(key_id: String, algorithm: String, created_by: Option<String>) -> Self { Self { key_id, version: 1, algorithm, usage: KeyUsage::EncryptDecrypt, status: KeyStatus::Active, description: None, metadata: HashMap::new(), created_at: Utc::now(), rotated_at: None, created_by, } } /// Create a new master key with description /// /// # Arguments /// * `key_id` - Unique identifier for the key /// * `algorithm` - Key algorithm (e.g., "AES-256") /// * `created_by` - Optional creator/owner of the key /// * `description` - Optional key description /// /// # Returns /// A new `MasterKey` instance with description /// pub fn new_with_description( key_id: String, algorithm: String, created_by: Option<String>, description: Option<String>, ) -> Self { Self { key_id, version: 1, algorithm, usage: KeyUsage::EncryptDecrypt, status: KeyStatus::Active, description, metadata: HashMap::new(), created_at: Utc::now(), rotated_at: None, created_by, } } } /// Key usage enumeration #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Eq)] pub enum KeyUsage { /// For encrypting and decrypting data EncryptDecrypt, /// For signing and verifying data SignVerify, } /// Key status #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Eq)] pub enum KeyStatus { /// Key is active and can be used Active, /// Key is disabled and cannot be used for new operations Disabled, /// Key is pending deletion PendingDeletion, /// Key has been deleted Deleted, } /// Information about a key #[derive(Debug, Clone, Serialize, Deserialize)] pub struct KeyInfo { /// Key identifier pub key_id: String, /// Key description pub description: Option<String>, /// Key algorithm pub algorithm: String, /// Key usage pub usage: KeyUsage, /// Key status pub status: KeyStatus, /// Key version pub version: u32, /// Associated metadata pub metadata: HashMap<String, String>, /// Key tags pub tags: HashMap<String, String>, /// Key creation timestamp pub created_at: DateTime<Utc>, /// Key last rotation timestamp pub rotated_at: Option<DateTime<Utc>>, /// Key creator pub created_by: Option<String>, } impl From<MasterKey> for KeyInfo { fn from(master_key: MasterKey) -> Self { Self { key_id: master_key.key_id, description: master_key.description, algorithm: master_key.algorithm, usage: master_key.usage, status: master_key.status, version: master_key.version, metadata: master_key.metadata.clone(), tags: master_key.metadata, created_at: master_key.created_at, rotated_at: master_key.rotated_at, created_by: master_key.created_by, } } } /// Request to generate a new data key #[derive(Debug, Clone, Serialize, Deserialize)] pub struct GenerateKeyRequest { /// Master key ID to use for encryption pub master_key_id: String, /// Key specification (e.g., "AES_256") pub key_spec: String, /// Number of bytes for the key (optional, derived from key_spec) pub key_length: Option<u32>, /// Encryption context for additional authenticated data pub encryption_context: HashMap<String, String>, /// Grant tokens for authorization (future use) pub grant_tokens: Vec<String>, } impl GenerateKeyRequest { /// Create a new generate key request /// /// # Arguments /// * `master_key_id` - Master key ID to use for encryption /// * `key_spec` - Key specification (e.g., "AES_256") /// /// # Returns /// A new `GenerateKeyRequest` instance /// pub fn new(master_key_id: String, key_spec: String) -> Self { Self { master_key_id, key_spec, key_length: None, encryption_context: HashMap::new(), grant_tokens: Vec::new(), } } /// Add encryption context /// /// # Arguments /// * `key` - Context key /// * `value` - Context value /// /// # Returns /// Updated `GenerateKeyRequest` instance with added context /// pub fn with_context(mut self, key: String, value: String) -> Self { self.encryption_context.insert(key, value); self } /// Set key length explicitly /// /// # Arguments /// * `length` - Key length in bytes /// /// # Returns /// Updated `GenerateKeyRequest` instance with specified key length /// pub fn with_length(mut self, length: u32) -> Self { self.key_length = Some(length); self } } /// Request to encrypt data #[derive(Debug, Clone, Serialize, Deserialize)] pub struct EncryptRequest { /// Key ID to use for encryption pub key_id: String, /// Plaintext data to encrypt pub plaintext: Vec<u8>, /// Encryption context pub encryption_context: HashMap<String, String>, /// Grant tokens for authorization pub grant_tokens: Vec<String>, } impl EncryptRequest { /// Create a new encrypt request /// /// # Arguments /// * `key_id` - Key ID to use for encryption /// * `plaintext` - Plaintext data to encrypt /// /// # Returns /// A new `EncryptRequest` instance /// pub fn new(key_id: String, plaintext: Vec<u8>) -> Self { Self { key_id, plaintext, encryption_context: HashMap::new(), grant_tokens: Vec::new(), } } /// Add encryption context /// /// # Arguments /// * `key` - Context key /// * `value` - Context value /// /// # Returns /// Updated `EncryptRequest` instance with added context /// pub fn with_context(mut self, key: String, value: String) -> Self { self.encryption_context.insert(key, value); self } } /// Response from encrypt operation #[derive(Debug, Clone, Serialize, Deserialize)] pub struct EncryptResponse { /// Encrypted data pub ciphertext: Vec<u8>, /// Key ID used for encryption pub key_id: String, /// Key version used pub key_version: u32, /// Encryption algorithm used pub algorithm: String, } /// Request to decrypt data #[derive(Debug, Clone, Serialize, Deserialize)] pub struct DecryptRequest { /// Ciphertext to decrypt pub ciphertext: Vec<u8>, /// Encryption context (must match the context used during encryption) pub encryption_context: HashMap<String, String>, /// Grant tokens for authorization pub grant_tokens: Vec<String>, } impl DecryptRequest { /// Create a new decrypt request /// /// # Arguments /// * `ciphertext` - Ciphertext to decrypt /// /// # Returns /// A new `DecryptRequest` instance /// pub fn new(ciphertext: Vec<u8>) -> Self { Self { ciphertext, encryption_context: HashMap::new(), grant_tokens: Vec::new(), } } /// Add encryption context /// /// # Arguments /// * `key` - Context key /// * `value` - Context value /// /// # Returns /// Updated `DecryptRequest` instance with added context /// pub fn with_context(mut self, key: String, value: String) -> Self { self.encryption_context.insert(key, value); self } } /// Request to list keys #[derive(Debug, Clone, Serialize, Deserialize)] pub struct ListKeysRequest { /// Maximum number of keys to return pub limit: Option<u32>, /// Pagination marker pub marker: Option<String>, /// Filter by key usage pub usage_filter: Option<KeyUsage>, /// Filter by key status pub status_filter: Option<KeyStatus>, } impl Default for ListKeysRequest { fn default() -> Self { Self { limit: Some(100), marker: None, usage_filter: None, status_filter: None, } } } /// Response from list keys operation #[derive(Debug, Clone, Serialize, Deserialize)] pub struct ListKeysResponse { /// List of keys pub keys: Vec<KeyInfo>, /// Pagination marker for next page pub next_marker: Option<String>, /// Whether there are more keys available pub truncated: bool, } /// Operation context for auditing and access control #[derive(Debug, Clone, Serialize, Deserialize)] pub struct OperationContext { /// Operation ID for tracking pub operation_id: Uuid, /// User or service performing the operation pub principal: String, /// Source IP address pub source_ip: Option<String>, /// User agent pub user_agent: Option<String>, /// Additional context information pub additional_context: HashMap<String, String>, } impl OperationContext { /// Create a new operation context /// /// # Arguments /// * `principal` - User or service performing the operation /// /// # Returns /// A new `OperationContext` instance /// pub fn new(principal: String) -> Self { Self { operation_id: Uuid::new_v4(), principal, source_ip: None, user_agent: None, additional_context: HashMap::new(), } } /// Add additional context /// /// # Arguments /// * `key` - Context key /// * `value` - Context value /// /// # Returns /// Updated `OperationContext` instance with added context /// pub fn with_context(mut self, key: String, value: String) -> Self { self.additional_context.insert(key, value); self } /// Set source IP /// /// # Arguments /// * `ip` - Source IP address /// /// # Returns /// Updated `OperationContext` instance with source IP /// pub fn with_source_ip(mut self, ip: String) -> Self { self.source_ip = Some(ip); self } /// Set user agent /// /// # Arguments /// * `agent` - User agent string /// /// # Returns /// Updated `OperationContext` instance with user agent /// pub fn with_user_agent(mut self, agent: String) -> Self { self.user_agent = Some(agent); self } } /// Object encryption context #[derive(Debug, Clone, Serialize, Deserialize)] pub struct ObjectEncryptionContext { /// Bucket name pub bucket: String, /// Object key pub object_key: String, /// Content type pub content_type: Option<String>, /// Object size in bytes pub size: Option<u64>, /// Additional encryption context pub encryption_context: HashMap<String, String>, } impl ObjectEncryptionContext { /// Create a new object encryption context /// /// # Arguments /// * `bucket` - Bucket name /// * `object_key` - Object key /// /// # Returns /// A new `ObjectEncryptionContext` instance /// pub fn new(bucket: String, object_key: String) -> Self { Self { bucket, object_key, content_type: None, size: None, encryption_context: HashMap::new(), } } /// Set content type /// /// # Arguments /// * `content_type` - Content type string /// /// # Returns /// Updated `ObjectEncryptionContext` instance with content type /// pub fn with_content_type(mut self, content_type: String) -> Self { self.content_type = Some(content_type); self } /// Set object size /// /// # Arguments /// * `size` - Object size in bytes /// /// # Returns /// Updated `ObjectEncryptionContext` instance with size /// pub fn with_size(mut self, size: u64) -> Self { self.size = Some(size); self } /// Add encryption context /// /// # Arguments /// * `key` - Context key /// * `value` - Context value /// /// # Returns /// Updated `ObjectEncryptionContext` instance with added context /// pub fn with_encryption_context(mut self, key: String, value: String) -> Self { self.encryption_context.insert(key, value); self } } /// Encryption metadata stored with encrypted objects #[derive(Debug, Clone, Serialize, Deserialize)] pub struct EncryptionMetadata { /// Encryption algorithm used pub algorithm: String, /// Key ID used for encryption pub key_id: String, /// Key version pub key_version: u32, /// Initialization vector pub iv: Vec<u8>, /// Authentication tag (for AEAD ciphers) pub tag: Option<Vec<u8>>, /// Encryption context pub encryption_context: HashMap<String, String>, /// Timestamp when encrypted pub encrypted_at: DateTime<Utc>, /// Size of original data pub original_size: u64, /// Encrypted data key pub encrypted_data_key: Vec<u8>, } /// Health status information #[derive(Debug, Clone, Serialize, Deserialize)] pub struct HealthStatus { /// Whether the KMS backend is healthy pub kms_healthy: bool, /// Whether encryption/decryption operations are working pub encryption_working: bool, /// Backend type (e.g., "local", "vault") pub backend_type: String, /// Additional health details pub details: HashMap<String, String>, } /// Supported encryption algorithms #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Eq)] pub enum EncryptionAlgorithm { /// AES-256-GCM #[serde(rename = "AES256")] Aes256, /// ChaCha20-Poly1305 #[serde(rename = "ChaCha20Poly1305")] ChaCha20Poly1305, /// AWS KMS managed encryption #[serde(rename = "aws:kms")] AwsKms, } /// Key specification for data keys #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Eq)] pub enum KeySpec { /// AES-256 key (32 bytes) Aes256, /// AES-128 key (16 bytes) Aes128, /// ChaCha20 key (32 bytes) ChaCha20, } impl KeySpec { /// Get the key size in bytes /// /// # Returns /// Key size in bytes /// pub fn key_size(&self) -> usize { match self { Self::Aes256 => 32, Self::Aes128 => 16, Self::ChaCha20 => 32, } } /// Get the string representation for backends /// /// # Returns /// Key specification as a string /// pub fn as_str(&self) -> &'static str { match self { Self::Aes256 => "AES_256", Self::Aes128 => "AES_128", Self::ChaCha20 => "ChaCha20", } } } /// Key metadata information #[derive(Debug, Clone, Serialize, Deserialize)] pub struct KeyMetadata { /// Key identifier pub key_id: String, /// Key state pub key_state: KeyState, /// Key usage type pub key_usage: KeyUsage, /// Key description pub description: Option<String>, /// Key creation timestamp pub creation_date: DateTime<Utc>, /// Key deletion timestamp pub deletion_date: Option<DateTime<Utc>>, /// Key origin pub origin: String, /// Key manager pub key_manager: String, /// Key tags pub tags: HashMap<String, String>, } /// Key state enumeration #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Eq)] pub enum KeyState { /// Key is enabled and can be used Enabled, /// Key is disabled Disabled, /// Key is pending deletion PendingDeletion, /// Key is pending import PendingImport, /// Key is unavailable Unavailable, } /// Request to create a new key #[derive(Debug, Clone, Serialize, Deserialize)] pub struct CreateKeyRequest { /// Custom key name (optional, will auto-generate UUID if not provided) pub key_name: Option<String>, /// Key usage type pub key_usage: KeyUsage, /// Key description pub description: Option<String>, /// Key policy pub policy: Option<String>, /// Tags for the key pub tags: HashMap<String, String>, /// Origin of the key pub origin: Option<String>, } impl Default for CreateKeyRequest { fn default() -> Self { Self { key_name: None, key_usage: KeyUsage::EncryptDecrypt, description: None, policy: None, tags: HashMap::new(), origin: None, } } } /// Response from create key operation #[derive(Debug, Clone, Serialize, Deserialize)] pub struct CreateKeyResponse { /// Created key ID pub key_id: String, /// Key metadata pub key_metadata: KeyMetadata, } /// Response from decrypt operation #[derive(Debug, Clone, Serialize, Deserialize)] pub struct DecryptResponse { /// Decrypted plaintext pub plaintext: Vec<u8>, /// Key ID used for decryption pub key_id: String, /// Encryption algorithm used pub encryption_algorithm: Option<String>, } /// Request to describe a key #[derive(Debug, Clone, Serialize, Deserialize)] pub struct DescribeKeyRequest { /// Key ID to describe pub key_id: String, } /// Response from describe key operation #[derive(Debug, Clone, Serialize, Deserialize)] pub struct DescribeKeyResponse { /// Key metadata pub key_metadata: KeyMetadata, } /// Request to generate a data key #[derive(Debug, Clone, Serialize, Deserialize)] pub struct GenerateDataKeyRequest { /// Key ID to use for encryption pub key_id: String, /// Key specification pub key_spec: KeySpec, /// Encryption context pub encryption_context: HashMap<String, String>, } impl GenerateDataKeyRequest { /// Create a new generate data key request /// /// # Arguments /// * `key_id` - Key ID to use for encryption /// * `key_spec` - Key specification /// /// # Returns /// A new `GenerateDataKeyRequest` instance /// pub fn new(key_id: String, key_spec: KeySpec) -> Self { Self { key_id, key_spec, encryption_context: HashMap::new(), } } } /// Response from generate data key operation #[derive(Debug, Clone, Serialize, Deserialize)] pub struct GenerateDataKeyResponse { /// Key ID used pub key_id: String, /// Plaintext data key pub plaintext_key: Vec<u8>, /// Encrypted data key pub ciphertext_blob: Vec<u8>, } impl EncryptionAlgorithm { /// Get the algorithm name as a string /// /// # Returns /// Algorithm name as a string /// pub fn as_str(&self) -> &'static str { match self { Self::Aes256 => "AES256", Self::ChaCha20Poly1305 => "ChaCha20Poly1305", Self::AwsKms => "aws:kms", } } /// Get the key size in bytes for this algorithm pub fn key_size(&self) -> usize { match self { Self::Aes256 => 32, // 256 bits Self::ChaCha20Poly1305 => 32, // 256 bits Self::AwsKms => 32, // 256 bits (uses AES-256 internally) } } /// Get the IV size in bytes for this algorithm pub fn iv_size(&self) -> usize { match self { Self::Aes256 => 12, // 96 bits for GCM Self::ChaCha20Poly1305 => 12, // 96 bits Self::AwsKms => 12, // 96 bits (uses AES-256-GCM internally) } } } impl std::str::FromStr for EncryptionAlgorithm { type Err = (); fn from_str(s: &str) -> Result<Self, Self::Err> { match s { "AES256" => Ok(Self::Aes256), "ChaCha20Poly1305" => Ok(Self::ChaCha20Poly1305), "aws:kms" => Ok(Self::AwsKms), _ => Err(()), } } } /// Request to delete a key #[derive(Debug, Clone, Serialize, Deserialize)] pub struct DeleteKeyRequest { /// Key ID to delete pub key_id: String, /// Number of days to wait before deletion (7-30 days, optional) pub pending_window_in_days: Option<u32>, /// Force immediate deletion (for development/testing only) pub force_immediate: Option<bool>, } /// Response from delete key operation #[derive(Debug, Clone, Serialize, Deserialize)] pub struct DeleteKeyResponse { /// Key ID that was deleted or scheduled for deletion pub key_id: String, /// Deletion date (if scheduled) pub deletion_date: Option<String>, /// Key metadata pub key_metadata: KeyMetadata, } /// Request to cancel key deletion #[derive(Debug, Clone, Serialize, Deserialize)] pub struct CancelKeyDeletionRequest { /// Key ID to cancel deletion for pub key_id: String, } /// Response from cancel key deletion operation #[derive(Debug, Clone, Serialize, Deserialize)] pub struct CancelKeyDeletionResponse { /// Key ID pub key_id: String, /// Key metadata pub key_metadata: KeyMetadata, } // SECURITY: Implement Drop to automatically zero sensitive data when DataKey is dropped impl Drop for DataKey { fn drop(&mut self) { self.clear_plaintext(); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/kms/src/service_manager.rs
crates/kms/src/service_manager.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! KMS service manager for dynamic configuration and runtime management use crate::backends::{KmsBackend, local::LocalKmsBackend}; use crate::config::{BackendConfig, KmsConfig}; use crate::encryption::service::ObjectEncryptionService; use crate::error::{KmsError, Result}; use crate::manager::KmsManager; use std::sync::{Arc, OnceLock}; use tokio::sync::RwLock; use tracing::{error, info, warn}; /// KMS service status #[derive(Debug, Clone, PartialEq, serde::Serialize, serde::Deserialize)] pub enum KmsServiceStatus { /// KMS is not configured NotConfigured, /// KMS is configured but not running Configured, /// KMS is running Running, /// KMS encountered an error Error(String), } /// Dynamic KMS service manager pub struct KmsServiceManager { /// Current KMS manager (if running) manager: Arc<RwLock<Option<Arc<KmsManager>>>>, /// Current encryption service (if running) encryption_service: Arc<RwLock<Option<Arc<ObjectEncryptionService>>>>, /// Current configuration config: Arc<RwLock<Option<KmsConfig>>>, /// Current status status: Arc<RwLock<KmsServiceStatus>>, } impl KmsServiceManager { /// Create a new KMS service manager (not configured) pub fn new() -> Self { Self { manager: Arc::new(RwLock::new(None)), encryption_service: Arc::new(RwLock::new(None)), config: Arc::new(RwLock::new(None)), status: Arc::new(RwLock::new(KmsServiceStatus::NotConfigured)), } } /// Get current service status pub async fn get_status(&self) -> KmsServiceStatus { self.status.read().await.clone() } /// Get current configuration (if any) pub async fn get_config(&self) -> Option<KmsConfig> { self.config.read().await.clone() } /// Configure KMS with new configuration pub async fn configure(&self, new_config: KmsConfig) -> Result<()> { // Update configuration { let mut config = self.config.write().await; *config = Some(new_config.clone()); } // Update status { let mut status = self.status.write().await; *status = KmsServiceStatus::Configured; } info!("KMS configuration updated successfully"); Ok(()) } /// Start KMS service with current configuration pub async fn start(&self) -> Result<()> { let config = { let config_guard = self.config.read().await; match config_guard.as_ref() { Some(config) => config.clone(), None => { let err_msg = "Cannot start KMS: no configuration provided"; error!("{}", err_msg); let mut status = self.status.write().await; *status = KmsServiceStatus::Error(err_msg.to_string()); return Err(KmsError::configuration_error(err_msg)); } } }; info!("Starting KMS service with backend: {:?}", config.backend); match self.create_backend(&config).await { Ok(backend) => { // Create KMS manager let kms_manager = Arc::new(KmsManager::new(backend, config)); // Create encryption service let encryption_service = Arc::new(ObjectEncryptionService::new((*kms_manager).clone())); // Update manager and service { let mut manager = self.manager.write().await; *manager = Some(kms_manager); } { let mut service = self.encryption_service.write().await; *service = Some(encryption_service); } // Update status { let mut status = self.status.write().await; *status = KmsServiceStatus::Running; } info!("KMS service started successfully"); Ok(()) } Err(e) => { let err_msg = format!("Failed to create KMS backend: {e}"); error!("{}", err_msg); let mut status = self.status.write().await; *status = KmsServiceStatus::Error(err_msg.clone()); Err(KmsError::backend_error(&err_msg)) } } } /// Stop KMS service pub async fn stop(&self) -> Result<()> { info!("Stopping KMS service"); // Clear manager and service { let mut manager = self.manager.write().await; *manager = None; } { let mut service = self.encryption_service.write().await; *service = None; } // Update status (keep configuration) { let mut status = self.status.write().await; if !matches!(*status, KmsServiceStatus::NotConfigured) { *status = KmsServiceStatus::Configured; } } info!("KMS service stopped successfully"); Ok(()) } /// Reconfigure and restart KMS service pub async fn reconfigure(&self, new_config: KmsConfig) -> Result<()> { info!("Reconfiguring KMS service"); // Stop current service if running if matches!(self.get_status().await, KmsServiceStatus::Running) { self.stop().await?; } // Configure with new config self.configure(new_config).await?; // Start with new configuration self.start().await?; info!("KMS service reconfigured successfully"); Ok(()) } /// Get KMS manager (if running) pub async fn get_manager(&self) -> Option<Arc<KmsManager>> { self.manager.read().await.clone() } /// Get encryption service (if running) pub async fn get_encryption_service(&self) -> Option<Arc<ObjectEncryptionService>> { self.encryption_service.read().await.clone() } /// Health check for the KMS service pub async fn health_check(&self) -> Result<bool> { let manager = self.get_manager().await; match manager { Some(manager) => { // Perform health check on the backend match manager.health_check().await { Ok(healthy) => { if !healthy { warn!("KMS backend health check failed"); } Ok(healthy) } Err(e) => { error!("KMS health check error: {}", e); // Update status to error let mut status = self.status.write().await; *status = KmsServiceStatus::Error(format!("Health check failed: {e}")); Err(e) } } } None => { warn!("Cannot perform health check: KMS service not running"); Ok(false) } } } /// Create backend from configuration async fn create_backend(&self, config: &KmsConfig) -> Result<Arc<dyn KmsBackend>> { match &config.backend_config { BackendConfig::Local(_) => { info!("Creating Local KMS backend"); let backend = LocalKmsBackend::new(config.clone()).await?; Ok(Arc::new(backend)) } BackendConfig::Vault(_) => { info!("Creating Vault KMS backend"); let backend = crate::backends::vault::VaultKmsBackend::new(config.clone()).await?; Ok(Arc::new(backend)) } } } } impl Default for KmsServiceManager { fn default() -> Self { Self::new() } } /// Global KMS service manager instance static GLOBAL_KMS_SERVICE_MANAGER: OnceLock<Arc<KmsServiceManager>> = OnceLock::new(); /// Initialize global KMS service manager pub fn init_global_kms_service_manager() -> Arc<KmsServiceManager> { GLOBAL_KMS_SERVICE_MANAGER .get_or_init(|| Arc::new(KmsServiceManager::new())) .clone() } /// Get global KMS service manager pub fn get_global_kms_service_manager() -> Option<Arc<KmsServiceManager>> { GLOBAL_KMS_SERVICE_MANAGER.get().cloned() } /// Get global encryption service (if KMS is running) pub async fn get_global_encryption_service() -> Option<Arc<ObjectEncryptionService>> { let manager = get_global_kms_service_manager().unwrap_or_else(init_global_kms_service_manager); manager.get_encryption_service().await }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/kms/src/api_types.rs
crates/kms/src/api_types.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! API types for KMS dynamic configuration use crate::config::{BackendConfig, CacheConfig, KmsBackend, KmsConfig, LocalConfig, TlsConfig, VaultAuthMethod, VaultConfig}; use crate::service_manager::KmsServiceStatus; use crate::types::{KeyMetadata, KeyUsage}; use serde::{Deserialize, Serialize}; use std::collections::HashMap; use std::path::PathBuf; use std::time::Duration; /// Request to configure KMS with Local backend #[derive(Debug, Clone, Serialize, Deserialize)] pub struct ConfigureLocalKmsRequest { /// Directory to store key files pub key_dir: PathBuf, /// Master key for encrypting stored keys (optional) pub master_key: Option<String>, /// File permissions for key files (octal, optional) pub file_permissions: Option<u32>, /// Default master key ID for auto-encryption pub default_key_id: Option<String>, /// Operation timeout in seconds pub timeout_seconds: Option<u64>, /// Number of retry attempts pub retry_attempts: Option<u32>, /// Enable caching pub enable_cache: Option<bool>, /// Maximum number of keys to cache pub max_cached_keys: Option<usize>, /// Cache TTL in seconds pub cache_ttl_seconds: Option<u64>, } /// Request to configure KMS with Vault backend #[derive(Debug, Clone, Serialize, Deserialize)] pub struct ConfigureVaultKmsRequest { /// Vault server URL pub address: String, /// Authentication method pub auth_method: VaultAuthMethod, /// Vault namespace (Vault Enterprise, optional) pub namespace: Option<String>, /// Transit engine mount path pub mount_path: Option<String>, /// KV engine mount path for storing keys pub kv_mount: Option<String>, /// Path prefix for keys in KV store pub key_path_prefix: Option<String>, /// Skip TLS verification (insecure, for development only) pub skip_tls_verify: Option<bool>, /// Default master key ID for auto-encryption pub default_key_id: Option<String>, /// Operation timeout in seconds pub timeout_seconds: Option<u64>, /// Number of retry attempts pub retry_attempts: Option<u32>, /// Enable caching pub enable_cache: Option<bool>, /// Maximum number of keys to cache pub max_cached_keys: Option<usize>, /// Cache TTL in seconds pub cache_ttl_seconds: Option<u64>, } /// Generic KMS configuration request #[derive(Debug, Clone, Serialize, Deserialize)] #[serde(tag = "backend_type", rename_all = "lowercase")] pub enum ConfigureKmsRequest { /// Configure with Local backend Local(ConfigureLocalKmsRequest), /// Configure with Vault backend Vault(ConfigureVaultKmsRequest), } /// KMS configuration response #[derive(Debug, Clone, Serialize, Deserialize)] pub struct ConfigureKmsResponse { /// Whether configuration was successful pub success: bool, /// Status message pub message: String, /// New service status pub status: KmsServiceStatus, } /// Request to start KMS service #[derive(Debug, Clone, Serialize, Deserialize)] pub struct StartKmsRequest { /// Whether to force start (restart if already running) pub force: Option<bool>, } /// KMS start response #[derive(Debug, Clone, Serialize, Deserialize)] pub struct StartKmsResponse { /// Whether start was successful pub success: bool, /// Status message pub message: String, /// New service status pub status: KmsServiceStatus, } /// KMS stop response #[derive(Debug, Clone, Serialize, Deserialize)] pub struct StopKmsResponse { /// Whether stop was successful pub success: bool, /// Status message pub message: String, /// New service status pub status: KmsServiceStatus, } /// KMS status response #[derive(Debug, Clone, Serialize, Deserialize)] pub struct KmsStatusResponse { /// Current service status pub status: KmsServiceStatus, /// Current backend type (if configured) pub backend_type: Option<KmsBackend>, /// Whether KMS is healthy (if running) pub healthy: Option<bool>, /// Configuration summary (if configured) pub config_summary: Option<KmsConfigSummary>, } /// Summary of KMS configuration (without sensitive data) #[derive(Debug, Clone, Serialize, Deserialize)] pub struct KmsConfigSummary { /// Backend type pub backend_type: KmsBackend, /// Default key ID (if configured) pub default_key_id: Option<String>, /// Operation timeout in seconds pub timeout_seconds: u64, /// Number of retry attempts pub retry_attempts: u32, /// Whether caching is enabled pub enable_cache: bool, /// Cache configuration summary pub cache_summary: Option<CacheSummary>, /// Backend-specific summary pub backend_summary: BackendSummary, } /// Cache configuration summary #[derive(Debug, Clone, Serialize, Deserialize)] pub struct CacheSummary { /// Maximum number of keys to cache pub max_keys: usize, /// Cache TTL in seconds pub ttl_seconds: u64, /// Whether cache metrics are enabled pub enable_metrics: bool, } /// Backend-specific configuration summary #[derive(Debug, Clone, Serialize, Deserialize)] #[serde(tag = "backend_type", rename_all = "lowercase")] pub enum BackendSummary { /// Local backend summary Local { /// Key directory path key_dir: PathBuf, /// Whether master key is configured has_master_key: bool, /// File permissions (octal) file_permissions: Option<u32>, }, /// Vault backend summary Vault { /// Vault server address address: String, /// Authentication method type auth_method_type: String, /// Namespace (if configured) namespace: Option<String>, /// Transit engine mount path mount_path: String, /// KV engine mount path kv_mount: String, /// Key path prefix key_path_prefix: String, }, } impl From<&KmsConfig> for KmsConfigSummary { fn from(config: &KmsConfig) -> Self { let cache_summary = if config.enable_cache { Some(CacheSummary { max_keys: config.cache_config.max_keys, ttl_seconds: config.cache_config.ttl.as_secs(), enable_metrics: config.cache_config.enable_metrics, }) } else { None }; let backend_summary = match &config.backend_config { BackendConfig::Local(local_config) => BackendSummary::Local { key_dir: local_config.key_dir.clone(), has_master_key: local_config.master_key.is_some(), file_permissions: local_config.file_permissions, }, BackendConfig::Vault(vault_config) => BackendSummary::Vault { address: vault_config.address.clone(), auth_method_type: match &vault_config.auth_method { VaultAuthMethod::Token { .. } => "token".to_string(), VaultAuthMethod::AppRole { .. } => "approle".to_string(), }, namespace: vault_config.namespace.clone(), mount_path: vault_config.mount_path.clone(), kv_mount: vault_config.kv_mount.clone(), key_path_prefix: vault_config.key_path_prefix.clone(), }, }; Self { backend_type: config.backend.clone(), default_key_id: config.default_key_id.clone(), timeout_seconds: config.timeout.as_secs(), retry_attempts: config.retry_attempts, enable_cache: config.enable_cache, cache_summary, backend_summary, } } } impl ConfigureLocalKmsRequest { /// Convert to KmsConfig pub fn to_kms_config(&self) -> KmsConfig { KmsConfig { backend: KmsBackend::Local, default_key_id: self.default_key_id.clone(), backend_config: BackendConfig::Local(LocalConfig { key_dir: self.key_dir.clone(), master_key: self.master_key.clone(), file_permissions: self.file_permissions, }), timeout: Duration::from_secs(self.timeout_seconds.unwrap_or(30)), retry_attempts: self.retry_attempts.unwrap_or(3), enable_cache: self.enable_cache.unwrap_or(true), cache_config: CacheConfig { max_keys: self.max_cached_keys.unwrap_or(1000), ttl: Duration::from_secs(self.cache_ttl_seconds.unwrap_or(3600)), enable_metrics: true, }, } } } impl ConfigureVaultKmsRequest { /// Convert to KmsConfig pub fn to_kms_config(&self) -> KmsConfig { KmsConfig { backend: KmsBackend::Vault, default_key_id: self.default_key_id.clone(), backend_config: BackendConfig::Vault(VaultConfig { address: self.address.clone(), auth_method: self.auth_method.clone(), namespace: self.namespace.clone(), mount_path: self.mount_path.clone().unwrap_or_else(|| "transit".to_string()), kv_mount: self.kv_mount.clone().unwrap_or_else(|| "secret".to_string()), key_path_prefix: self.key_path_prefix.clone().unwrap_or_else(|| "rustfs/kms/keys".to_string()), tls: if self.skip_tls_verify.unwrap_or(false) { Some(TlsConfig { ca_cert_path: None, client_cert_path: None, client_key_path: None, skip_verify: true, }) } else { None }, }), timeout: Duration::from_secs(self.timeout_seconds.unwrap_or(30)), retry_attempts: self.retry_attempts.unwrap_or(3), enable_cache: self.enable_cache.unwrap_or(true), cache_config: CacheConfig { max_keys: self.max_cached_keys.unwrap_or(1000), ttl: Duration::from_secs(self.cache_ttl_seconds.unwrap_or(3600)), enable_metrics: true, }, } } } impl ConfigureKmsRequest { /// Convert to KmsConfig pub fn to_kms_config(&self) -> KmsConfig { match self { ConfigureKmsRequest::Local(req) => req.to_kms_config(), ConfigureKmsRequest::Vault(req) => req.to_kms_config(), } } } // ======================================== // Key Management API Types // ======================================== /// Request to create a new key with optional custom name #[derive(Debug, Clone, Serialize, Deserialize)] pub struct CreateKeyRequest { /// Custom key name (optional, will auto-generate UUID if not provided) pub key_name: Option<String>, /// Key usage type pub key_usage: KeyUsage, /// Key description pub description: Option<String>, /// Key policy JSON string pub policy: Option<String>, /// Tags for the key pub tags: HashMap<String, String>, /// Origin of the key pub origin: Option<String>, } impl Default for CreateKeyRequest { fn default() -> Self { Self { key_name: None, key_usage: KeyUsage::EncryptDecrypt, description: None, policy: None, tags: HashMap::new(), origin: None, } } } /// Response from create key operation #[derive(Debug, Clone, Serialize, Deserialize)] pub struct CreateKeyResponse { /// Success flag pub success: bool, /// Status message pub message: String, /// Created key ID (either custom name or auto-generated UUID) pub key_id: String, /// Key metadata pub key_metadata: KeyMetadata, } /// Request to delete a key #[derive(Debug, Clone, Serialize, Deserialize)] pub struct DeleteKeyRequest { /// Key ID to delete pub key_id: String, /// Number of days to wait before deletion (7-30 days, optional) pub pending_window_in_days: Option<u32>, /// Force immediate deletion (for development/testing only) pub force_immediate: Option<bool>, } /// Response from delete key operation #[derive(Debug, Clone, Serialize, Deserialize)] pub struct DeleteKeyResponse { /// Success flag pub success: bool, /// Status message pub message: String, /// Key ID that was deleted or scheduled for deletion pub key_id: String, /// Deletion date (if scheduled) pub deletion_date: Option<String>, } /// Request to list all keys #[derive(Debug, Clone, Serialize, Deserialize)] pub struct ListKeysRequest { /// Maximum number of keys to return (1-1000) pub limit: Option<u32>, /// Pagination marker pub marker: Option<String>, } /// Response from list keys operation #[derive(Debug, Clone, Serialize, Deserialize)] pub struct ListKeysResponse { /// Success flag pub success: bool, /// Status message pub message: String, /// List of key IDs pub keys: Vec<String>, /// Whether more keys are available pub truncated: bool, /// Next marker for pagination pub next_marker: Option<String>, } /// Request to describe a key #[derive(Debug, Clone, Serialize, Deserialize)] pub struct DescribeKeyRequest { /// Key ID to describe pub key_id: String, } /// Response from describe key operation #[derive(Debug, Clone, Serialize, Deserialize)] pub struct DescribeKeyResponse { /// Success flag pub success: bool, /// Status message pub message: String, /// Key metadata pub key_metadata: Option<KeyMetadata>, } /// Request to cancel key deletion #[derive(Debug, Clone, Serialize, Deserialize)] pub struct CancelKeyDeletionRequest { /// Key ID to cancel deletion for pub key_id: String, } /// Response from cancel key deletion operation #[derive(Debug, Clone, Serialize, Deserialize)] pub struct CancelKeyDeletionResponse { /// Success flag pub success: bool, /// Status message pub message: String, /// Key ID pub key_id: String, } /// Request to update key description #[derive(Debug, Clone, Serialize, Deserialize)] pub struct UpdateKeyDescriptionRequest { /// Key ID to update pub key_id: String, /// New description pub description: String, } /// Response from update key description operation #[derive(Debug, Clone, Serialize, Deserialize)] pub struct UpdateKeyDescriptionResponse { /// Success flag pub success: bool, /// Status message pub message: String, /// Key ID pub key_id: String, } /// Request to add/update key tags #[derive(Debug, Clone, Serialize, Deserialize)] pub struct TagKeyRequest { /// Key ID to tag pub key_id: String, /// Tags to add/update pub tags: HashMap<String, String>, } /// Response from tag key operation #[derive(Debug, Clone, Serialize, Deserialize)] pub struct TagKeyResponse { /// Success flag pub success: bool, /// Status message pub message: String, /// Key ID pub key_id: String, } /// Request to remove key tags #[derive(Debug, Clone, Serialize, Deserialize)] pub struct UntagKeyRequest { /// Key ID to untag pub key_id: String, /// Tag keys to remove pub tag_keys: Vec<String>, } /// Response from untag key operation #[derive(Debug, Clone, Serialize, Deserialize)] pub struct UntagKeyResponse { /// Success flag pub success: bool, /// Status message pub message: String, /// Key ID pub key_id: String, }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/kms/src/cache.rs
crates/kms/src/cache.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Caching layer for KMS operations to improve performance use crate::types::{KeyMetadata, KeySpec}; use moka::future::Cache; use std::time::Duration; /// Cached data key entry #[derive(Clone, Debug)] pub struct CachedDataKey { pub plaintext: Vec<u8>, pub ciphertext: Vec<u8>, pub key_spec: KeySpec, } /// KMS cache for storing frequently accessed keys and metadata pub struct KmsCache { key_metadata_cache: Cache<String, KeyMetadata>, data_key_cache: Cache<String, CachedDataKey>, } impl KmsCache { /// Create a new KMS cache with the specified capacity /// /// # Arguments /// * `capacity` - Maximum number of entries in the cache /// /// # Returns /// A new instance of `KmsCache` /// pub fn new(capacity: u64) -> Self { Self { key_metadata_cache: Cache::builder() .max_capacity(capacity / 2) .time_to_live(Duration::from_secs(300)) // 5 minutes default TTL .build(), data_key_cache: Cache::builder() .max_capacity(capacity / 2) .time_to_live(Duration::from_secs(60)) // 1 minute for data keys (shorter for security) .build(), } } /// Get key metadata from cache /// /// # Arguments /// * `key_id` - The ID of the key to retrieve metadata for /// /// # Returns /// An `Option` containing the `KeyMetadata` if found, or `None` if not found /// pub async fn get_key_metadata(&self, key_id: &str) -> Option<KeyMetadata> { self.key_metadata_cache.get(key_id).await } /// Put key metadata into cache /// /// # Arguments /// * `key_id` - The ID of the key to store metadata for /// * `metadata` - The `KeyMetadata` to store in the cache /// pub async fn put_key_metadata(&mut self, key_id: &str, metadata: &KeyMetadata) { self.key_metadata_cache.insert(key_id.to_string(), metadata.clone()).await; self.key_metadata_cache.run_pending_tasks().await; } /// Get data key from cache /// /// # Arguments /// * `key_id` - The ID of the key to retrieve the data key for /// /// # Returns /// An `Option` containing the `CachedDataKey` if found, or `None` if not found /// pub async fn get_data_key(&self, key_id: &str) -> Option<CachedDataKey> { self.data_key_cache.get(key_id).await } /// Put data key into cache /// /// # Arguments /// * `key_id` - The ID of the key to store the data key for /// * `plaintext` - The plaintext data key bytes /// * `ciphertext` - The ciphertext data key bytes /// pub async fn put_data_key(&mut self, key_id: &str, plaintext: &[u8], ciphertext: &[u8]) { let cached_key = CachedDataKey { plaintext: plaintext.to_vec(), ciphertext: ciphertext.to_vec(), key_spec: KeySpec::Aes256, // Default to AES-256 }; self.data_key_cache.insert(key_id.to_string(), cached_key).await; self.data_key_cache.run_pending_tasks().await; } /// Remove key metadata from cache /// /// # Arguments /// * `key_id` - The ID of the key to remove metadata for /// pub async fn remove_key_metadata(&mut self, key_id: &str) { self.key_metadata_cache.remove(key_id).await; } /// Remove data key from cache /// /// # Arguments /// * `key_id` - The ID of the key to remove the data key for /// pub async fn remove_data_key(&mut self, key_id: &str) { self.data_key_cache.remove(key_id).await; } /// Clear all cached entries pub async fn clear(&mut self) { self.key_metadata_cache.invalidate_all(); self.data_key_cache.invalidate_all(); // Wait for invalidation to complete self.key_metadata_cache.run_pending_tasks().await; self.data_key_cache.run_pending_tasks().await; } /// Get cache statistics (hit count, miss count) /// /// # Returns /// A tuple containing total entries and total misses /// pub fn stats(&self) -> (u64, u64) { let metadata_stats = ( self.key_metadata_cache.entry_count(), 0u64, // moka doesn't provide miss count directly ); let data_key_stats = (self.data_key_cache.entry_count(), 0u64); (metadata_stats.0 + data_key_stats.0, metadata_stats.1 + data_key_stats.1) } } #[cfg(test)] mod tests { use super::*; use crate::types::{KeyState, KeyUsage}; use std::time::Duration; #[derive(Debug, Clone)] struct CacheInfo { key_metadata_count: u64, data_key_count: u64, } impl CacheInfo { fn total_entries(&self) -> u64 { self.key_metadata_count + self.data_key_count } } impl KmsCache { fn with_ttl_for_tests(capacity: u64, metadata_ttl: Duration, data_key_ttl: Duration) -> Self { Self { key_metadata_cache: Cache::builder().max_capacity(capacity / 2).time_to_live(metadata_ttl).build(), data_key_cache: Cache::builder().max_capacity(capacity / 2).time_to_live(data_key_ttl).build(), } } fn info_for_tests(&self) -> CacheInfo { CacheInfo { key_metadata_count: self.key_metadata_cache.entry_count(), data_key_count: self.data_key_cache.entry_count(), } } fn contains_key_metadata_for_tests(&self, key_id: &str) -> bool { self.key_metadata_cache.contains_key(key_id) } fn contains_data_key_for_tests(&self, key_id: &str) -> bool { self.data_key_cache.contains_key(key_id) } } #[tokio::test] async fn test_cache_operations() { let mut cache = KmsCache::new(100); // Test key metadata caching let metadata = KeyMetadata { key_id: "test-key-1".to_string(), key_state: KeyState::Enabled, key_usage: KeyUsage::EncryptDecrypt, description: Some("Test key".to_string()), creation_date: chrono::Utc::now(), deletion_date: None, origin: "KMS".to_string(), key_manager: "CUSTOMER".to_string(), tags: std::collections::HashMap::new(), }; // Put and get metadata cache.put_key_metadata("test-key-1", &metadata).await; let retrieved = cache.get_key_metadata("test-key-1").await; assert!(retrieved.is_some()); assert_eq!(retrieved.expect("metadata should be cached").key_id, "test-key-1"); // Test data key caching let plaintext = vec![1, 2, 3, 4]; let ciphertext = vec![5, 6, 7, 8]; cache.put_data_key("test-key-1", &plaintext, &ciphertext).await; let cached_data_key = cache.get_data_key("test-key-1").await; assert!(cached_data_key.is_some()); let cached_data_key = cached_data_key.expect("data key should be cached"); assert_eq!(cached_data_key.plaintext, plaintext); assert_eq!(cached_data_key.ciphertext, ciphertext); assert_eq!(cached_data_key.key_spec, KeySpec::Aes256); // Test cache info let info = cache.info_for_tests(); assert_eq!(info.key_metadata_count, 1); assert_eq!(info.data_key_count, 1); assert_eq!(info.total_entries(), 2); // Test cache clearing cache.clear().await; let info_after_clear = cache.info_for_tests(); assert_eq!(info_after_clear.total_entries(), 0); } #[tokio::test] async fn test_cache_with_custom_ttl() { let mut cache = KmsCache::with_ttl_for_tests( 100, Duration::from_millis(100), // Short TTL for testing Duration::from_millis(50), ); let metadata = KeyMetadata { key_id: "ttl-test-key".to_string(), key_state: KeyState::Enabled, key_usage: KeyUsage::EncryptDecrypt, description: Some("TTL test key".to_string()), creation_date: chrono::Utc::now(), deletion_date: None, origin: "KMS".to_string(), key_manager: "CUSTOMER".to_string(), tags: std::collections::HashMap::new(), }; cache.put_key_metadata("ttl-test-key", &metadata).await; // Should be present immediately assert!(cache.get_key_metadata("ttl-test-key").await.is_some()); // Wait for TTL to expire tokio::time::sleep(Duration::from_millis(150)).await; // Should be expired now assert!(cache.get_key_metadata("ttl-test-key").await.is_none()); } #[tokio::test] async fn test_cache_contains_methods() { let mut cache = KmsCache::new(100); assert!(!cache.contains_key_metadata_for_tests("nonexistent")); assert!(!cache.contains_data_key_for_tests("nonexistent")); let metadata = KeyMetadata { key_id: "contains-test".to_string(), key_state: KeyState::Enabled, key_usage: KeyUsage::EncryptDecrypt, description: None, creation_date: chrono::Utc::now(), deletion_date: None, origin: "KMS".to_string(), key_manager: "CUSTOMER".to_string(), tags: std::collections::HashMap::new(), }; cache.put_key_metadata("contains-test", &metadata).await; cache.put_data_key("contains-test", &[1, 2, 3], &[4, 5, 6]).await; assert!(cache.contains_key_metadata_for_tests("contains-test")); assert!(cache.contains_data_key_for_tests("contains-test")); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/kms/src/encryption/service.rs
crates/kms/src/encryption/service.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Object encryption service for S3-compatible encryption use crate::encryption::ciphers::{create_cipher, generate_iv}; use crate::error::{KmsError, Result}; use crate::manager::KmsManager; use crate::types::*; use base64::Engine; use rand::random; use std::collections::HashMap; use std::io::Cursor; use tokio::io::{AsyncRead, AsyncReadExt}; use tracing::{debug, info}; use zeroize::Zeroize; /// Data key for object encryption /// SECURITY: This struct automatically zeros sensitive key material when dropped #[derive(Debug, Clone)] pub struct DataKey { /// 256-bit encryption key - automatically zeroed on drop pub plaintext_key: [u8; 32], /// 96-bit nonce for GCM mode - not secret so no need to zero pub nonce: [u8; 12], } // SECURITY: Implement Drop to automatically zero sensitive key material impl Drop for DataKey { fn drop(&mut self) { self.plaintext_key.zeroize(); } } /// Service for encrypting and decrypting S3 objects with KMS integration pub struct ObjectEncryptionService { kms_manager: KmsManager, } /// Result of object encryption #[derive(Debug, Clone)] pub struct EncryptionResult { /// Encrypted data pub ciphertext: Vec<u8>, /// Encryption metadata to be stored with the object pub metadata: EncryptionMetadata, } impl ObjectEncryptionService { /// Create a new object encryption service /// /// # Arguments /// * `kms_manager` - KMS manager to use for key operations /// /// # Returns /// New ObjectEncryptionService instance /// pub fn new(kms_manager: KmsManager) -> Self { Self { kms_manager } } /// Create a new master key (delegates to KMS manager) /// /// # Arguments /// * `request` - CreateKeyRequest with key parameters /// /// # Returns /// CreateKeyResponse with created key details /// pub async fn create_key(&self, request: CreateKeyRequest) -> Result<CreateKeyResponse> { self.kms_manager.create_key(request).await } /// Describe a master key (delegates to KMS manager) /// /// # Arguments /// * `request` - DescribeKeyRequest with key ID /// /// # Returns /// DescribeKeyResponse with key metadata /// pub async fn describe_key(&self, request: DescribeKeyRequest) -> Result<DescribeKeyResponse> { self.kms_manager.describe_key(request).await } /// List master keys (delegates to KMS manager) /// /// # Arguments /// * `request` - ListKeysRequest with listing parameters /// /// # Returns /// ListKeysResponse with list of keys /// pub async fn list_keys(&self, request: ListKeysRequest) -> Result<ListKeysResponse> { self.kms_manager.list_keys(request).await } /// Generate a data encryption key (delegates to KMS manager) /// /// # Arguments /// * `request` - GenerateDataKeyRequest with key parameters /// /// # Returns /// GenerateDataKeyResponse with generated key details /// pub async fn generate_data_key(&self, request: GenerateDataKeyRequest) -> Result<GenerateDataKeyResponse> { self.kms_manager.generate_data_key(request).await } /// Get the default key ID /// /// # Returns /// Option with default key ID if configured /// pub fn get_default_key_id(&self) -> Option<&String> { self.kms_manager.get_default_key_id() } /// Get cache statistics /// /// # Returns /// Option with (hits, misses) if caching is enabled /// pub async fn cache_stats(&self) -> Option<(u64, u64)> { self.kms_manager.cache_stats().await } /// Clear the cache /// /// # Returns /// Result indicating success or failure /// pub async fn clear_cache(&self) -> Result<()> { self.kms_manager.clear_cache().await } /// Get backend health status /// /// # Returns /// Result indicating if backend is healthy /// pub async fn health_check(&self) -> Result<bool> { self.kms_manager.health_check().await } /// Create a data encryption key for object encryption /// /// # Arguments /// * `kms_key_id` - Optional KMS key ID to use (uses default if None) /// * `context` - ObjectEncryptionContext with bucket and object key /// /// # Returns /// Tuple with DataKey and encrypted key blob /// pub async fn create_data_key( &self, kms_key_id: &Option<String>, context: &ObjectEncryptionContext, ) -> Result<(DataKey, Vec<u8>)> { // Determine the KMS key ID to use let actual_key_id = kms_key_id .as_ref() .map(|s| s.as_str()) .or_else(|| self.kms_manager.get_default_key_id().map(|s| s.as_str())) .ok_or_else(|| KmsError::configuration_error("No KMS key ID specified and no default configured"))?; // Build encryption context let mut enc_context = context.encryption_context.clone(); enc_context.insert("bucket".to_string(), context.bucket.clone()); enc_context.insert("object_key".to_string(), context.object_key.clone()); let request = GenerateDataKeyRequest { key_id: actual_key_id.to_string(), key_spec: KeySpec::Aes256, encryption_context: enc_context, }; let data_key_response = self.kms_manager.generate_data_key(request).await?; // Generate a unique random nonce for this data key // This ensures each object/part gets a unique base nonce for streaming encryption let nonce: [u8; 12] = random(); tracing::info!("Generated random nonce for data key: {:02x?}", nonce); let data_key = DataKey { plaintext_key: data_key_response .plaintext_key .try_into() .map_err(|_| KmsError::internal_error("Invalid key length"))?, nonce, }; Ok((data_key, data_key_response.ciphertext_blob)) } /// Decrypt a data encryption key /// /// # Arguments /// * `encrypted_key` - Encrypted data key blob /// * `context` - ObjectEncryptionContext with bucket and object key /// /// # Returns /// DataKey with decrypted key /// pub async fn decrypt_data_key(&self, encrypted_key: &[u8], _context: &ObjectEncryptionContext) -> Result<DataKey> { let decrypt_request = DecryptRequest { ciphertext: encrypted_key.to_vec(), encryption_context: HashMap::new(), grant_tokens: Vec::new(), }; let decrypt_response = self.kms_manager.decrypt(decrypt_request).await?; let data_key = DataKey { plaintext_key: decrypt_response .plaintext .try_into() .map_err(|_| KmsError::internal_error("Invalid key length"))?, nonce: [0u8; 12], // This will be replaced by stored nonce during GET }; Ok(data_key) } /// Encrypt object data using server-side encryption /// /// # Arguments /// * `bucket` - S3 bucket name /// * `object_key` - S3 object key /// * `reader` - Data reader /// * `algorithm` - Encryption algorithm to use /// * `kms_key_id` - Optional KMS key ID (uses default if None) /// * `encryption_context` - Additional encryption context /// /// # Returns /// EncryptionResult containing encrypted data and metadata pub async fn encrypt_object<R>( &self, bucket: &str, object_key: &str, mut reader: R, algorithm: &EncryptionAlgorithm, kms_key_id: Option<&str>, encryption_context: Option<&HashMap<String, String>>, ) -> Result<EncryptionResult> where R: AsyncRead + Unpin, { debug!("Encrypting object {}/{} with algorithm {:?}", bucket, object_key, algorithm); // Read all data (for simplicity - in production, use streaming) let mut data = Vec::new(); reader.read_to_end(&mut data).await?; let original_size = data.len() as u64; // Determine the KMS key ID to use let actual_key_id = kms_key_id .or_else(|| self.kms_manager.get_default_key_id().map(|s| s.as_str())) .ok_or_else(|| KmsError::configuration_error("No KMS key ID specified and no default configured"))?; // Build encryption context let mut context = encryption_context.cloned().unwrap_or_default(); context.insert("bucket".to_string(), bucket.to_string()); context.insert("object".to_string(), object_key.to_string()); context.insert("algorithm".to_string(), algorithm.as_str().to_string()); // Auto-create key for SSE-S3 if it doesn't exist if algorithm == &EncryptionAlgorithm::Aes256 { let describe_req = DescribeKeyRequest { key_id: actual_key_id.to_string(), }; if let Err(KmsError::KeyNotFound { .. }) = self.kms_manager.describe_key(describe_req).await { info!("Auto-creating SSE-S3 key: {}", actual_key_id); let create_req = CreateKeyRequest { key_name: Some(actual_key_id.to_string()), key_usage: KeyUsage::EncryptDecrypt, description: Some("Auto-created SSE-S3 key".to_string()), policy: None, tags: HashMap::new(), origin: None, }; self.kms_manager .create_key(create_req) .await .map_err(|e| KmsError::backend_error(format!("Failed to auto-create SSE-S3 key {actual_key_id}: {e}")))?; } } else { // For SSE-KMS, key must exist let describe_req = DescribeKeyRequest { key_id: actual_key_id.to_string(), }; self.kms_manager.describe_key(describe_req).await.map_err(|_| { KmsError::invalid_operation(format!("SSE-KMS key '{actual_key_id}' not found. Please create it first.")) })?; } // Generate data encryption key let request = GenerateDataKeyRequest { key_id: actual_key_id.to_string(), key_spec: KeySpec::Aes256, encryption_context: context.clone(), }; let data_key = self .kms_manager .generate_data_key(request) .await .map_err(|e| KmsError::backend_error(format!("Failed to generate data key: {e}")))?; let plaintext_key = data_key.plaintext_key; // Create cipher and generate IV let cipher = create_cipher(algorithm, &plaintext_key)?; let iv = generate_iv(algorithm); // Build AAD from encryption context let aad = serde_json::to_vec(&context)?; // Encrypt the data let (ciphertext, tag) = cipher.encrypt(&data, &iv, &aad)?; // Create encryption metadata let metadata = EncryptionMetadata { algorithm: algorithm.as_str().to_string(), key_id: actual_key_id.to_string(), key_version: 1, // Default to version 1 for now iv, tag: Some(tag), encryption_context: context, encrypted_at: chrono::Utc::now(), original_size, encrypted_data_key: data_key.ciphertext_blob, }; info!("Successfully encrypted object {}/{} ({} bytes)", bucket, object_key, original_size); Ok(EncryptionResult { ciphertext, metadata }) } /// Decrypt object data /// /// # Arguments /// * `bucket` - S3 bucket name /// * `object_key` - S3 object key /// * `ciphertext` - Encrypted data /// * `metadata` - Encryption metadata /// * `expected_context` - Expected encryption context for validation /// /// # Returns /// Decrypted data as a reader pub async fn decrypt_object( &self, bucket: &str, object_key: &str, ciphertext: Vec<u8>, metadata: &EncryptionMetadata, expected_context: Option<&HashMap<String, String>>, ) -> Result<Box<dyn AsyncRead + Send + Sync + Unpin>> { debug!("Decrypting object {}/{} with algorithm {}", bucket, object_key, metadata.algorithm); // Validate encryption context if provided if let Some(expected) = expected_context { self.validate_encryption_context(&metadata.encryption_context, expected)?; } // Parse algorithm let algorithm = metadata .algorithm .parse::<EncryptionAlgorithm>() .map_err(|_| KmsError::unsupported_algorithm(&metadata.algorithm))?; // Decrypt the data key let decrypt_request = DecryptRequest { ciphertext: metadata.encrypted_data_key.clone(), encryption_context: metadata.encryption_context.clone(), grant_tokens: Vec::new(), }; let decrypt_response = self .kms_manager .decrypt(decrypt_request) .await .map_err(|e| KmsError::backend_error(format!("Failed to decrypt data key: {e}")))?; // Create cipher let cipher = create_cipher(&algorithm, &decrypt_response.plaintext)?; // Build AAD from encryption context let aad = serde_json::to_vec(&metadata.encryption_context)?; // Get tag from metadata let tag = metadata .tag .as_ref() .ok_or_else(|| KmsError::invalid_operation("Missing authentication tag"))?; // Decrypt the data let plaintext = cipher.decrypt(&ciphertext, &metadata.iv, tag, &aad)?; info!("Successfully decrypted object {}/{} ({} bytes)", bucket, object_key, plaintext.len()); Ok(Box::new(Cursor::new(plaintext))) } /// Encrypt object with customer-provided key (SSE-C) /// /// # Arguments /// * `bucket` - S3 bucket name /// * `object_key` - S3 object key /// * `reader` - Data reader /// * `customer_key` - Customer-provided 256-bit key /// * `customer_key_md5` - Optional MD5 hash of the customer key for validation /// /// # Returns /// EncryptionResult with SSE-C metadata pub async fn encrypt_object_with_customer_key<R>( &self, bucket: &str, object_key: &str, mut reader: R, customer_key: &[u8], customer_key_md5: Option<&str>, ) -> Result<EncryptionResult> where R: AsyncRead + Unpin, { debug!("Encrypting object {}/{} with customer-provided key (SSE-C)", bucket, object_key); // Validate key size if customer_key.len() != 32 { return Err(KmsError::invalid_key_size(32, customer_key.len())); } // Validate key MD5 if provided if let Some(expected_md5) = customer_key_md5 { let actual_md5 = md5::compute(customer_key); let actual_md5_hex = format!("{actual_md5:x}"); if actual_md5_hex != expected_md5.to_lowercase() { return Err(KmsError::validation_error("Customer key MD5 mismatch")); } } // Read all data let mut data = Vec::new(); reader.read_to_end(&mut data).await?; let original_size = data.len() as u64; // Create cipher and generate IV let algorithm = EncryptionAlgorithm::Aes256; let cipher = create_cipher(&algorithm, customer_key)?; let iv = generate_iv(&algorithm); // Build minimal encryption context for SSE-C let context = HashMap::from([ ("bucket".to_string(), bucket.to_string()), ("object".to_string(), object_key.to_string()), ("sse_type".to_string(), "customer".to_string()), ]); let aad = serde_json::to_vec(&context)?; // Encrypt the data let (ciphertext, tag) = cipher.encrypt(&data, &iv, &aad)?; // Create metadata (no encrypted data key for SSE-C) let metadata = EncryptionMetadata { algorithm: algorithm.as_str().to_string(), key_id: "sse-c".to_string(), // Special marker for SSE-C key_version: 1, iv, tag: Some(tag), encryption_context: context, encrypted_at: chrono::Utc::now(), original_size, encrypted_data_key: Vec::new(), // Empty for SSE-C }; info!( "Successfully encrypted object {}/{} with SSE-C ({} bytes)", bucket, object_key, original_size ); Ok(EncryptionResult { ciphertext, metadata }) } /// Decrypt object with customer-provided key (SSE-C) /// /// # Arguments /// * `bucket` - S3 bucket name /// * `object_key` - S3 object key /// * `ciphertext` - Encrypted data /// * `metadata` - Encryption metadata /// * `customer_key` - Customer-provided 256-bit key /// /// # Returns /// Decrypted data as a reader /// pub async fn decrypt_object_with_customer_key( &self, bucket: &str, object_key: &str, ciphertext: Vec<u8>, metadata: &EncryptionMetadata, customer_key: &[u8], ) -> Result<Box<dyn AsyncRead + Send + Sync + Unpin>> { debug!("Decrypting object {}/{} with customer-provided key (SSE-C)", bucket, object_key); // Validate key size if customer_key.len() != 32 { return Err(KmsError::invalid_key_size(32, customer_key.len())); } // Validate that this is SSE-C if metadata.key_id != "sse-c" { return Err(KmsError::invalid_operation("This object was not encrypted with SSE-C")); } // Parse algorithm let algorithm = metadata .algorithm .parse::<EncryptionAlgorithm>() .map_err(|_| KmsError::unsupported_algorithm(&metadata.algorithm))?; // Create cipher let cipher = create_cipher(&algorithm, customer_key)?; // Build AAD from encryption context let aad = serde_json::to_vec(&metadata.encryption_context)?; // Get tag from metadata let tag = metadata .tag .as_ref() .ok_or_else(|| KmsError::invalid_operation("Missing authentication tag"))?; // Decrypt the data let plaintext = cipher.decrypt(&ciphertext, &metadata.iv, tag, &aad)?; info!( "Successfully decrypted SSE-C object {}/{} ({} bytes)", bucket, object_key, plaintext.len() ); Ok(Box::new(Cursor::new(plaintext))) } /// Validate encryption context /// /// # Arguments /// * `actual` - Actual encryption context from metadata /// * `expected` - Expected encryption context to validate against /// /// # Returns /// Result indicating success or context mismatch /// fn validate_encryption_context(&self, actual: &HashMap<String, String>, expected: &HashMap<String, String>) -> Result<()> { for (key, expected_value) in expected { match actual.get(key) { Some(actual_value) if actual_value == expected_value => continue, Some(actual_value) => { return Err(KmsError::context_mismatch(format!( "Context mismatch for '{key}': expected '{expected_value}', got '{actual_value}'" ))); } None => { return Err(KmsError::context_mismatch(format!("Missing context key '{key}'"))); } } } Ok(()) } /// Convert encryption metadata to HTTP headers for S3 compatibility /// /// # Arguments /// * `metadata` - EncryptionMetadata to convert /// /// # Returns /// HashMap of HTTP headers /// pub fn metadata_to_headers(&self, metadata: &EncryptionMetadata) -> HashMap<String, String> { let mut headers = HashMap::new(); // Standard S3 encryption headers if metadata.key_id == "sse-c" { headers.insert("x-amz-server-side-encryption".to_string(), "AES256".to_string()); headers.insert("x-amz-server-side-encryption-customer-algorithm".to_string(), "AES256".to_string()); } else if metadata.algorithm == "AES256" { headers.insert("x-amz-server-side-encryption".to_string(), "AES256".to_string()); // For SSE-S3, we still need to store the key ID for internal use headers.insert("x-amz-server-side-encryption-aws-kms-key-id".to_string(), metadata.key_id.clone()); } else { headers.insert("x-amz-server-side-encryption".to_string(), "aws:kms".to_string()); headers.insert("x-amz-server-side-encryption-aws-kms-key-id".to_string(), metadata.key_id.clone()); } // Internal headers for decryption headers.insert( "x-rustfs-encryption-iv".to_string(), base64::engine::general_purpose::STANDARD.encode(&metadata.iv), ); if let Some(ref tag) = metadata.tag { headers.insert( "x-rustfs-encryption-tag".to_string(), base64::engine::general_purpose::STANDARD.encode(tag), ); } headers.insert( "x-rustfs-encryption-key".to_string(), base64::engine::general_purpose::STANDARD.encode(&metadata.encrypted_data_key), ); headers.insert( "x-rustfs-encryption-context".to_string(), serde_json::to_string(&metadata.encryption_context).unwrap_or_default(), ); headers } /// Parse encryption metadata from HTTP headers /// /// # Arguments /// * `headers` - HashMap of HTTP headers /// /// # Returns /// EncryptionMetadata parsed from headers /// pub fn headers_to_metadata(&self, headers: &HashMap<String, String>) -> Result<EncryptionMetadata> { let algorithm = headers .get("x-amz-server-side-encryption") .ok_or_else(|| KmsError::validation_error("Missing encryption algorithm header"))? .clone(); let key_id = if algorithm == "AES256" && headers.contains_key("x-amz-server-side-encryption-customer-algorithm") { "sse-c".to_string() } else if let Some(kms_key_id) = headers.get("x-amz-server-side-encryption-aws-kms-key-id") { kms_key_id.clone() } else { return Err(KmsError::validation_error("Missing key ID")); }; let iv = headers .get("x-rustfs-encryption-iv") .ok_or_else(|| KmsError::validation_error("Missing IV header"))?; let iv = base64::engine::general_purpose::STANDARD .decode(iv) .map_err(|e| KmsError::validation_error(format!("Invalid IV: {e}")))?; let tag = if let Some(tag_str) = headers.get("x-rustfs-encryption-tag") { Some( base64::engine::general_purpose::STANDARD .decode(tag_str) .map_err(|e| KmsError::validation_error(format!("Invalid tag: {e}")))?, ) } else { None }; let encrypted_data_key = if let Some(key_str) = headers.get("x-rustfs-encryption-key") { base64::engine::general_purpose::STANDARD .decode(key_str) .map_err(|e| KmsError::validation_error(format!("Invalid encrypted key: {e}")))? } else { Vec::new() // Empty for SSE-C }; let encryption_context = if let Some(context_str) = headers.get("x-rustfs-encryption-context") { serde_json::from_str(context_str) .map_err(|e| KmsError::validation_error(format!("Invalid encryption context: {e}")))? } else { HashMap::new() }; Ok(EncryptionMetadata { algorithm, key_id, key_version: 1, // Default for parsing iv, tag, encryption_context, encrypted_at: chrono::Utc::now(), original_size: 0, // Not available from headers encrypted_data_key, }) } } #[cfg(test)] mod tests { use super::*; use crate::config::KmsConfig; use std::sync::Arc; use tempfile::TempDir; async fn create_test_service() -> (ObjectEncryptionService, TempDir) { let temp_dir = TempDir::new().expect("Failed to create temp dir"); let config = KmsConfig::local(temp_dir.path().to_path_buf()).with_default_key("test-key".to_string()); let backend = Arc::new( crate::backends::local::LocalKmsBackend::new(config.clone()) .await .expect("local backend should initialize"), ); let kms_manager = KmsManager::new(backend, config); let service = ObjectEncryptionService::new(kms_manager); (service, temp_dir) } #[tokio::test] async fn test_sse_s3_encryption() { let (service, _temp_dir) = create_test_service().await; let bucket = "test-bucket"; let object_key = "test-object"; let data = b"Hello, SSE-S3!"; let reader = Cursor::new(data.to_vec()); // Encrypt with SSE-S3 (auto-create key) let result = service .encrypt_object( bucket, object_key, reader, &EncryptionAlgorithm::Aes256, None, // Use default key None, ) .await .expect("Encryption failed"); assert!(!result.ciphertext.is_empty()); assert_eq!(result.metadata.algorithm, "AES256"); assert_eq!(result.metadata.original_size, data.len() as u64); // Decrypt let decrypted_reader = service .decrypt_object(bucket, object_key, result.ciphertext, &result.metadata, None) .await .expect("Decryption failed"); let mut decrypted_data = Vec::new(); let mut reader = decrypted_reader; reader .read_to_end(&mut decrypted_data) .await .expect("Failed to read decrypted data"); assert_eq!(decrypted_data, data); } #[tokio::test] async fn test_sse_c_encryption() { let (service, _temp_dir) = create_test_service().await; let bucket = "test-bucket"; let object_key = "test-object"; let data = b"Hello, SSE-C!"; let reader = Cursor::new(data.to_vec()); let customer_key = [0u8; 32]; // 256-bit key // Encrypt with SSE-C let result = service .encrypt_object_with_customer_key(bucket, object_key, reader, &customer_key, None) .await .expect("SSE-C encryption failed"); assert!(!result.ciphertext.is_empty()); assert_eq!(result.metadata.key_id, "sse-c"); assert_eq!(result.metadata.original_size, data.len() as u64); // Decrypt with same customer key let decrypted_reader = service .decrypt_object_with_customer_key(bucket, object_key, result.ciphertext, &result.metadata, &customer_key) .await .expect("SSE-C decryption failed"); let mut decrypted_data = Vec::new(); let mut reader = decrypted_reader; reader .read_to_end(&mut decrypted_data) .await .expect("Failed to read decrypted data"); assert_eq!(decrypted_data, data); } #[tokio::test] async fn test_metadata_headers_conversion() { let (service, _temp_dir) = create_test_service().await; let metadata = EncryptionMetadata { algorithm: "AES256".to_string(), key_id: "test-key".to_string(), key_version: 1, iv: vec![1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12], tag: Some(vec![1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]), encryption_context: HashMap::from([("bucket".to_string(), "test-bucket".to_string())]), encrypted_at: chrono::Utc::now(), original_size: 100, encrypted_data_key: vec![1, 2, 3, 4], }; // Convert to headers let headers = service.metadata_to_headers(&metadata); assert!(headers.contains_key("x-amz-server-side-encryption")); assert!(headers.contains_key("x-rustfs-encryption-iv")); // Convert back to metadata let parsed_metadata = service.headers_to_metadata(&headers).expect("Failed to parse headers"); assert_eq!(parsed_metadata.algorithm, metadata.algorithm); assert_eq!(parsed_metadata.key_id, metadata.key_id); assert_eq!(parsed_metadata.iv, metadata.iv); assert_eq!(parsed_metadata.tag, metadata.tag); } #[tokio::test] async fn test_encryption_context_validation() { let (service, _temp_dir) = create_test_service().await; let actual_context = HashMap::from([ ("bucket".to_string(), "test-bucket".to_string()), ("object".to_string(), "test-object".to_string()), ]); let valid_expected = HashMap::from([("bucket".to_string(), "test-bucket".to_string())]); let invalid_expected = HashMap::from([("bucket".to_string(), "wrong-bucket".to_string())]); // Valid context should pass assert!(service.validate_encryption_context(&actual_context, &valid_expected).is_ok()); // Invalid context should fail assert!( service .validate_encryption_context(&actual_context, &invalid_expected) .is_err() ); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/kms/src/encryption/mod.rs
crates/kms/src/encryption/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Object encryption service implementation mod ciphers; pub mod service; pub use service::ObjectEncryptionService;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/kms/src/encryption/ciphers.rs
crates/kms/src/encryption/ciphers.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Cipher implementations for object encryption use crate::error::{KmsError, Result}; use crate::types::EncryptionAlgorithm; use aes_gcm::{ Aes256Gcm, Key, Nonce, aead::{Aead, KeyInit}, }; use chacha20poly1305::ChaCha20Poly1305; use rand::Rng; /// Trait for object encryption ciphers #[cfg_attr(not(test), allow(dead_code))] pub trait ObjectCipher: Send + Sync { /// Encrypt data with the given IV and AAD fn encrypt(&self, plaintext: &[u8], iv: &[u8], aad: &[u8]) -> Result<(Vec<u8>, Vec<u8>)>; /// Decrypt data with the given IV, tag, and AAD fn decrypt(&self, ciphertext: &[u8], iv: &[u8], tag: &[u8], aad: &[u8]) -> Result<Vec<u8>>; /// Get the algorithm name fn algorithm(&self) -> &'static str; /// Get the required key size in bytes fn key_size(&self) -> usize; /// Get the required IV size in bytes fn iv_size(&self) -> usize; /// Get the tag size in bytes fn tag_size(&self) -> usize; } /// AES-256-GCM cipher implementation pub struct AesCipher { cipher: Aes256Gcm, } impl AesCipher { /// Create a new AES cipher with the given key /// /// #Arguments /// * `key` - A byte slice representing the AES-256 key (32 bytes) /// /// #Errors /// Returns `KmsError` if the key size is invalid /// /// #Returns /// A Result containing the AesCipher instance /// pub fn new(key: &[u8]) -> Result<Self> { if key.len() != 32 { return Err(KmsError::invalid_key_size(32, key.len())); } let key = Key::<Aes256Gcm>::try_from(key).map_err(|_| KmsError::cryptographic_error("key", "Invalid key length"))?; let cipher = Aes256Gcm::new(&key); Ok(Self { cipher }) } } impl ObjectCipher for AesCipher { fn encrypt(&self, plaintext: &[u8], iv: &[u8], aad: &[u8]) -> Result<(Vec<u8>, Vec<u8>)> { if iv.len() != 12 { return Err(KmsError::invalid_key_size(12, iv.len())); } let nonce = Nonce::try_from(iv).map_err(|_| KmsError::cryptographic_error("nonce", "Invalid nonce length"))?; // AES-GCM includes the tag in the ciphertext let ciphertext_with_tag = self .cipher .encrypt(&nonce, aes_gcm::aead::Payload { msg: plaintext, aad }) .map_err(KmsError::from_aes_gcm_error)?; // Split ciphertext and tag let tag_size = self.tag_size(); if ciphertext_with_tag.len() < tag_size { return Err(KmsError::cryptographic_error("AES-GCM encrypt", "Ciphertext too short for tag")); } let (ciphertext, tag) = ciphertext_with_tag.split_at(ciphertext_with_tag.len() - tag_size); Ok((ciphertext.to_vec(), tag.to_vec())) } fn decrypt(&self, ciphertext: &[u8], iv: &[u8], tag: &[u8], aad: &[u8]) -> Result<Vec<u8>> { if iv.len() != 12 { return Err(KmsError::invalid_key_size(12, iv.len())); } if tag.len() != self.tag_size() { return Err(KmsError::invalid_key_size(self.tag_size(), tag.len())); } let nonce = Nonce::try_from(iv).map_err(|_| KmsError::cryptographic_error("nonce", "Invalid nonce length"))?; // Combine ciphertext and tag for AES-GCM let mut ciphertext_with_tag = ciphertext.to_vec(); ciphertext_with_tag.extend_from_slice(tag); let plaintext = self .cipher .decrypt( &nonce, aes_gcm::aead::Payload { msg: &ciphertext_with_tag, aad, }, ) .map_err(KmsError::from_aes_gcm_error)?; Ok(plaintext) } fn algorithm(&self) -> &'static str { "AES-256-GCM" } fn key_size(&self) -> usize { 32 // 256 bits } fn iv_size(&self) -> usize { 12 // 96 bits for GCM } fn tag_size(&self) -> usize { 16 // 128 bits } } /// ChaCha20-Poly1305 cipher implementation pub struct ChaCha20Cipher { cipher: ChaCha20Poly1305, } impl ChaCha20Cipher { /// Create a new ChaCha20 cipher with the given key /// /// #Arguments /// * `key` - A byte slice representing the ChaCha20-Poly1305 key (32 bytes) /// /// #Errors /// Returns `KmsError` if the key size is invalid /// /// #Returns /// A Result containing the ChaCha20Cipher instance /// pub fn new(key: &[u8]) -> Result<Self> { if key.len() != 32 { return Err(KmsError::invalid_key_size(32, key.len())); } let key = chacha20poly1305::Key::try_from(key).map_err(|_| KmsError::cryptographic_error("key", "Invalid key length"))?; let cipher = ChaCha20Poly1305::new(&key); Ok(Self { cipher }) } } impl ObjectCipher for ChaCha20Cipher { fn encrypt(&self, plaintext: &[u8], iv: &[u8], aad: &[u8]) -> Result<(Vec<u8>, Vec<u8>)> { if iv.len() != 12 { return Err(KmsError::invalid_key_size(12, iv.len())); } let nonce = chacha20poly1305::Nonce::try_from(iv).map_err(|_| KmsError::cryptographic_error("nonce", "Invalid nonce length"))?; // ChaCha20-Poly1305 includes the tag in the ciphertext let ciphertext_with_tag = self .cipher .encrypt(&nonce, chacha20poly1305::aead::Payload { msg: plaintext, aad }) .map_err(KmsError::from_chacha20_error)?; // Split ciphertext and tag let tag_size = self.tag_size(); if ciphertext_with_tag.len() < tag_size { return Err(KmsError::cryptographic_error("ChaCha20-Poly1305 encrypt", "Ciphertext too short for tag")); } let (ciphertext, tag) = ciphertext_with_tag.split_at(ciphertext_with_tag.len() - tag_size); Ok((ciphertext.to_vec(), tag.to_vec())) } fn decrypt(&self, ciphertext: &[u8], iv: &[u8], tag: &[u8], aad: &[u8]) -> Result<Vec<u8>> { if iv.len() != 12 { return Err(KmsError::invalid_key_size(12, iv.len())); } if tag.len() != self.tag_size() { return Err(KmsError::invalid_key_size(self.tag_size(), tag.len())); } let nonce = chacha20poly1305::Nonce::try_from(iv).map_err(|_| KmsError::cryptographic_error("nonce", "Invalid nonce length"))?; // Combine ciphertext and tag for ChaCha20-Poly1305 let mut ciphertext_with_tag = ciphertext.to_vec(); ciphertext_with_tag.extend_from_slice(tag); let plaintext = self .cipher .decrypt( &nonce, chacha20poly1305::aead::Payload { msg: &ciphertext_with_tag, aad, }, ) .map_err(KmsError::from_chacha20_error)?; Ok(plaintext) } fn algorithm(&self) -> &'static str { "ChaCha20-Poly1305" } fn key_size(&self) -> usize { 32 // 256 bits } fn iv_size(&self) -> usize { 12 // 96 bits } fn tag_size(&self) -> usize { 16 // 128 bits } } /// Create a cipher instance for the given algorithm and key /// /// #Arguments /// * `algorithm` - The encryption algorithm to use /// * `key` - A byte slice representing the encryption key /// /// #Returns /// A Result containing a boxed ObjectCipher instance /// pub fn create_cipher(algorithm: &EncryptionAlgorithm, key: &[u8]) -> Result<Box<dyn ObjectCipher>> { match algorithm { EncryptionAlgorithm::Aes256 | EncryptionAlgorithm::AwsKms => Ok(Box::new(AesCipher::new(key)?)), EncryptionAlgorithm::ChaCha20Poly1305 => Ok(Box::new(ChaCha20Cipher::new(key)?)), } } /// Generate a random IV for the given algorithm /// /// #Arguments /// * `algorithm` - The encryption algorithm for which to generate the IV /// /// #Returns /// A vector containing the generated IV bytes /// pub fn generate_iv(algorithm: &EncryptionAlgorithm) -> Vec<u8> { let iv_size = match algorithm { EncryptionAlgorithm::Aes256 | EncryptionAlgorithm::AwsKms => 12, EncryptionAlgorithm::ChaCha20Poly1305 => 12, }; let mut iv = vec![0u8; iv_size]; rand::rng().fill(&mut iv[..]); iv } #[cfg(test)] mod tests { use super::*; #[test] fn test_aes_cipher() { let key = [0u8; 32]; // 256-bit key let cipher = AesCipher::new(&key).expect("Failed to create AES cipher"); let plaintext = b"Hello, World!"; let iv = [0u8; 12]; // 96-bit IV let aad = b"additional data"; // Test encryption let (ciphertext, tag) = cipher.encrypt(plaintext, &iv, aad).expect("Encryption failed"); assert!(!ciphertext.is_empty()); assert_eq!(tag.len(), 16); // 128-bit tag // Test decryption let decrypted = cipher.decrypt(&ciphertext, &iv, &tag, aad).expect("Decryption failed"); assert_eq!(decrypted, plaintext); // Test properties assert_eq!(cipher.algorithm(), "AES-256-GCM"); assert_eq!(cipher.key_size(), 32); assert_eq!(cipher.iv_size(), 12); assert_eq!(cipher.tag_size(), 16); } #[test] fn test_chacha20_cipher() { let key = [0u8; 32]; // 256-bit key let cipher = ChaCha20Cipher::new(&key).expect("Failed to create ChaCha20 cipher"); let plaintext = b"Hello, ChaCha20!"; let iv = [0u8; 12]; // 96-bit IV let aad = b"additional data"; // Test encryption let (ciphertext, tag) = cipher.encrypt(plaintext, &iv, aad).expect("Encryption failed"); assert!(!ciphertext.is_empty()); assert_eq!(tag.len(), 16); // 128-bit tag // Test decryption let decrypted = cipher.decrypt(&ciphertext, &iv, &tag, aad).expect("Decryption failed"); assert_eq!(decrypted, plaintext); // Test properties assert_eq!(cipher.algorithm(), "ChaCha20-Poly1305"); assert_eq!(cipher.key_size(), 32); assert_eq!(cipher.iv_size(), 12); assert_eq!(cipher.tag_size(), 16); } #[test] fn test_create_cipher() { let key = [0u8; 32]; // Test AES creation let aes_cipher = create_cipher(&EncryptionAlgorithm::Aes256, &key).expect("Failed to create AES cipher"); assert_eq!(aes_cipher.algorithm(), "AES-256-GCM"); // Test ChaCha20 creation let chacha_cipher = create_cipher(&EncryptionAlgorithm::ChaCha20Poly1305, &key).expect("Failed to create ChaCha20 cipher"); assert_eq!(chacha_cipher.algorithm(), "ChaCha20-Poly1305"); } #[test] fn test_generate_iv() { let aes_iv = generate_iv(&EncryptionAlgorithm::Aes256); assert_eq!(aes_iv.len(), 12); let chacha_iv = generate_iv(&EncryptionAlgorithm::ChaCha20Poly1305); assert_eq!(chacha_iv.len(), 12); // IVs should be different let another_aes_iv = generate_iv(&EncryptionAlgorithm::Aes256); assert_ne!(aes_iv, another_aes_iv); } #[test] fn test_invalid_key_size() { let short_key = [0u8; 16]; // Too short assert!(AesCipher::new(&short_key).is_err()); assert!(ChaCha20Cipher::new(&short_key).is_err()); } #[test] fn test_invalid_iv_size() { let key = [0u8; 32]; let cipher = AesCipher::new(&key).expect("Failed to create cipher"); let plaintext = b"test"; let short_iv = [0u8; 8]; // Too short let aad = b""; assert!(cipher.encrypt(plaintext, &short_iv, aad).is_err()); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/kms/src/backends/vault.rs
crates/kms/src/backends/vault.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Vault-based KMS backend implementation using vaultrs use crate::backends::{BackendInfo, KmsBackend, KmsClient}; use crate::config::{KmsConfig, VaultConfig}; use crate::error::{KmsError, Result}; use crate::types::*; use async_trait::async_trait; use base64::{Engine as _, engine::general_purpose}; use rand::RngCore; use serde::{Deserialize, Serialize}; use std::collections::HashMap; use tracing::{debug, info, warn}; use vaultrs::{ client::{VaultClient, VaultClientSettingsBuilder}, kv2, }; /// Vault KMS client implementation pub struct VaultKmsClient { client: VaultClient, config: VaultConfig, /// Mount path for the KV engine (typically "kv" or "secret") kv_mount: String, /// Path prefix for storing keys key_path_prefix: String, } /// Key data stored in Vault #[derive(Debug, Clone, Serialize, Deserialize)] struct VaultKeyData { /// Key algorithm algorithm: String, /// Key usage type usage: KeyUsage, /// Key creation timestamp created_at: chrono::DateTime<chrono::Utc>, /// Key status status: KeyStatus, /// Key version version: u32, /// Key description description: Option<String>, /// Key metadata metadata: HashMap<String, String>, /// Key tags tags: HashMap<String, String>, /// Encrypted key material (base64 encoded) encrypted_key_material: String, } impl VaultKmsClient { /// Create a new Vault KMS client pub async fn new(config: VaultConfig) -> Result<Self> { // Create client settings let mut settings_builder = VaultClientSettingsBuilder::default(); settings_builder.address(&config.address); // Set authentication token based on method let token = match &config.auth_method { crate::config::VaultAuthMethod::Token { token } => token.clone(), crate::config::VaultAuthMethod::AppRole { .. } => { // For AppRole authentication, we would need to first authenticate // and get a token. For simplicity, we'll require a token for now. return Err(KmsError::backend_error( "AppRole authentication not yet implemented. Please use token authentication.", )); } }; settings_builder.token(&token); if let Some(namespace) = &config.namespace { settings_builder.namespace(Some(namespace.clone())); } let settings = settings_builder .build() .map_err(|e| KmsError::backend_error(format!("Failed to build Vault client settings: {e}")))?; let client = VaultClient::new(settings).map_err(|e| KmsError::backend_error(format!("Failed to create Vault client: {e}")))?; info!("Successfully connected to Vault at {}", config.address); Ok(Self { client, kv_mount: config.kv_mount.clone(), key_path_prefix: config.key_path_prefix.clone(), config, }) } /// Get the full path for a key in Vault fn key_path(&self, key_id: &str) -> String { format!("{}/{}", self.key_path_prefix, key_id) } /// Generate key material for the given algorithm fn generate_key_material(algorithm: &str) -> Result<Vec<u8>> { let key_size = match algorithm { "AES_256" => 32, "AES_128" => 16, _ => return Err(KmsError::unsupported_algorithm(algorithm)), }; let mut key_material = vec![0u8; key_size]; rand::rng().fill_bytes(&mut key_material); Ok(key_material) } /// Encrypt key material using Vault's transit engine async fn encrypt_key_material(&self, key_material: &[u8]) -> Result<String> { // For simplicity, we'll base64 encode the key material // In a production setup, you would use Vault's transit engine for additional encryption Ok(general_purpose::STANDARD.encode(key_material)) } /// Decrypt key material async fn decrypt_key_material(&self, encrypted_material: &str) -> Result<Vec<u8>> { // For simplicity, we'll base64 decode the key material // In a production setup, you would use Vault's transit engine for decryption general_purpose::STANDARD .decode(encrypted_material) .map_err(|e| KmsError::cryptographic_error("decrypt", e.to_string())) } /// Store key data in Vault async fn store_key_data(&self, key_id: &str, key_data: &VaultKeyData) -> Result<()> { let path = self.key_path(key_id); kv2::set(&self.client, &self.kv_mount, &path, key_data) .await .map_err(|e| KmsError::backend_error(format!("Failed to store key in Vault: {e}")))?; debug!("Stored key {} in Vault at path {}", key_id, path); Ok(()) } async fn store_key_metadata(&self, key_id: &str, request: &CreateKeyRequest) -> Result<()> { debug!("Storing key metadata for {}, input tags: {:?}", key_id, request.tags); let key_data = VaultKeyData { algorithm: "AES_256".to_string(), usage: request.key_usage.clone(), created_at: chrono::Utc::now(), status: KeyStatus::Active, version: 1, description: request.description.clone(), metadata: HashMap::new(), tags: request.tags.clone(), encrypted_key_material: String::new(), // Not used for transit keys }; debug!("VaultKeyData tags before storage: {:?}", key_data.tags); self.store_key_data(key_id, &key_data).await } /// Retrieve key data from Vault async fn get_key_data(&self, key_id: &str) -> Result<VaultKeyData> { let path = self.key_path(key_id); let secret: VaultKeyData = kv2::read(&self.client, &self.kv_mount, &path).await.map_err(|e| match e { vaultrs::error::ClientError::ResponseWrapError => KmsError::key_not_found(key_id), vaultrs::error::ClientError::APIError { code: 404, .. } => KmsError::key_not_found(key_id), _ => KmsError::backend_error(format!("Failed to read key from Vault: {e}")), })?; debug!("Retrieved key {} from Vault, tags: {:?}", key_id, secret.tags); Ok(secret) } /// List all keys stored in Vault async fn list_vault_keys(&self) -> Result<Vec<String>> { // List keys under the prefix match kv2::list(&self.client, &self.kv_mount, &self.key_path_prefix).await { Ok(keys) => { debug!("Found {} keys in Vault", keys.len()); Ok(keys) } Err(vaultrs::error::ClientError::ResponseWrapError) => { // No keys exist yet Ok(Vec::new()) } Err(vaultrs::error::ClientError::APIError { code: 404, .. }) => { // Path doesn't exist - no keys exist yet debug!("Key path doesn't exist in Vault (404), returning empty list"); Ok(Vec::new()) } Err(e) => Err(KmsError::backend_error(format!("Failed to list keys in Vault: {e}"))), } } /// Physically delete a key from Vault storage async fn delete_key(&self, key_id: &str) -> Result<()> { let path = self.key_path(key_id); // For this specific key path, we can safely delete the metadata // since each key has its own unique path under the prefix kv2::delete_metadata(&self.client, &self.kv_mount, &path) .await .map_err(|e| match e { vaultrs::error::ClientError::APIError { code: 404, .. } => KmsError::key_not_found(key_id), _ => KmsError::backend_error(format!("Failed to delete key metadata from Vault: {e}")), })?; debug!("Permanently deleted key {} metadata from Vault at path {}", key_id, path); Ok(()) } } #[async_trait] impl KmsClient for VaultKmsClient { async fn generate_data_key(&self, request: &GenerateKeyRequest, context: Option<&OperationContext>) -> Result<DataKey> { debug!("Generating data key for master key: {}", request.master_key_id); // Verify master key exists let _master_key = self.describe_key(&request.master_key_id, context).await?; // Generate data key material let key_length = match request.key_spec.as_str() { "AES_256" => 32, "AES_128" => 16, _ => return Err(KmsError::unsupported_algorithm(&request.key_spec)), }; let mut plaintext_key = vec![0u8; key_length]; rand::rng().fill_bytes(&mut plaintext_key); // Encrypt the data key with the master key let encrypted_key = self.encrypt_key_material(&plaintext_key).await?; Ok(DataKey { key_id: request.master_key_id.clone(), version: 1, plaintext: Some(plaintext_key), ciphertext: general_purpose::STANDARD .decode(&encrypted_key) .map_err(|e| KmsError::cryptographic_error("decode", e.to_string()))?, key_spec: request.key_spec.clone(), metadata: request.encryption_context.clone(), created_at: chrono::Utc::now(), }) } async fn encrypt(&self, request: &EncryptRequest, _context: Option<&OperationContext>) -> Result<EncryptResponse> { debug!("Encrypting data with key: {}", request.key_id); // Get the master key let key_data = self.get_key_data(&request.key_id).await?; let key_material = self.decrypt_key_material(&key_data.encrypted_key_material).await?; // For simplicity, we'll use a basic encryption approach // In practice, you'd use proper AEAD encryption let mut ciphertext = request.plaintext.clone(); for (i, byte) in ciphertext.iter_mut().enumerate() { *byte ^= key_material[i % key_material.len()]; } Ok(EncryptResponse { ciphertext, key_id: request.key_id.clone(), key_version: key_data.version, algorithm: key_data.algorithm, }) } async fn decrypt(&self, _request: &DecryptRequest, _context: Option<&OperationContext>) -> Result<Vec<u8>> { debug!("Decrypting data"); // For this simple implementation, we assume the key ID is embedded in the ciphertext metadata // In practice, you'd extract this from the ciphertext envelope Err(KmsError::invalid_operation("Decrypt not fully implemented for Vault backend")) } async fn create_key(&self, key_id: &str, algorithm: &str, _context: Option<&OperationContext>) -> Result<MasterKey> { debug!("Creating master key: {} with algorithm: {}", key_id, algorithm); // Check if key already exists if self.get_key_data(key_id).await.is_ok() { return Err(KmsError::key_already_exists(key_id)); } // Generate key material let key_material = Self::generate_key_material(algorithm)?; let encrypted_material = self.encrypt_key_material(&key_material).await?; // Create key data let key_data = VaultKeyData { algorithm: algorithm.to_string(), usage: KeyUsage::EncryptDecrypt, created_at: chrono::Utc::now(), status: KeyStatus::Active, version: 1, description: None, metadata: HashMap::new(), tags: HashMap::new(), encrypted_key_material: encrypted_material, }; // Store in Vault self.store_key_data(key_id, &key_data).await?; let master_key = MasterKey { key_id: key_id.to_string(), version: key_data.version, algorithm: key_data.algorithm.clone(), usage: key_data.usage, status: key_data.status, description: None, // This method doesn't receive description parameter metadata: key_data.metadata.clone(), created_at: key_data.created_at, rotated_at: None, created_by: None, }; info!("Successfully created master key: {}", key_id); Ok(master_key) } async fn describe_key(&self, key_id: &str, _context: Option<&OperationContext>) -> Result<KeyInfo> { debug!("Describing key: {}", key_id); let key_data = self.get_key_data(key_id).await?; Ok(KeyInfo { key_id: key_id.to_string(), description: key_data.description, algorithm: key_data.algorithm, usage: key_data.usage, status: key_data.status, version: key_data.version, metadata: key_data.metadata, tags: key_data.tags, created_at: key_data.created_at, rotated_at: None, created_by: None, }) } async fn list_keys(&self, request: &ListKeysRequest, _context: Option<&OperationContext>) -> Result<ListKeysResponse> { debug!("Listing keys with limit: {:?}", request.limit); let all_keys = self.list_vault_keys().await?; let limit = request.limit.unwrap_or(100) as usize; // Simple pagination implementation let start_idx = request .marker .as_ref() .and_then(|m| all_keys.iter().position(|k| k == m)) .map(|idx| idx + 1) .unwrap_or(0); let end_idx = std::cmp::min(start_idx + limit, all_keys.len()); let keys_page = &all_keys[start_idx..end_idx]; let mut key_infos = Vec::new(); for key_id in keys_page { if let Ok(key_info) = self.describe_key(key_id, None).await { key_infos.push(key_info); } } let next_marker = if end_idx < all_keys.len() { Some(all_keys[end_idx - 1].clone()) } else { None }; Ok(ListKeysResponse { keys: key_infos, next_marker, truncated: end_idx < all_keys.len(), }) } async fn enable_key(&self, key_id: &str, _context: Option<&OperationContext>) -> Result<()> { debug!("Enabling key: {}", key_id); let mut key_data = self.get_key_data(key_id).await?; key_data.status = KeyStatus::Active; self.store_key_data(key_id, &key_data).await?; info!("Enabled key: {}", key_id); Ok(()) } async fn disable_key(&self, key_id: &str, _context: Option<&OperationContext>) -> Result<()> { debug!("Disabling key: {}", key_id); let mut key_data = self.get_key_data(key_id).await?; key_data.status = KeyStatus::Disabled; self.store_key_data(key_id, &key_data).await?; info!("Disabled key: {}", key_id); Ok(()) } async fn schedule_key_deletion( &self, key_id: &str, _pending_window_days: u32, _context: Option<&OperationContext>, ) -> Result<()> { debug!("Scheduling key deletion: {}", key_id); let mut key_data = self.get_key_data(key_id).await?; key_data.status = KeyStatus::PendingDeletion; self.store_key_data(key_id, &key_data).await?; info!("Scheduled key deletion: {}", key_id); Ok(()) } async fn cancel_key_deletion(&self, key_id: &str, _context: Option<&OperationContext>) -> Result<()> { debug!("Canceling key deletion: {}", key_id); let mut key_data = self.get_key_data(key_id).await?; key_data.status = KeyStatus::Active; self.store_key_data(key_id, &key_data).await?; info!("Canceled key deletion: {}", key_id); Ok(()) } async fn rotate_key(&self, key_id: &str, _context: Option<&OperationContext>) -> Result<MasterKey> { debug!("Rotating key: {}", key_id); let mut key_data = self.get_key_data(key_id).await?; key_data.version += 1; // Generate new key material let key_material = Self::generate_key_material(&key_data.algorithm)?; key_data.encrypted_key_material = self.encrypt_key_material(&key_material).await?; self.store_key_data(key_id, &key_data).await?; let master_key = MasterKey { key_id: key_id.to_string(), version: key_data.version, algorithm: key_data.algorithm, usage: key_data.usage, status: key_data.status, description: None, // Rotate preserves existing description (would need key lookup) metadata: key_data.metadata, created_at: key_data.created_at, rotated_at: Some(chrono::Utc::now()), created_by: None, }; info!("Successfully rotated key: {}", key_id); Ok(master_key) } async fn health_check(&self) -> Result<()> { debug!("Performing Vault health check"); // Use list_vault_keys but handle the case where no keys exist (which is normal) match self.list_vault_keys().await { Ok(_) => { debug!("Vault health check passed - successfully listed keys"); Ok(()) } Err(e) => { // Check if the error is specifically about "no keys found" or 404 let error_msg = e.to_string(); if error_msg.contains("status code 404") || error_msg.contains("No such key") { debug!("Vault health check passed - 404 error is expected when no keys exist yet"); Ok(()) } else { warn!("Vault health check failed: {}", e); Err(e) } } } } fn backend_info(&self) -> BackendInfo { BackendInfo::new("vault".to_string(), "0.1.0".to_string(), self.config.address.clone(), true) .with_metadata("kv_mount".to_string(), self.kv_mount.clone()) .with_metadata("key_prefix".to_string(), self.key_path_prefix.clone()) } } /// VaultKmsBackend wraps VaultKmsClient and implements the KmsBackend trait pub struct VaultKmsBackend { client: VaultKmsClient, } impl VaultKmsBackend { /// Create a new VaultKmsBackend pub async fn new(config: KmsConfig) -> Result<Self> { let vault_config = match &config.backend_config { crate::config::BackendConfig::Vault(vault_config) => vault_config.clone(), _ => return Err(KmsError::configuration_error("Expected Vault backend configuration")), }; let client = VaultKmsClient::new(vault_config).await?; Ok(Self { client }) } /// Update key metadata in Vault storage async fn update_key_metadata_in_storage(&self, key_id: &str, metadata: &KeyMetadata) -> Result<()> { // Get the current key data from Vault let mut key_data = self.client.get_key_data(key_id).await?; // Update the status based on the new metadata key_data.status = match metadata.key_state { KeyState::Enabled => KeyStatus::Active, KeyState::Disabled => KeyStatus::Disabled, KeyState::PendingDeletion => KeyStatus::PendingDeletion, KeyState::Unavailable => KeyStatus::Deleted, KeyState::PendingImport => KeyStatus::Disabled, // Treat as disabled until import completes }; // Update the key data in Vault storage self.client.store_key_data(key_id, &key_data).await?; Ok(()) } } #[async_trait] impl KmsBackend for VaultKmsBackend { async fn create_key(&self, request: CreateKeyRequest) -> Result<CreateKeyResponse> { let key_id = request.key_name.clone().unwrap_or_else(|| uuid::Uuid::new_v4().to_string()); // Create key in Vault transit engine let _master_key = self.client.create_key(&key_id, "AES_256", None).await?; // Also store key metadata in KV store with tags self.client.store_key_metadata(&key_id, &request).await?; let metadata = KeyMetadata { key_id: key_id.clone(), key_state: KeyState::Enabled, key_usage: request.key_usage, description: request.description, creation_date: chrono::Utc::now(), deletion_date: None, origin: "VAULT".to_string(), key_manager: "VAULT".to_string(), tags: request.tags, }; Ok(CreateKeyResponse { key_id, key_metadata: metadata, }) } async fn encrypt(&self, request: EncryptRequest) -> Result<EncryptResponse> { let encrypt_request = crate::types::EncryptRequest { key_id: request.key_id.clone(), plaintext: request.plaintext, encryption_context: request.encryption_context, grant_tokens: request.grant_tokens, }; let response = self.client.encrypt(&encrypt_request, None).await?; Ok(EncryptResponse { ciphertext: response.ciphertext, key_id: response.key_id, key_version: response.key_version, algorithm: response.algorithm, }) } async fn decrypt(&self, request: DecryptRequest) -> Result<DecryptResponse> { let plaintext = self.client.decrypt(&request, None).await?; Ok(DecryptResponse { plaintext, key_id: "unknown".to_string(), // Would be extracted from ciphertext metadata encryption_algorithm: Some("AES-256-GCM".to_string()), }) } async fn generate_data_key(&self, request: GenerateDataKeyRequest) -> Result<GenerateDataKeyResponse> { let generate_request = GenerateKeyRequest { master_key_id: request.key_id.clone(), key_spec: request.key_spec.as_str().to_string(), key_length: Some(request.key_spec.key_size() as u32), encryption_context: request.encryption_context, grant_tokens: Vec::new(), }; let data_key = self.client.generate_data_key(&generate_request, None).await?; Ok(GenerateDataKeyResponse { key_id: request.key_id, plaintext_key: data_key.plaintext.clone().unwrap_or_default(), ciphertext_blob: data_key.ciphertext.clone(), }) } async fn describe_key(&self, request: DescribeKeyRequest) -> Result<DescribeKeyResponse> { let key_info = self.client.describe_key(&request.key_id, None).await?; // Also get key metadata from KV store to retrieve tags let key_data = self.client.get_key_data(&request.key_id).await?; let metadata = KeyMetadata { key_id: key_info.key_id, key_state: match key_info.status { KeyStatus::Active => KeyState::Enabled, KeyStatus::Disabled => KeyState::Disabled, KeyStatus::PendingDeletion => KeyState::PendingDeletion, KeyStatus::Deleted => KeyState::Unavailable, }, key_usage: key_info.usage, description: key_info.description, creation_date: key_info.created_at, deletion_date: None, origin: "VAULT".to_string(), key_manager: "VAULT".to_string(), tags: key_data.tags, }; Ok(DescribeKeyResponse { key_metadata: metadata }) } async fn list_keys(&self, request: ListKeysRequest) -> Result<ListKeysResponse> { let response = self.client.list_keys(&request, None).await?; Ok(response) } async fn delete_key(&self, request: DeleteKeyRequest) -> Result<DeleteKeyResponse> { // For Vault backend, we'll mark keys for deletion but not physically delete them // This allows for recovery during the pending window let key_id = &request.key_id; // First, check if the key exists and get its metadata let describe_request = DescribeKeyRequest { key_id: key_id.clone() }; let mut key_metadata = match self.describe_key(describe_request).await { Ok(response) => response.key_metadata, Err(_) => { return Err(crate::error::KmsError::key_not_found(format!("Key {key_id} not found"))); } }; let deletion_date = if request.force_immediate.unwrap_or(false) { // Check if key is already in PendingDeletion state if key_metadata.key_state == KeyState::PendingDeletion { // Force immediate deletion: physically delete the key from Vault storage self.client.delete_key(key_id).await?; // Return empty deletion_date to indicate key was permanently deleted None } else { // For non-pending keys, mark as PendingDeletion key_metadata.key_state = KeyState::PendingDeletion; key_metadata.deletion_date = Some(chrono::Utc::now()); // Update the key metadata in Vault storage to reflect the new state self.update_key_metadata_in_storage(key_id, &key_metadata).await?; None } } else { // Schedule for deletion (default 30 days) let days = request.pending_window_in_days.unwrap_or(30); if !(7..=30).contains(&days) { return Err(crate::error::KmsError::invalid_parameter( "pending_window_in_days must be between 7 and 30".to_string(), )); } let deletion_date = chrono::Utc::now() + chrono::Duration::days(days as i64); key_metadata.key_state = KeyState::PendingDeletion; key_metadata.deletion_date = Some(deletion_date); // Update the key metadata in Vault storage to reflect the new state self.update_key_metadata_in_storage(key_id, &key_metadata).await?; Some(deletion_date.to_rfc3339()) }; Ok(DeleteKeyResponse { key_id: key_id.clone(), deletion_date, key_metadata, }) } async fn cancel_key_deletion(&self, request: CancelKeyDeletionRequest) -> Result<CancelKeyDeletionResponse> { let key_id = &request.key_id; // Check if the key exists and is pending deletion let describe_request = DescribeKeyRequest { key_id: key_id.clone() }; let mut key_metadata = match self.describe_key(describe_request).await { Ok(response) => response.key_metadata, Err(_) => { return Err(crate::error::KmsError::key_not_found(format!("Key {key_id} not found"))); } }; if key_metadata.key_state != KeyState::PendingDeletion { return Err(crate::error::KmsError::invalid_key_state(format!("Key {key_id} is not pending deletion"))); } // Cancel the deletion by resetting the state key_metadata.key_state = KeyState::Enabled; key_metadata.deletion_date = None; Ok(CancelKeyDeletionResponse { key_id: key_id.clone(), key_metadata, }) } async fn health_check(&self) -> Result<bool> { self.client.health_check().await.map(|_| true) } } #[cfg(test)] mod tests { use super::*; use crate::config::{VaultAuthMethod, VaultConfig}; #[tokio::test] #[ignore] // Requires a running Vault instance async fn test_vault_client_integration() { let config = VaultConfig { address: "http://127.0.0.1:8200".to_string(), auth_method: VaultAuthMethod::Token { token: "dev-only-token".to_string(), }, kv_mount: "secret".to_string(), key_path_prefix: "rustfs/kms/keys".to_string(), mount_path: "transit".to_string(), namespace: None, tls: None, }; let client = VaultKmsClient::new(config).await.expect("Failed to create Vault client"); // Test key operations let key_id = "test-key-vault"; let master_key = client .create_key(key_id, "AES_256", None) .await .expect("Failed to create key"); assert_eq!(master_key.key_id, key_id); assert_eq!(master_key.algorithm, "AES_256"); // Test key description let key_info = client.describe_key(key_id, None).await.expect("Failed to describe key"); assert_eq!(key_info.key_id, key_id); // Test data key generation let data_key_request = GenerateKeyRequest { master_key_id: key_id.to_string(), key_spec: "AES_256".to_string(), key_length: Some(32), encryption_context: Default::default(), grant_tokens: Vec::new(), }; let data_key = client .generate_data_key(&data_key_request, None) .await .expect("Failed to generate data key"); assert!(data_key.plaintext.is_some()); assert!(!data_key.ciphertext.is_empty()); // Test health check client.health_check().await.expect("Health check failed"); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/kms/src/backends/local.rs
crates/kms/src/backends/local.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Local file-based KMS backend implementation use crate::backends::{BackendInfo, KmsBackend, KmsClient}; use crate::config::KmsConfig; use crate::config::LocalConfig; use crate::error::{KmsError, Result}; use crate::types::*; use aes_gcm::{ Aes256Gcm, Key, Nonce, aead::{Aead, KeyInit}, }; use async_trait::async_trait; use rand::Rng; use serde::{Deserialize, Serialize}; use std::collections::HashMap; use std::path::PathBuf; use tokio::fs; use tokio::sync::RwLock; use tracing::{debug, info, warn}; /// Local KMS client that stores keys in local files pub struct LocalKmsClient { config: LocalConfig, /// In-memory cache of loaded keys for performance key_cache: RwLock<HashMap<String, MasterKey>>, /// Master encryption key for encrypting stored keys master_cipher: Option<Aes256Gcm>, } /// Serializable representation of a master key stored on disk #[derive(Debug, Clone, Serialize, Deserialize)] struct StoredMasterKey { key_id: String, version: u32, algorithm: String, usage: KeyUsage, status: KeyStatus, description: Option<String>, metadata: HashMap<String, String>, created_at: chrono::DateTime<chrono::Utc>, rotated_at: Option<chrono::DateTime<chrono::Utc>>, created_by: Option<String>, /// Encrypted key material (32 bytes for AES-256) encrypted_key_material: Vec<u8>, /// Nonce used for encryption nonce: Vec<u8>, } /// Data key envelope stored with each data key generation #[derive(Debug, Clone, Serialize, Deserialize)] struct DataKeyEnvelope { key_id: String, master_key_id: String, key_spec: String, encrypted_key: Vec<u8>, nonce: Vec<u8>, encryption_context: HashMap<String, String>, created_at: chrono::DateTime<chrono::Utc>, } impl LocalKmsClient { /// Create a new local KMS client pub async fn new(config: LocalConfig) -> Result<Self> { // Create key directory if it doesn't exist if !config.key_dir.exists() { fs::create_dir_all(&config.key_dir).await?; info!("Created KMS key directory: {:?}", config.key_dir); } // Initialize master cipher if master key is provided let master_cipher = if let Some(ref master_key) = config.master_key { let key = Self::derive_master_key(master_key)?; Some(Aes256Gcm::new(&key)) } else { warn!("No master key provided - stored keys will not be encrypted at rest"); None }; Ok(Self { config, key_cache: RwLock::new(HashMap::new()), master_cipher, }) } /// Derive a 256-bit key from the master key string fn derive_master_key(master_key: &str) -> Result<Key<Aes256Gcm>> { use sha2::{Digest, Sha256}; let mut hasher = Sha256::new(); hasher.update(master_key.as_bytes()); hasher.update(b"rustfs-kms-local"); // Salt to prevent rainbow tables let hash = hasher.finalize(); let key = Key::<Aes256Gcm>::try_from(hash.as_slice()) .map_err(|_| KmsError::cryptographic_error("key", "Invalid key length"))?; Ok(key) } /// Get the file path for a master key fn master_key_path(&self, key_id: &str) -> PathBuf { self.config.key_dir.join(format!("{key_id}.key")) } /// Load a master key from disk async fn load_master_key(&self, key_id: &str) -> Result<MasterKey> { let key_path = self.master_key_path(key_id); if !key_path.exists() { return Err(KmsError::key_not_found(key_id)); } let content = fs::read(&key_path).await?; let stored_key: StoredMasterKey = serde_json::from_slice(&content)?; // Decrypt key material if master cipher is available let _key_material = if let Some(ref cipher) = self.master_cipher { if stored_key.nonce.len() != 12 { return Err(KmsError::cryptographic_error("nonce", "Invalid nonce length")); } let mut nonce_array = [0u8; 12]; nonce_array.copy_from_slice(&stored_key.nonce); let nonce = Nonce::from(nonce_array); cipher .decrypt(&nonce, stored_key.encrypted_key_material.as_ref()) .map_err(|e| KmsError::cryptographic_error("decrypt", e.to_string()))? } else { stored_key.encrypted_key_material }; Ok(MasterKey { key_id: stored_key.key_id, version: stored_key.version, algorithm: stored_key.algorithm, usage: stored_key.usage, status: stored_key.status, description: stored_key.description, metadata: stored_key.metadata, created_at: stored_key.created_at, rotated_at: stored_key.rotated_at, created_by: stored_key.created_by, }) } /// Save a master key to disk async fn save_master_key(&self, master_key: &MasterKey, key_material: &[u8]) -> Result<()> { let key_path = self.master_key_path(&master_key.key_id); // Encrypt key material if master cipher is available let (encrypted_key_material, nonce) = if let Some(ref cipher) = self.master_cipher { let mut nonce_bytes = [0u8; 12]; rand::rng().fill(&mut nonce_bytes[..]); let nonce = Nonce::from(nonce_bytes); let encrypted = cipher .encrypt(&nonce, key_material) .map_err(|e| KmsError::cryptographic_error("encrypt", e.to_string()))?; (encrypted, nonce.to_vec()) } else { (key_material.to_vec(), Vec::new()) }; let stored_key = StoredMasterKey { key_id: master_key.key_id.clone(), version: master_key.version, algorithm: master_key.algorithm.clone(), usage: master_key.usage.clone(), status: master_key.status.clone(), description: master_key.description.clone(), metadata: master_key.metadata.clone(), created_at: master_key.created_at, rotated_at: master_key.rotated_at, created_by: master_key.created_by.clone(), encrypted_key_material, nonce, }; let content = serde_json::to_vec_pretty(&stored_key)?; // Write to temporary file first, then rename for atomicity let temp_path = key_path.with_extension("tmp"); fs::write(&temp_path, &content).await?; // Set file permissions if specified #[cfg(unix)] if let Some(permissions) = self.config.file_permissions { use std::os::unix::fs::PermissionsExt; let perms = std::fs::Permissions::from_mode(permissions); std::fs::set_permissions(&temp_path, perms)?; } fs::rename(&temp_path, &key_path).await?; info!("Saved master key {} to {:?}", master_key.key_id, key_path); Ok(()) } /// Generate a random 256-bit key fn generate_key_material() -> Vec<u8> { let mut key_material = vec![0u8; 32]; // 256 bits rand::rng().fill(&mut key_material[..]); key_material } /// Get the actual key material for a master key async fn get_key_material(&self, key_id: &str) -> Result<Vec<u8>> { let key_path = self.master_key_path(key_id); if !key_path.exists() { return Err(KmsError::key_not_found(key_id)); } let content = fs::read(&key_path).await?; let stored_key: StoredMasterKey = serde_json::from_slice(&content)?; // Decrypt key material if master cipher is available let key_material = if let Some(ref cipher) = self.master_cipher { if stored_key.nonce.len() != 12 { return Err(KmsError::cryptographic_error("nonce", "Invalid nonce length")); } let mut nonce_array = [0u8; 12]; nonce_array.copy_from_slice(&stored_key.nonce); let nonce = Nonce::from(nonce_array); cipher .decrypt(&nonce, stored_key.encrypted_key_material.as_ref()) .map_err(|e| KmsError::cryptographic_error("decrypt", e.to_string()))? } else { stored_key.encrypted_key_material }; Ok(key_material) } /// Encrypt data using a master key async fn encrypt_with_master_key(&self, key_id: &str, plaintext: &[u8]) -> Result<(Vec<u8>, Vec<u8>)> { // Load the actual master key material let key_material = self.get_key_material(key_id).await?; let key = Key::<Aes256Gcm>::try_from(key_material.as_slice()) .map_err(|_| KmsError::cryptographic_error("key", "Invalid key length"))?; let cipher = Aes256Gcm::new(&key); let mut nonce_bytes = [0u8; 12]; rand::rng().fill(&mut nonce_bytes[..]); let nonce = Nonce::from(nonce_bytes); let ciphertext = cipher .encrypt(&nonce, plaintext) .map_err(|e| KmsError::cryptographic_error("encrypt", e.to_string()))?; Ok((ciphertext, nonce_bytes.to_vec())) } /// Decrypt data using a master key async fn decrypt_with_master_key(&self, key_id: &str, ciphertext: &[u8], nonce: &[u8]) -> Result<Vec<u8>> { if nonce.len() != 12 { return Err(KmsError::cryptographic_error("nonce", "Invalid nonce length")); } // Load the actual master key material let key_material = self.get_key_material(key_id).await?; let key = Key::<Aes256Gcm>::try_from(key_material.as_slice()) .map_err(|_| KmsError::cryptographic_error("key", "Invalid key length"))?; let cipher = Aes256Gcm::new(&key); let mut nonce_array = [0u8; 12]; nonce_array.copy_from_slice(nonce); let nonce_ref = Nonce::from(nonce_array); let plaintext = cipher .decrypt(&nonce_ref, ciphertext) .map_err(|e| KmsError::cryptographic_error("decrypt", e.to_string()))?; Ok(plaintext) } } #[async_trait] impl KmsClient for LocalKmsClient { async fn generate_data_key(&self, request: &GenerateKeyRequest, context: Option<&OperationContext>) -> Result<DataKey> { debug!("Generating data key for master key: {}", request.master_key_id); // Verify master key exists let _master_key = self.describe_key(&request.master_key_id, context).await?; // Generate random data key material let key_length = match request.key_spec.as_str() { "AES_256" => 32, "AES_128" => 16, _ => return Err(KmsError::unsupported_algorithm(&request.key_spec)), }; let mut plaintext_key = vec![0u8; key_length]; rand::rng().fill(&mut plaintext_key[..]); // Encrypt the data key with the master key let (encrypted_key, nonce) = self.encrypt_with_master_key(&request.master_key_id, &plaintext_key).await?; // Create data key envelope let envelope = DataKeyEnvelope { key_id: uuid::Uuid::new_v4().to_string(), master_key_id: request.master_key_id.clone(), key_spec: request.key_spec.clone(), encrypted_key: encrypted_key.clone(), nonce, encryption_context: request.encryption_context.clone(), created_at: chrono::Utc::now(), }; // Serialize the envelope as the ciphertext let ciphertext = serde_json::to_vec(&envelope)?; let data_key = DataKey::new(envelope.key_id, 1, Some(plaintext_key), ciphertext, request.key_spec.clone()); info!("Generated data key for master key: {}", request.master_key_id); Ok(data_key) } async fn encrypt(&self, request: &EncryptRequest, context: Option<&OperationContext>) -> Result<EncryptResponse> { debug!("Encrypting data with key: {}", request.key_id); // Verify key exists and is active let key_info = self.describe_key(&request.key_id, context).await?; if key_info.status != KeyStatus::Active { return Err(KmsError::invalid_operation(format!( "Key {} is not active (status: {:?})", request.key_id, key_info.status ))); } let (ciphertext, _nonce) = self.encrypt_with_master_key(&request.key_id, &request.plaintext).await?; Ok(EncryptResponse { ciphertext, key_id: request.key_id.clone(), key_version: key_info.version, algorithm: key_info.algorithm, }) } async fn decrypt(&self, request: &DecryptRequest, _context: Option<&OperationContext>) -> Result<Vec<u8>> { debug!("Decrypting data"); // Parse the data key envelope from ciphertext let envelope: DataKeyEnvelope = serde_json::from_slice(&request.ciphertext)?; // Verify encryption context matches if !request.encryption_context.is_empty() { for (key, expected_value) in &request.encryption_context { if let Some(actual_value) = envelope.encryption_context.get(key) { if actual_value != expected_value { return Err(KmsError::context_mismatch(format!( "Context mismatch for key '{key}': expected '{expected_value}', got '{actual_value}'" ))); } } else { return Err(KmsError::context_mismatch(format!("Missing context key '{key}'"))); } } } // Decrypt the data key let plaintext = self .decrypt_with_master_key(&envelope.master_key_id, &envelope.encrypted_key, &envelope.nonce) .await?; info!("Successfully decrypted data"); Ok(plaintext) } async fn create_key(&self, key_id: &str, algorithm: &str, context: Option<&OperationContext>) -> Result<MasterKey> { debug!("Creating master key: {}", key_id); // Check if key already exists if self.master_key_path(key_id).exists() { return Err(KmsError::key_already_exists(key_id)); } // Validate algorithm if algorithm != "AES_256" { return Err(KmsError::unsupported_algorithm(algorithm)); } // Generate key material let key_material = Self::generate_key_material(); let created_by = context .map(|ctx| ctx.principal.clone()) .unwrap_or_else(|| "local-kms".to_string()); let master_key = MasterKey::new_with_description(key_id.to_string(), algorithm.to_string(), Some(created_by), None); // Save to disk self.save_master_key(&master_key, &key_material).await?; // Cache the key let mut cache = self.key_cache.write().await; cache.insert(key_id.to_string(), master_key.clone()); info!("Created master key: {}", key_id); Ok(master_key) } async fn describe_key(&self, key_id: &str, _context: Option<&OperationContext>) -> Result<KeyInfo> { debug!("Describing key: {}", key_id); // Check cache first { let cache = self.key_cache.read().await; if let Some(master_key) = cache.get(key_id) { return Ok(master_key.clone().into()); } } // Load from disk let master_key = self.load_master_key(key_id).await?; // Update cache { let mut cache = self.key_cache.write().await; cache.insert(key_id.to_string(), master_key.clone()); } Ok(master_key.into()) } async fn list_keys(&self, request: &ListKeysRequest, _context: Option<&OperationContext>) -> Result<ListKeysResponse> { debug!("Listing keys"); let mut keys = Vec::new(); let limit = request.limit.unwrap_or(100) as usize; let mut count = 0; let mut entries = fs::read_dir(&self.config.key_dir).await?; while let Some(entry) = entries.next_entry().await? { if count >= limit { break; } let path = entry.path(); if path.extension().is_some_and(|ext| ext == "key") && let Some(stem) = path.file_stem() && let Some(key_id) = stem.to_str() && let Ok(key_info) = self.describe_key(key_id, None).await { // Apply filters if let Some(ref status_filter) = request.status_filter && &key_info.status != status_filter { continue; } if let Some(ref usage_filter) = request.usage_filter && &key_info.usage != usage_filter { continue; } keys.push(key_info); count += 1; } } Ok(ListKeysResponse { keys, next_marker: None, // Simple implementation without pagination truncated: false, }) } async fn enable_key(&self, key_id: &str, _context: Option<&OperationContext>) -> Result<()> { debug!("Enabling key: {}", key_id); let mut master_key = self.load_master_key(key_id).await?; master_key.status = KeyStatus::Active; // For simplicity, we'll regenerate key material // In a real implementation, we'd preserve the original key material let key_material = Self::generate_key_material(); self.save_master_key(&master_key, &key_material).await?; // Update cache let mut cache = self.key_cache.write().await; cache.insert(key_id.to_string(), master_key); info!("Enabled key: {}", key_id); Ok(()) } async fn disable_key(&self, key_id: &str, _context: Option<&OperationContext>) -> Result<()> { debug!("Disabling key: {}", key_id); let mut master_key = self.load_master_key(key_id).await?; master_key.status = KeyStatus::Disabled; let key_material = Self::generate_key_material(); self.save_master_key(&master_key, &key_material).await?; // Update cache let mut cache = self.key_cache.write().await; cache.insert(key_id.to_string(), master_key); info!("Disabled key: {}", key_id); Ok(()) } async fn schedule_key_deletion( &self, key_id: &str, _pending_window_days: u32, _context: Option<&OperationContext>, ) -> Result<()> { debug!("Scheduling deletion for key: {}", key_id); let mut master_key = self.load_master_key(key_id).await?; master_key.status = KeyStatus::PendingDeletion; let key_material = Self::generate_key_material(); self.save_master_key(&master_key, &key_material).await?; // Update cache let mut cache = self.key_cache.write().await; cache.insert(key_id.to_string(), master_key); warn!("Scheduled key deletion: {}", key_id); Ok(()) } async fn cancel_key_deletion(&self, key_id: &str, _context: Option<&OperationContext>) -> Result<()> { debug!("Canceling deletion for key: {}", key_id); let mut master_key = self.load_master_key(key_id).await?; master_key.status = KeyStatus::Active; let key_material = Self::generate_key_material(); self.save_master_key(&master_key, &key_material).await?; // Update cache let mut cache = self.key_cache.write().await; cache.insert(key_id.to_string(), master_key); info!("Canceled deletion for key: {}", key_id); Ok(()) } async fn rotate_key(&self, key_id: &str, _context: Option<&OperationContext>) -> Result<MasterKey> { debug!("Rotating key: {}", key_id); let mut master_key = self.load_master_key(key_id).await?; master_key.version += 1; master_key.rotated_at = Some(chrono::Utc::now()); // Generate new key material let key_material = Self::generate_key_material(); self.save_master_key(&master_key, &key_material).await?; // Update cache let mut cache = self.key_cache.write().await; cache.insert(key_id.to_string(), master_key.clone()); info!("Rotated key: {}", key_id); Ok(master_key) } async fn health_check(&self) -> Result<()> { // Check if key directory is accessible if !self.config.key_dir.exists() { return Err(KmsError::backend_error("Key directory does not exist")); } // Try to read the directory let _ = fs::read_dir(&self.config.key_dir).await?; Ok(()) } fn backend_info(&self) -> BackendInfo { BackendInfo::new( "local".to_string(), env!("CARGO_PKG_VERSION").to_string(), self.config.key_dir.to_string_lossy().to_string(), true, // We'll assume healthy for now ) .with_metadata("key_dir".to_string(), self.config.key_dir.to_string_lossy().to_string()) .with_metadata("encrypted_at_rest".to_string(), self.master_cipher.is_some().to_string()) } } /// LocalKmsBackend wraps LocalKmsClient and implements the KmsBackend trait pub struct LocalKmsBackend { client: LocalKmsClient, } impl LocalKmsBackend { /// Create a new LocalKmsBackend pub async fn new(config: KmsConfig) -> Result<Self> { let local_config = match &config.backend_config { crate::config::BackendConfig::Local(local_config) => local_config.clone(), _ => return Err(KmsError::configuration_error("Expected Local backend configuration")), }; let client = LocalKmsClient::new(local_config).await?; Ok(Self { client }) } } #[async_trait] impl KmsBackend for LocalKmsBackend { async fn create_key(&self, request: CreateKeyRequest) -> Result<CreateKeyResponse> { let key_id = request.key_name.unwrap_or_else(|| uuid::Uuid::new_v4().to_string()); // Create master key with description directly let _master_key = { // Generate key material let key_material = LocalKmsClient::generate_key_material(); let master_key = MasterKey::new_with_description( key_id.clone(), "AES_256".to_string(), Some("local-kms".to_string()), request.description.clone(), ); // Save to disk and cache self.client.save_master_key(&master_key, &key_material).await?; let mut cache = self.client.key_cache.write().await; cache.insert(key_id.clone(), master_key.clone()); master_key }; let metadata = KeyMetadata { key_id: key_id.clone(), key_state: KeyState::Enabled, key_usage: request.key_usage, description: request.description, creation_date: chrono::Utc::now(), deletion_date: None, origin: "KMS".to_string(), key_manager: "CUSTOMER".to_string(), tags: request.tags, }; Ok(CreateKeyResponse { key_id, key_metadata: metadata, }) } async fn encrypt(&self, request: EncryptRequest) -> Result<EncryptResponse> { let encrypt_request = EncryptRequest { key_id: request.key_id.clone(), plaintext: request.plaintext, encryption_context: request.encryption_context, grant_tokens: request.grant_tokens, }; let response = self.client.encrypt(&encrypt_request, None).await?; Ok(EncryptResponse { ciphertext: response.ciphertext, key_id: response.key_id, key_version: response.key_version, algorithm: response.algorithm, }) } async fn decrypt(&self, request: DecryptRequest) -> Result<DecryptResponse> { let plaintext = self.client.decrypt(&request, None).await?; // For simplicity, return basic response - in real implementation would extract more info from ciphertext Ok(DecryptResponse { plaintext, key_id: "unknown".to_string(), // Would be extracted from ciphertext metadata encryption_algorithm: Some("AES-256-GCM".to_string()), }) } async fn generate_data_key(&self, request: GenerateDataKeyRequest) -> Result<GenerateDataKeyResponse> { let generate_request = GenerateKeyRequest { master_key_id: request.key_id.clone(), key_spec: request.key_spec.as_str().to_string(), key_length: Some(request.key_spec.key_size() as u32), encryption_context: request.encryption_context, grant_tokens: Vec::new(), }; let data_key = self.client.generate_data_key(&generate_request, None).await?; Ok(GenerateDataKeyResponse { key_id: request.key_id, plaintext_key: data_key.plaintext.clone().unwrap_or_default(), ciphertext_blob: data_key.ciphertext.clone(), }) } async fn describe_key(&self, request: DescribeKeyRequest) -> Result<DescribeKeyResponse> { let key_info = self.client.describe_key(&request.key_id, None).await?; let metadata = KeyMetadata { key_id: key_info.key_id, key_state: match key_info.status { KeyStatus::Active => KeyState::Enabled, KeyStatus::Disabled => KeyState::Disabled, KeyStatus::PendingDeletion => KeyState::PendingDeletion, KeyStatus::Deleted => KeyState::Unavailable, }, key_usage: key_info.usage, description: key_info.description, creation_date: key_info.created_at, deletion_date: None, origin: "KMS".to_string(), key_manager: "CUSTOMER".to_string(), tags: key_info.tags, }; Ok(DescribeKeyResponse { key_metadata: metadata }) } async fn list_keys(&self, request: ListKeysRequest) -> Result<ListKeysResponse> { let response = self.client.list_keys(&request, None).await?; Ok(response) } async fn delete_key(&self, request: DeleteKeyRequest) -> Result<DeleteKeyResponse> { // For local backend, we'll implement immediate deletion by default // unless a pending window is specified let key_id = &request.key_id; // First, load the key from disk to get the master key let mut master_key = self .client .load_master_key(key_id) .await .map_err(|_| KmsError::key_not_found(format!("Key {key_id} not found")))?; let (deletion_date_str, deletion_date_dt) = if request.force_immediate.unwrap_or(false) { // For immediate deletion, actually delete the key from filesystem let key_path = self.client.master_key_path(key_id); tokio::fs::remove_file(&key_path) .await .map_err(|e| KmsError::internal_error(format!("Failed to delete key file: {e}")))?; // Remove from cache let mut cache = self.client.key_cache.write().await; cache.remove(key_id); info!("Immediately deleted key: {}", key_id); // Return success response for immediate deletion let key_metadata = KeyMetadata { key_id: master_key.key_id.clone(), description: master_key.description.clone(), key_usage: master_key.usage, key_state: KeyState::PendingDeletion, // AWS KMS compatibility creation_date: master_key.created_at, deletion_date: Some(chrono::Utc::now()), key_manager: "CUSTOMER".to_string(), origin: "AWS_KMS".to_string(), tags: master_key.metadata, }; return Ok(DeleteKeyResponse { key_id: key_id.clone(), deletion_date: None, // No deletion date for immediate deletion key_metadata, }); } else { // Schedule for deletion (default 30 days) let days = request.pending_window_in_days.unwrap_or(30); if !(7..=30).contains(&days) { return Err(KmsError::invalid_parameter("pending_window_in_days must be between 7 and 30".to_string())); } let deletion_date = chrono::Utc::now() + chrono::Duration::days(days as i64); master_key.status = KeyStatus::PendingDeletion; (Some(deletion_date.to_rfc3339()), Some(deletion_date)) }; // Save the updated key to disk - preserve existing key material! // Load the stored key from disk to get the existing key material let key_path = self.client.master_key_path(key_id); let content = tokio::fs::read(&key_path) .await .map_err(|e| KmsError::internal_error(format!("Failed to read key file: {e}")))?; let stored_key: StoredMasterKey = serde_json::from_slice(&content).map_err(|e| KmsError::internal_error(format!("Failed to parse stored key: {e}")))?; // Decrypt the existing key material to preserve it let existing_key_material = if let Some(ref cipher) = self.client.master_cipher { if stored_key.nonce.len() != 12 { return Err(KmsError::cryptographic_error("nonce", "Invalid nonce length")); } let mut nonce_array = [0u8; 12]; nonce_array.copy_from_slice(&stored_key.nonce); let nonce = Nonce::from(nonce_array); cipher .decrypt(&nonce, stored_key.encrypted_key_material.as_ref()) .map_err(|e| KmsError::cryptographic_error("decrypt", e.to_string()))? } else { stored_key.encrypted_key_material }; self.client.save_master_key(&master_key, &existing_key_material).await?; // Update cache let mut cache = self.client.key_cache.write().await; cache.insert(key_id.to_string(), master_key.clone()); // Convert master_key to KeyMetadata for response let key_metadata = KeyMetadata { key_id: master_key.key_id.clone(), description: master_key.description.clone(), key_usage: master_key.usage, key_state: KeyState::PendingDeletion, creation_date: master_key.created_at, deletion_date: deletion_date_dt, key_manager: "CUSTOMER".to_string(), origin: "AWS_KMS".to_string(), tags: master_key.metadata, }; Ok(DeleteKeyResponse { key_id: key_id.clone(), deletion_date: deletion_date_str, key_metadata, }) } async fn cancel_key_deletion(&self, request: CancelKeyDeletionRequest) -> Result<CancelKeyDeletionResponse> { let key_id = &request.key_id; // Load the key from disk to get the master key let mut master_key = self .client .load_master_key(key_id) .await .map_err(|_| KmsError::key_not_found(format!("Key {key_id} not found")))?; if master_key.status != KeyStatus::PendingDeletion { return Err(KmsError::invalid_key_state(format!("Key {key_id} is not pending deletion"))); } // Cancel the deletion by resetting the state master_key.status = KeyStatus::Active; // Save the updated key to disk - this is the missing critical step! let key_material = LocalKmsClient::generate_key_material(); self.client.save_master_key(&master_key, &key_material).await?; // Update cache let mut cache = self.client.key_cache.write().await; cache.insert(key_id.to_string(), master_key.clone()); // Convert master_key to KeyMetadata for response let key_metadata = KeyMetadata {
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
true
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/kms/src/backends/mod.rs
crates/kms/src/backends/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! KMS backend implementations use crate::error::Result; use crate::types::*; use async_trait::async_trait; use std::collections::HashMap; pub mod local; pub mod vault; /// Abstract KMS client interface that all backends must implement #[async_trait] pub trait KmsClient: Send + Sync { /// Generate a new data encryption key (DEK) /// /// Creates a new data key using the specified master key. The returned DataKey /// contains both the plaintext and encrypted versions of the key. /// /// # Arguments /// * `request` - The key generation request /// * `context` - Optional operation context for auditing /// /// # Returns /// Returns a DataKey containing both plaintext and encrypted key material async fn generate_data_key(&self, request: &GenerateKeyRequest, context: Option<&OperationContext>) -> Result<DataKey>; /// Encrypt data directly using a master key /// /// Encrypts the provided plaintext using the specified master key. /// This is different from generate_data_key as it encrypts user data directly. /// /// # Arguments /// * `request` - The encryption request containing plaintext and key ID /// * `context` - Optional operation context for auditing async fn encrypt(&self, request: &EncryptRequest, context: Option<&OperationContext>) -> Result<EncryptResponse>; /// Decrypt data using a master key /// /// Decrypts the provided ciphertext. The KMS automatically determines /// which key was used for encryption based on the ciphertext metadata. /// /// # Arguments /// * `request` - The decryption request containing ciphertext /// * `context` - Optional operation context for auditing async fn decrypt(&self, request: &DecryptRequest, context: Option<&OperationContext>) -> Result<Vec<u8>>; /// Create a new master key /// /// Creates a new master key in the KMS with the specified ID. /// Returns an error if a key with the same ID already exists. /// /// # Arguments /// * `key_id` - Unique identifier for the new key /// * `algorithm` - Key algorithm (e.g., "AES_256") /// * `context` - Optional operation context for auditing async fn create_key(&self, key_id: &str, algorithm: &str, context: Option<&OperationContext>) -> Result<MasterKey>; /// Get information about a specific key /// /// Returns metadata and information about the specified key. /// /// # Arguments /// * `key_id` - The key identifier /// * `context` - Optional operation context for auditing async fn describe_key(&self, key_id: &str, context: Option<&OperationContext>) -> Result<KeyInfo>; /// List available keys /// /// Returns a paginated list of keys available in the KMS. /// /// # Arguments /// * `request` - List request parameters (pagination, filters) /// * `context` - Optional operation context for auditing async fn list_keys(&self, request: &ListKeysRequest, context: Option<&OperationContext>) -> Result<ListKeysResponse>; /// Enable a key /// /// Enables a previously disabled key, allowing it to be used for cryptographic operations. /// /// # Arguments /// * `key_id` - The key identifier /// * `context` - Optional operation context for auditing async fn enable_key(&self, key_id: &str, context: Option<&OperationContext>) -> Result<()>; /// Disable a key /// /// Disables a key, preventing it from being used for new cryptographic operations. /// Existing encrypted data can still be decrypted. /// /// # Arguments /// * `key_id` - The key identifier /// * `context` - Optional operation context for auditing async fn disable_key(&self, key_id: &str, context: Option<&OperationContext>) -> Result<()>; /// Schedule key deletion /// /// Schedules a key for deletion after a specified number of days. /// This allows for a grace period to recover the key if needed. /// /// # Arguments /// * `key_id` - The key identifier /// * `pending_window_days` - Number of days before actual deletion /// * `context` - Optional operation context for auditing async fn schedule_key_deletion( &self, key_id: &str, pending_window_days: u32, context: Option<&OperationContext>, ) -> Result<()>; /// Cancel key deletion /// /// Cancels a previously scheduled key deletion. /// /// # Arguments /// * `key_id` - The key identifier /// * `context` - Optional operation context for auditing async fn cancel_key_deletion(&self, key_id: &str, context: Option<&OperationContext>) -> Result<()>; /// Rotate a key /// /// Creates a new version of the specified key. Previous versions remain /// available for decryption but new operations will use the new version. /// /// # Arguments /// * `key_id` - The key identifier /// * `context` - Optional operation context for auditing async fn rotate_key(&self, key_id: &str, context: Option<&OperationContext>) -> Result<MasterKey>; /// Health check /// /// Performs a health check on the KMS backend to ensure it's operational. async fn health_check(&self) -> Result<()>; /// Get backend information /// /// Returns information about the KMS backend (type, version, etc.). fn backend_info(&self) -> BackendInfo; } /// Simplified KMS backend interface for manager #[async_trait] pub trait KmsBackend: Send + Sync { /// Create a new master key async fn create_key(&self, request: CreateKeyRequest) -> Result<CreateKeyResponse>; /// Encrypt data async fn encrypt(&self, request: EncryptRequest) -> Result<EncryptResponse>; /// Decrypt data async fn decrypt(&self, request: DecryptRequest) -> Result<DecryptResponse>; /// Generate a data key async fn generate_data_key(&self, request: GenerateDataKeyRequest) -> Result<GenerateDataKeyResponse>; /// Describe a key async fn describe_key(&self, request: DescribeKeyRequest) -> Result<DescribeKeyResponse>; /// List keys async fn list_keys(&self, request: ListKeysRequest) -> Result<ListKeysResponse>; /// Delete a key async fn delete_key(&self, request: DeleteKeyRequest) -> Result<DeleteKeyResponse>; /// Cancel key deletion async fn cancel_key_deletion(&self, request: CancelKeyDeletionRequest) -> Result<CancelKeyDeletionResponse>; /// Health check async fn health_check(&self) -> Result<bool>; } /// Information about a KMS backend #[derive(Debug, Clone)] pub struct BackendInfo { /// Backend type name (e.g., "local", "vault") pub backend_type: String, /// Backend version pub version: String, /// Backend endpoint or location pub endpoint: String, /// Whether the backend is currently healthy pub healthy: bool, /// Additional metadata about the backend pub metadata: HashMap<String, String>, } impl BackendInfo { /// Create a new backend info /// /// # Arguments /// * `backend_type` - The type of the backend /// * `version` - The version of the backend /// * `endpoint` - The endpoint or location of the backend /// * `healthy` - Whether the backend is healthy /// /// # Returns /// A new BackendInfo instance /// pub fn new(backend_type: String, version: String, endpoint: String, healthy: bool) -> Self { Self { backend_type, version, endpoint, healthy, metadata: HashMap::new(), } } /// Add metadata to the backend info /// /// # Arguments /// * `key` - Metadata key /// * `value` - Metadata value /// /// # Returns /// Updated BackendInfo instance /// pub fn with_metadata(mut self, key: String, value: String) -> Self { self.metadata.insert(key, value); self } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/path.rs
crates/utils/src/path.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::path::Path; use std::path::PathBuf; pub const GLOBAL_DIR_SUFFIX: &str = "__XLDIR__"; pub const SLASH_SEPARATOR: &str = "/"; pub const GLOBAL_DIR_SUFFIX_WITH_SLASH: &str = "__XLDIR__/"; pub fn has_suffix(s: &str, suffix: &str) -> bool { if cfg!(target_os = "windows") { s.to_lowercase().ends_with(&suffix.to_lowercase()) } else { s.ends_with(suffix) } } pub fn encode_dir_object(object: &str) -> String { if has_suffix(object, SLASH_SEPARATOR) { format!("{}{}", object.trim_end_matches(SLASH_SEPARATOR), GLOBAL_DIR_SUFFIX) } else { object.to_string() } } pub fn is_dir_object(object: &str) -> bool { let obj = encode_dir_object(object); obj.ends_with(GLOBAL_DIR_SUFFIX) } #[allow(dead_code)] pub fn decode_dir_object(object: &str) -> String { if has_suffix(object, GLOBAL_DIR_SUFFIX) { format!("{}{}", object.trim_end_matches(GLOBAL_DIR_SUFFIX), SLASH_SEPARATOR) } else { object.to_string() } } pub fn retain_slash(s: &str) -> String { if s.is_empty() { return s.to_string(); } if s.ends_with(SLASH_SEPARATOR) { s.to_string() } else { format!("{s}{SLASH_SEPARATOR}") } } pub fn strings_has_prefix_fold(s: &str, prefix: &str) -> bool { s.len() >= prefix.len() && (s[..prefix.len()] == *prefix || s[..prefix.len()].eq_ignore_ascii_case(prefix)) } pub fn has_prefix(s: &str, prefix: &str) -> bool { if cfg!(target_os = "windows") { return strings_has_prefix_fold(s, prefix); } s.starts_with(prefix) } pub fn path_join(elem: &[PathBuf]) -> PathBuf { let mut joined_path = PathBuf::new(); for path in elem { joined_path.push(path); } joined_path } pub fn path_join_buf(elements: &[&str]) -> String { let trailing_slash = !elements.is_empty() && elements.last().unwrap().ends_with(SLASH_SEPARATOR); let mut dst = String::new(); let mut added = 0; for e in elements { if added > 0 || !e.is_empty() { if added > 0 { dst.push_str(SLASH_SEPARATOR); } dst.push_str(e); added += e.len(); } } let result = dst.to_string(); let cpath = Path::new(&result).components().collect::<PathBuf>(); let clean_path = cpath.to_string_lossy(); if trailing_slash { return format!("{clean_path}{SLASH_SEPARATOR}"); } clean_path.to_string() } pub fn path_to_bucket_object_with_base_path(bash_path: &str, path: &str) -> (String, String) { let path = path.trim_start_matches(bash_path).trim_start_matches(SLASH_SEPARATOR); if let Some(m) = path.find(SLASH_SEPARATOR) { return (path[..m].to_string(), path[m + SLASH_SEPARATOR.len()..].to_string()); } (path.to_string(), "".to_string()) } pub fn path_to_bucket_object(s: &str) -> (String, String) { path_to_bucket_object_with_base_path("", s) } pub fn base_dir_from_prefix(prefix: &str) -> String { let mut base_dir = dir(prefix).to_owned(); if base_dir == "." || base_dir == "./" || base_dir == "/" { base_dir = "".to_owned(); } if !prefix.contains('/') { base_dir = "".to_owned(); } if !base_dir.is_empty() && !base_dir.ends_with(SLASH_SEPARATOR) { base_dir.push_str(SLASH_SEPARATOR); } base_dir } pub struct LazyBuf { s: String, buf: Option<Vec<u8>>, w: usize, } impl LazyBuf { pub fn new(s: String) -> Self { LazyBuf { s, buf: None, w: 0 } } pub fn index(&self, i: usize) -> u8 { if let Some(ref buf) = self.buf { buf[i] } else { self.s.as_bytes()[i] } } pub fn append(&mut self, c: u8) { if self.buf.is_none() { if self.w < self.s.len() && self.s.as_bytes()[self.w] == c { self.w += 1; return; } let mut new_buf = vec![0; self.s.len()]; new_buf[..self.w].copy_from_slice(&self.s.as_bytes()[..self.w]); self.buf = Some(new_buf); } if let Some(ref mut buf) = self.buf { buf[self.w] = c; self.w += 1; } } pub fn string(&self) -> String { if let Some(ref buf) = self.buf { String::from_utf8(buf[..self.w].to_vec()).unwrap() } else { self.s[..self.w].to_string() } } } pub fn clean(path: &str) -> String { if path.is_empty() { return ".".to_string(); } let rooted = path.starts_with('/'); let n = path.len(); let mut out = LazyBuf::new(path.to_string()); let mut r = 0; let mut dotdot = 0; if rooted { out.append(b'/'); r = 1; dotdot = 1; } while r < n { match path.as_bytes()[r] { b'/' => { // Empty path element r += 1; } b'.' if r + 1 == n || path.as_bytes()[r + 1] == b'/' => { // . element r += 1; } b'.' if path.as_bytes()[r + 1] == b'.' && (r + 2 == n || path.as_bytes()[r + 2] == b'/') => { // .. element: remove to last / r += 2; if out.w > dotdot { // Can backtrack out.w -= 1; while out.w > dotdot && out.index(out.w) != b'/' { out.w -= 1; } } else if !rooted { // Cannot backtrack but not rooted, so append .. element. if out.w > 0 { out.append(b'/'); } out.append(b'.'); out.append(b'.'); dotdot = out.w; } } _ => { // Real path element. // Add slash if needed if (rooted && out.w != 1) || (!rooted && out.w != 0) { out.append(b'/'); } // Copy element while r < n && path.as_bytes()[r] != b'/' { out.append(path.as_bytes()[r]); r += 1; } } } } // Turn empty string into "." if out.w == 0 { return ".".to_string(); } out.string() } pub fn split(path: &str) -> (&str, &str) { // Find the last occurrence of the '/' character if let Some(i) = path.rfind('/') { // Return the directory (up to and including the last '/') and the file name return (&path[..i + 1], &path[i + 1..]); } // If no '/' is found, return an empty string for the directory and the whole path as the file name (path, "") } pub fn dir(path: &str) -> String { let (a, _) = split(path); clean(a) } pub fn trim_etag(etag: &str) -> String { etag.trim_matches('"').to_string() } #[cfg(test)] mod tests { use super::*; #[test] fn test_trim_etag() { // Test with quoted ETag assert_eq!(trim_etag("\"abc123\""), "abc123"); // Test with unquoted ETag assert_eq!(trim_etag("abc123"), "abc123"); // Test with empty string assert_eq!(trim_etag(""), ""); // Test with only quotes assert_eq!(trim_etag("\"\""), ""); // Test with MD5 hash assert_eq!(trim_etag("\"2c7ab85a893283e98c931e9511add182\""), "2c7ab85a893283e98c931e9511add182"); // Test with multipart ETag format assert_eq!(trim_etag("\"098f6bcd4621d373cade4e832627b4f6-2\""), "098f6bcd4621d373cade4e832627b4f6-2"); } #[test] fn test_base_dir_from_prefix() { let a = "da/"; // Test base_dir_from_prefix function let result = base_dir_from_prefix(a); assert!(!result.is_empty()); } #[test] fn test_clean() { assert_eq!(clean(""), "."); assert_eq!(clean("abc"), "abc"); assert_eq!(clean("abc/def"), "abc/def"); assert_eq!(clean("a/b/c"), "a/b/c"); assert_eq!(clean("."), "."); assert_eq!(clean(".."), ".."); assert_eq!(clean("../.."), "../.."); assert_eq!(clean("../../abc"), "../../abc"); assert_eq!(clean("/abc"), "/abc"); assert_eq!(clean("/"), "/"); assert_eq!(clean("abc/"), "abc"); assert_eq!(clean("abc/def/"), "abc/def"); assert_eq!(clean("a/b/c/"), "a/b/c"); assert_eq!(clean("./"), "."); assert_eq!(clean("../"), ".."); assert_eq!(clean("../../"), "../.."); assert_eq!(clean("/abc/"), "/abc"); assert_eq!(clean("abc//def//ghi"), "abc/def/ghi"); assert_eq!(clean("//abc"), "/abc"); assert_eq!(clean("///abc"), "/abc"); assert_eq!(clean("//abc//"), "/abc"); assert_eq!(clean("abc//"), "abc"); assert_eq!(clean("abc/./def"), "abc/def"); assert_eq!(clean("/./abc/def"), "/abc/def"); assert_eq!(clean("abc/."), "abc"); assert_eq!(clean("abc/./../def"), "def"); assert_eq!(clean("abc//./../def"), "def"); assert_eq!(clean("abc/../../././../def"), "../../def"); assert_eq!(clean("abc/def/ghi/../jkl"), "abc/def/jkl"); assert_eq!(clean("abc/def/../ghi/../jkl"), "abc/jkl"); assert_eq!(clean("abc/def/.."), "abc"); assert_eq!(clean("abc/def/../.."), "."); assert_eq!(clean("/abc/def/../.."), "/"); assert_eq!(clean("abc/def/../../.."), ".."); assert_eq!(clean("/abc/def/../../.."), "/"); assert_eq!(clean("abc/def/../../../ghi/jkl/../../../mno"), "../../mno"); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/compress.rs
crates/utils/src/compress.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::io::Write; use std::{fmt, str}; use tokio::io; #[derive(Debug, Clone, Copy, PartialEq, Eq, Hash, Default)] pub enum CompressionAlgorithm { None, Gzip, Deflate, Zstd, #[default] Lz4, Brotli, Snappy, } impl CompressionAlgorithm { pub fn as_str(&self) -> &str { match self { CompressionAlgorithm::None => "none", CompressionAlgorithm::Gzip => "gzip", CompressionAlgorithm::Deflate => "deflate", CompressionAlgorithm::Zstd => "zstd", CompressionAlgorithm::Lz4 => "lz4", CompressionAlgorithm::Brotli => "brotli", CompressionAlgorithm::Snappy => "snappy", } } } impl fmt::Display for CompressionAlgorithm { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { write!(f, "{}", self.as_str()) } } impl str::FromStr for CompressionAlgorithm { type Err = io::Error; fn from_str(s: &str) -> Result<Self, Self::Err> { match s.to_lowercase().as_str() { "gzip" => Ok(CompressionAlgorithm::Gzip), "deflate" => Ok(CompressionAlgorithm::Deflate), "zstd" => Ok(CompressionAlgorithm::Zstd), "lz4" => Ok(CompressionAlgorithm::Lz4), "brotli" => Ok(CompressionAlgorithm::Brotli), "snappy" => Ok(CompressionAlgorithm::Snappy), "none" => Ok(CompressionAlgorithm::None), _ => Err(std::io::Error::other(format!("Unsupported compression algorithm: {s}"))), } } } /// Compress a block of data using the specified compression algorithm. /// Returns the compressed data as a Vec<u8>. /// /// # Arguments /// * `input` - The input data to be compressed. /// * `algorithm` - The compression algorithm to use. /// /// # Returns /// * A Vec<u8> containing the compressed data. /// pub fn compress_block(input: &[u8], algorithm: CompressionAlgorithm) -> Vec<u8> { match algorithm { CompressionAlgorithm::Gzip => { let mut encoder = flate2::write::GzEncoder::new(Vec::new(), flate2::Compression::default()); let _ = encoder.write_all(input); let _ = encoder.flush(); encoder.finish().unwrap_or_default() } CompressionAlgorithm::Deflate => { let mut encoder = flate2::write::DeflateEncoder::new(Vec::new(), flate2::Compression::default()); let _ = encoder.write_all(input); let _ = encoder.flush(); encoder.finish().unwrap_or_default() } CompressionAlgorithm::Zstd => { let mut encoder = zstd::Encoder::new(Vec::new(), 0).expect("zstd encoder"); let _ = encoder.write_all(input); encoder.finish().unwrap_or_default() } CompressionAlgorithm::Lz4 => { let mut encoder = lz4::EncoderBuilder::new().build(Vec::new()).expect("lz4 encoder"); let _ = encoder.write_all(input); let (out, result) = encoder.finish(); result.expect("lz4 finish"); out } CompressionAlgorithm::Brotli => { let mut out = Vec::new(); brotli::CompressorWriter::new(&mut out, 4096, 5, 22) .write_all(input) .expect("brotli compress"); out } CompressionAlgorithm::Snappy => { let mut encoder = snap::write::FrameEncoder::new(Vec::new()); let _ = encoder.write_all(input); encoder.into_inner().unwrap_or_default() } CompressionAlgorithm::None => input.to_vec(), } } /// Decompress a block of data using the specified compression algorithm. /// Returns the decompressed data as a Vec<u8>. /// /// # Arguments /// * `compressed` - The compressed data to be decompressed. /// * `algorithm` - The compression algorithm used for compression. /// /// # Returns /// * A Result containing a Vec<u8> with the decompressed data, or an io::Error. /// pub fn decompress_block(compressed: &[u8], algorithm: CompressionAlgorithm) -> io::Result<Vec<u8>> { match algorithm { CompressionAlgorithm::Gzip => { let mut decoder = flate2::read::GzDecoder::new(std::io::Cursor::new(compressed)); let mut out = Vec::new(); std::io::Read::read_to_end(&mut decoder, &mut out)?; Ok(out) } CompressionAlgorithm::Deflate => { let mut decoder = flate2::read::DeflateDecoder::new(std::io::Cursor::new(compressed)); let mut out = Vec::new(); std::io::Read::read_to_end(&mut decoder, &mut out)?; Ok(out) } CompressionAlgorithm::Zstd => { let mut decoder = zstd::Decoder::new(std::io::Cursor::new(compressed))?; let mut out = Vec::new(); std::io::Read::read_to_end(&mut decoder, &mut out)?; Ok(out) } CompressionAlgorithm::Lz4 => { let mut decoder = lz4::Decoder::new(std::io::Cursor::new(compressed)).expect("lz4 decoder"); let mut out = Vec::new(); std::io::Read::read_to_end(&mut decoder, &mut out)?; Ok(out) } CompressionAlgorithm::Brotli => { let mut out = Vec::new(); let mut decoder = brotli::Decompressor::new(std::io::Cursor::new(compressed), 4096); std::io::Read::read_to_end(&mut decoder, &mut out)?; Ok(out) } CompressionAlgorithm::Snappy => { let mut decoder = snap::read::FrameDecoder::new(std::io::Cursor::new(compressed)); let mut out = Vec::new(); std::io::Read::read_to_end(&mut decoder, &mut out)?; Ok(out) } CompressionAlgorithm::None => Ok(Vec::new()), } } #[cfg(test)] mod tests { use super::*; use std::str::FromStr; use std::time::Instant; #[test] fn test_compress_decompress_gzip() { let data = b"hello gzip compress"; let compressed = compress_block(data, CompressionAlgorithm::Gzip); let decompressed = decompress_block(&compressed, CompressionAlgorithm::Gzip).unwrap(); assert_eq!(decompressed, data); } #[test] fn test_compress_decompress_deflate() { let data = b"hello deflate compress"; let compressed = compress_block(data, CompressionAlgorithm::Deflate); let decompressed = decompress_block(&compressed, CompressionAlgorithm::Deflate).unwrap(); assert_eq!(decompressed, data); } #[test] fn test_compress_decompress_zstd() { let data = b"hello zstd compress"; let compressed = compress_block(data, CompressionAlgorithm::Zstd); let decompressed = decompress_block(&compressed, CompressionAlgorithm::Zstd).unwrap(); assert_eq!(decompressed, data); } #[test] fn test_compress_decompress_lz4() { let data = b"hello lz4 compress"; let compressed = compress_block(data, CompressionAlgorithm::Lz4); let decompressed = decompress_block(&compressed, CompressionAlgorithm::Lz4).unwrap(); assert_eq!(decompressed, data); } #[test] fn test_compress_decompress_brotli() { let data = b"hello brotli compress"; let compressed = compress_block(data, CompressionAlgorithm::Brotli); let decompressed = decompress_block(&compressed, CompressionAlgorithm::Brotli).unwrap(); assert_eq!(decompressed, data); } #[test] fn test_compress_decompress_snappy() { let data = b"hello snappy compress"; let compressed = compress_block(data, CompressionAlgorithm::Snappy); let decompressed = decompress_block(&compressed, CompressionAlgorithm::Snappy).unwrap(); assert_eq!(decompressed, data); } #[test] fn test_from_str() { assert_eq!(CompressionAlgorithm::from_str("gzip").unwrap(), CompressionAlgorithm::Gzip); assert_eq!(CompressionAlgorithm::from_str("deflate").unwrap(), CompressionAlgorithm::Deflate); assert_eq!(CompressionAlgorithm::from_str("zstd").unwrap(), CompressionAlgorithm::Zstd); assert_eq!(CompressionAlgorithm::from_str("lz4").unwrap(), CompressionAlgorithm::Lz4); assert_eq!(CompressionAlgorithm::from_str("brotli").unwrap(), CompressionAlgorithm::Brotli); assert_eq!(CompressionAlgorithm::from_str("snappy").unwrap(), CompressionAlgorithm::Snappy); assert!(CompressionAlgorithm::from_str("unknown").is_err()); } #[test] fn test_compare_compression_algorithms() { use std::time::Instant; let data = vec![42u8; 1024 * 100]; // 100KB of repetitive data // let mut data = vec![0u8; 1024 * 1024]; // rand::thread_rng().fill(&mut data[..]); let start = Instant::now(); let mut times = Vec::new(); times.push(("original", start.elapsed(), data.len())); let start = Instant::now(); let gzip = compress_block(&data, CompressionAlgorithm::Gzip); let gzip_time = start.elapsed(); times.push(("gzip", gzip_time, gzip.len())); let start = Instant::now(); let deflate = compress_block(&data, CompressionAlgorithm::Deflate); let deflate_time = start.elapsed(); times.push(("deflate", deflate_time, deflate.len())); let start = Instant::now(); let zstd = compress_block(&data, CompressionAlgorithm::Zstd); let zstd_time = start.elapsed(); times.push(("zstd", zstd_time, zstd.len())); let start = Instant::now(); let lz4 = compress_block(&data, CompressionAlgorithm::Lz4); let lz4_time = start.elapsed(); times.push(("lz4", lz4_time, lz4.len())); let start = Instant::now(); let brotli = compress_block(&data, CompressionAlgorithm::Brotli); let brotli_time = start.elapsed(); times.push(("brotli", brotli_time, brotli.len())); let start = Instant::now(); let snappy = compress_block(&data, CompressionAlgorithm::Snappy); let snappy_time = start.elapsed(); times.push(("snappy", snappy_time, snappy.len())); println!("Compression results:"); for (name, dur, size) in &times { println!("{name}: {size} bytes, {dur:?}"); } // All should decompress to the original assert_eq!(decompress_block(&gzip, CompressionAlgorithm::Gzip).unwrap(), data); assert_eq!(decompress_block(&deflate, CompressionAlgorithm::Deflate).unwrap(), data); assert_eq!(decompress_block(&zstd, CompressionAlgorithm::Zstd).unwrap(), data); assert_eq!(decompress_block(&lz4, CompressionAlgorithm::Lz4).unwrap(), data); assert_eq!(decompress_block(&brotli, CompressionAlgorithm::Brotli).unwrap(), data); assert_eq!(decompress_block(&snappy, CompressionAlgorithm::Snappy).unwrap(), data); // All compressed results should not be empty assert!( !gzip.is_empty() && !deflate.is_empty() && !zstd.is_empty() && !lz4.is_empty() && !brotli.is_empty() && !snappy.is_empty() ); } #[test] fn test_compression_benchmark() { let sizes = [128 * 1024, 512 * 1024, 1024 * 1024]; let algorithms = [ CompressionAlgorithm::Gzip, CompressionAlgorithm::Deflate, CompressionAlgorithm::Zstd, CompressionAlgorithm::Lz4, CompressionAlgorithm::Brotli, CompressionAlgorithm::Snappy, ]; println!("\nCompression algorithm benchmark results:"); println!( "{:<10} {:<10} {:<15} {:<15} {:<15}", "Data Size", "Algorithm", "Compress Time(ms)", "Compressed Size", "Compression Ratio" ); for size in sizes { // Generate compressible data (repeated text pattern) let pattern = b"Hello, this is a test pattern that will be repeated multiple times to create compressible data. "; let data: Vec<u8> = pattern.iter().cycle().take(size).copied().collect(); for algo in algorithms { // Compression test let start = Instant::now(); let compressed = compress_block(&data, algo); let compression_time = start.elapsed(); // Decompression test let start = Instant::now(); let _decompressed = decompress_block(&compressed, algo).unwrap(); let _decompression_time = start.elapsed(); // Calculate compression ratio let compression_ratio = (size as f64 / compressed.len() as f64) as f32; println!( "{:<10} {:<10} {:<15.2} {:<15} {:<15.2}x", format!("{}KB", size / 1024), algo.as_str(), compression_time.as_secs_f64() * 1000.0, compressed.len(), compression_ratio ); // Verify decompression result assert_eq!(_decompressed, data); } println!(); // Add blank line to separate results of different sizes } } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/lib.rs
crates/utils/src/lib.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #[cfg(feature = "tls")] pub mod certs; #[cfg(feature = "ip")] pub mod ip; #[cfg(feature = "net")] pub mod net; #[cfg(feature = "http")] pub mod http; #[cfg(feature = "net")] pub use net::*; #[cfg(all(feature = "net", feature = "io"))] pub mod retry; #[cfg(feature = "io")] pub mod io; #[cfg(feature = "hash")] pub mod hash; #[cfg(feature = "os")] pub mod os; #[cfg(feature = "path")] pub mod path; #[cfg(feature = "string")] pub mod string; #[cfg(feature = "crypto")] pub mod crypto; #[cfg(feature = "compress")] pub mod compress; #[cfg(feature = "path")] pub mod dirs; #[cfg(feature = "tls")] pub use certs::*; #[cfg(feature = "hash")] pub use hash::*; #[cfg(feature = "io")] pub use io::*; #[cfg(feature = "ip")] pub use ip::*; #[cfg(feature = "crypto")] pub use crypto::*; #[cfg(feature = "compress")] pub use compress::*; #[cfg(feature = "notify")] mod notify; #[cfg(feature = "sys")] pub mod sys; #[cfg(feature = "sys")] pub use sys::user_agent::*; #[cfg(feature = "notify")] pub use notify::*; #[cfg(feature = "obj")] pub mod obj; mod envs; pub use envs::*;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/string.rs
crates/utils/src/string.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use regex::Regex; use std::io::{Error, Result}; use std::sync::LazyLock; /// Parses a boolean value from a string. /// /// # Arguments /// `str` - A string slice representing the boolean value. /// /// # Returns /// A `Result` containing the parsed boolean value or an error if parsing fails. /// /// Examples /// ```no_run /// use rustfs_utils::string::parse_bool; /// /// let true_values = ["1", "t", "T", "true", "TRUE", "True", "on", "ON", "On", "enabled"]; /// let false_values = ["0", "f", "F", "false", "FALSE", "False", "off", "OFF", "Off", "disabled"]; /// /// for val in true_values.iter() { /// assert_eq!(parse_bool(val).unwrap(), true); /// } /// for val in false_values.iter() { /// assert_eq!(parse_bool(val).unwrap(), false); /// } /// ``` /// pub fn parse_bool(str: &str) -> Result<bool> { match str { "1" | "t" | "T" | "true" | "TRUE" | "True" | "on" | "ON" | "On" | "enabled" => Ok(true), "0" | "f" | "F" | "false" | "FALSE" | "False" | "off" | "OFF" | "Off" | "disabled" => Ok(false), _ => Err(Error::other(format!("ParseBool: parsing {str}"))), } } pub fn parse_bool_with_default(str: &str, default: bool) -> bool { match str { "1" | "t" | "T" | "true" | "TRUE" | "True" | "on" | "ON" | "On" | "enabled" => true, "0" | "f" | "F" | "false" | "FALSE" | "False" | "off" | "OFF" | "Off" | "disabled" => false, _ => default, } } /// Matches a simple pattern against a name using wildcards. /// /// # Arguments /// * `pattern` - The pattern to match, which may include wildcards '*' and '?' /// * `name` - The name to match against the pattern /// /// # Returns /// * `true` if the name matches the pattern, `false` otherwise /// /// Examples /// ```no_run /// use rustfs_utils::string::match_simple; /// assert!(match_simple("file*", "file123")); /// assert!(match_simple("file?", "file1")); /// assert!(!match_simple("file?", "file12")); /// ``` /// pub fn match_simple(pattern: &str, name: &str) -> bool { if pattern.is_empty() { return name == pattern; } if pattern == "*" { return true; } // Do an extended wildcard '*' and '?' match. deep_match_rune(name.as_bytes(), pattern.as_bytes(), true) } /// Matches a pattern against a name using wildcards. /// /// # Arguments /// * `pattern` - The pattern to match, which may include wildcards '*' and '?' /// * `name` - The name to match against the pattern /// /// # Returns /// * `true` if the name matches the pattern, `false` otherwise /// /// Examples /// ```no_run /// use rustfs_utils::string::match_pattern; /// /// assert!(match_pattern("file*", "file123")); /// assert!(match_pattern("file?", "file1")); /// assert!(!match_pattern("file?", "file12")); /// ``` /// pub fn match_pattern(pattern: &str, name: &str) -> bool { if pattern.is_empty() { return name == pattern; } if pattern == "*" { return true; } // Do an extended wildcard '*' and '?' match. deep_match_rune(name.as_bytes(), pattern.as_bytes(), false) } /// Checks if any pattern in the list matches the given string. /// /// # Arguments /// * `patterns` - A slice of patterns to match against /// * `match_str` - The string to match against the patterns /// /// # Returns /// * `true` if any pattern matches the string, `false` otherwise /// /// Examples /// ```no_run /// use rustfs_utils::string::has_pattern; /// /// let patterns = vec!["file*", "data?", "image*"]; /// assert!(has_pattern(&patterns, "file123")); /// assert!(has_pattern(&patterns, "data1")); /// assert!(!has_pattern(&patterns, "video1")); /// ``` /// pub fn has_pattern(patterns: &[&str], match_str: &str) -> bool { for pattern in patterns { if match_simple(pattern, match_str) { return true; } } false } /// Checks if the given string has any suffix from the provided list, ignoring case. /// /// # Arguments /// * `str` - The string to check /// * `list` - A slice of suffixes to check against /// /// # Returns /// * `true` if the string ends with any of the suffixes in the list, `false` otherwise /// /// Examples /// ```no_run /// use rustfs_utils::string::has_string_suffix_in_slice; /// /// let suffixes = vec![".txt", ".md", ".rs"]; /// assert!(has_string_suffix_in_slice("document.TXT", &suffixes)); /// assert!(!has_string_suffix_in_slice("image.png", &suffixes)); /// ``` pub fn has_string_suffix_in_slice(str: &str, list: &[&str]) -> bool { let str = str.to_lowercase(); for v in list { if *v == "*" { return true; } if str.ends_with(&v.to_lowercase()) { return true; } } false } fn deep_match_rune(str_: &[u8], pattern: &[u8], simple: bool) -> bool { let (mut str_, mut pattern) = (str_, pattern); while !pattern.is_empty() { match pattern[0] as char { '*' => { return if pattern.len() == 1 { true } else { deep_match_rune(str_, &pattern[1..], simple) || (!str_.is_empty() && deep_match_rune(&str_[1..], pattern, simple)) }; } '?' => { if str_.is_empty() { return simple; } } _ => { if str_.is_empty() || str_[0] != pattern[0] { return false; } } } str_ = &str_[1..]; pattern = &pattern[1..]; } str_.is_empty() && pattern.is_empty() } /// Matches a pattern as a prefix against the given text. /// /// # Arguments /// * `pattern` - The pattern to match, which may include wildcards '*' and '?' /// * `text` - The text to match against the pattern /// /// # Returns /// * `true` if the text matches the pattern as a prefix, `false` otherwise /// /// Examples /// ```no_run /// use rustfs_utils::string::match_as_pattern_prefix; /// /// assert!(match_as_pattern_prefix("file*", "file123")); /// assert!(match_as_pattern_prefix("file?", "file1")); /// assert!(!match_as_pattern_prefix("file?", "file12")); /// ``` /// pub fn match_as_pattern_prefix(pattern: &str, text: &str) -> bool { let mut i = 0; while i < text.len() && i < pattern.len() { match pattern.as_bytes()[i] as char { '*' => return true, '?' => i += 1, _ => { if pattern.as_bytes()[i] != text.as_bytes()[i] { return false; } } } i += 1; } text.len() <= pattern.len() } static ELLIPSES_RE: LazyLock<Regex> = LazyLock::new(|| Regex::new(r"(.*)(\{[0-9a-z]*\.\.\.[0-9a-z]*\})(.*)").unwrap()); /// Ellipses constants const OPEN_BRACES: &str = "{"; const CLOSE_BRACES: &str = "}"; const ELLIPSES: &str = "..."; /// ellipses pattern, describes the range and also the /// associated prefix and suffixes. #[derive(Debug, Default, PartialEq, Eq)] pub struct Pattern { pub prefix: String, pub suffix: String, pub seq: Vec<String>, } impl Pattern { /// expands a ellipses pattern. pub fn expand(&self) -> Vec<String> { let mut ret = Vec::with_capacity(self.suffix.len()); for v in self.seq.iter() { match (self.prefix.is_empty(), self.suffix.is_empty()) { (false, true) => ret.push(format!("{}{}", self.prefix, v)), (true, false) => ret.push(format!("{}{}", v, self.suffix)), (true, true) => ret.push(v.to_string()), (false, false) => ret.push(format!("{}{}{}", self.prefix, v, self.suffix)), } } ret } pub fn len(&self) -> usize { self.seq.len() } pub fn is_empty(&self) -> bool { self.seq.is_empty() } } /// contains a list of patterns provided in the input. #[derive(Debug, PartialEq, Eq)] pub struct ArgPattern { inner: Vec<Pattern>, } impl AsRef<Vec<Pattern>> for ArgPattern { fn as_ref(&self) -> &Vec<Pattern> { &self.inner } } impl AsMut<Vec<Pattern>> for ArgPattern { fn as_mut(&mut self) -> &mut Vec<Pattern> { &mut self.inner } } impl ArgPattern { pub fn new(inner: Vec<Pattern>) -> Self { Self { inner } } /// expands all the ellipses patterns in the given argument. pub fn expand(&self) -> Vec<Vec<String>> { let ret: Vec<Vec<String>> = self.inner.iter().map(|v| v.expand()).collect(); Self::arg_expander(&ret) } /// recursively expands labels into its respective forms. fn arg_expander(lbs: &[Vec<String>]) -> Vec<Vec<String>> { if lbs.len() == 1 { return lbs[0].iter().map(|v| vec![v.to_string()]).collect(); } let mut ret = Vec::new(); let (first, others) = lbs.split_at(1); for bs in first[0].iter() { let ots = Self::arg_expander(others); for mut obs in ots { obs.push(bs.to_string()); ret.push(obs); } } ret } /// returns the total number of sizes in the given patterns. pub fn total_sizes(&self) -> usize { self.inner.iter().fold(1, |acc, v| acc * v.seq.len()) } } /// finds all ellipses patterns, recursively and parses the ranges numerically. /// /// # Arguments /// * `arg` - The argument string to search for ellipses patterns /// /// # Returns /// * `Result<ArgPattern>` - A result containing the parsed ArgPattern or an error if parsing fails /// /// # Errors /// This function will return an error if the ellipses pattern format is invalid. /// /// Examples /// ```no_run /// use rustfs_utils::string::find_ellipses_patterns; /// /// let pattern = "http://rustfs{2...3}/export/set{1...64}"; /// let arg_pattern = find_ellipses_patterns(pattern).unwrap(); /// assert_eq!(arg_pattern.total_sizes(), 128); /// ``` pub fn find_ellipses_patterns(arg: &str) -> Result<ArgPattern> { let mut parts = match ELLIPSES_RE.captures(arg) { Some(caps) => caps, None => { return Err(Error::other(format!( "Invalid ellipsis format in ({arg}), Ellipsis range must be provided in format {{N...M}} where N and M are positive integers, M must be greater than N, with an allowed minimum range of 4" ))); } }; let mut patterns = Vec::new(); while let Some(prefix) = parts.get(1) { let seq = parse_ellipses_range(parts[2].into())?; match ELLIPSES_RE.captures(prefix.into()) { Some(cs) => { patterns.push(Pattern { seq, prefix: String::new(), suffix: parts[3].into(), }); parts = cs; } None => { patterns.push(Pattern { seq, prefix: prefix.as_str().to_owned(), suffix: parts[3].into(), }); break; } }; } // Check if any of the prefix or suffixes now have flower braces // left over, in such a case we generally think that there is // perhaps a typo in users input and error out accordingly. for p in patterns.iter() { if p.prefix.contains(OPEN_BRACES) || p.prefix.contains(CLOSE_BRACES) || p.suffix.contains(OPEN_BRACES) || p.suffix.contains(CLOSE_BRACES) { return Err(Error::other(format!( "Invalid ellipsis format in ({arg}), Ellipsis range must be provided in format {{N...M}} where N and M are positive integers, M must be greater than N, with an allowed minimum range of 4" ))); } } Ok(ArgPattern::new(patterns)) } /// returns true if input arg has ellipses type pattern. /// /// # Arguments /// * `s` - A slice of strings to check for ellipses patterns /// /// # Returns /// * `true` if any string contains ellipses patterns, `false` otherwise /// /// Examples /// ```no_run /// use rustfs_utils::string::has_ellipses; /// /// let args = vec!["http://rustfs{2...3}/export/set{1...64}", "mydisk-{a...z}{1...20}"]; /// assert!(has_ellipses(&args)); /// ``` /// pub fn has_ellipses<T: AsRef<str>>(s: &[T]) -> bool { let pattern = [ELLIPSES, OPEN_BRACES, CLOSE_BRACES]; s.iter().any(|v| pattern.iter().any(|p| v.as_ref().contains(p))) } /// Parses an ellipses range pattern of following style /// /// example: /// {1...64} /// {33...64} /// /// # Arguments /// * `pattern` - A string slice representing the ellipses range pattern /// /// # Returns /// * `Result<Vec<String>>` - A result containing a vector of strings representing the expanded range or an error if parsing fails /// /// # Errors /// This function will return an error if the ellipses range format is invalid. /// /// Examples /// ```no_run /// use rustfs_utils::string::parse_ellipses_range; /// /// let range = parse_ellipses_range("{1...5}").unwrap(); /// assert_eq!(range, vec!["1", "2", "3", "4", "5"]); /// ``` /// pub fn parse_ellipses_range(pattern: &str) -> Result<Vec<String>> { if !pattern.contains(OPEN_BRACES) { return Err(Error::other("Invalid argument")); } if !pattern.contains(CLOSE_BRACES) { return Err(Error::other("Invalid argument")); } let ellipses_range: Vec<&str> = pattern .trim_start_matches(OPEN_BRACES) .trim_end_matches(CLOSE_BRACES) .split(ELLIPSES) .collect(); if ellipses_range.len() != 2 { return Err(Error::other("Invalid argument")); } // TODO: Add support for hexadecimals. let start = ellipses_range[0].parse::<usize>().map_err(Error::other)?; let end = ellipses_range[1].parse::<usize>().map_err(Error::other)?; if start > end { return Err(Error::other("Invalid argument:range start cannot be bigger than end")); } let mut ret: Vec<String> = Vec::with_capacity(end - start + 1); for i in start..=end { if ellipses_range[0].starts_with('0') && ellipses_range[0].len() > 1 { ret.push(format!("{:0width$}", i, width = ellipses_range[1].len())); } else { ret.push(format!("{i}")); } } Ok(ret) } /// Tests whether the string s begins with prefix ignoring case /// /// # Arguments /// * `s` - The string to test /// * `prefix` - The prefix to look for /// /// # Returns /// * `true` if s starts with prefix (case-insensitive), `false` otherwise /// /// # Examples /// ```no_run /// use rustfs_utils::string::strings_has_prefix_fold; /// /// assert!(strings_has_prefix_fold("HelloWorld", "hello")); /// assert!(!strings_has_prefix_fold("HelloWorld", "world")); /// ``` /// pub fn strings_has_prefix_fold(s: &str, prefix: &str) -> bool { if s.len() < prefix.len() { return false; } let s_prefix = &s[..prefix.len()]; // Test match with case first, then case-insensitive s_prefix == prefix || s_prefix.to_lowercase() == prefix.to_lowercase() } #[cfg(test)] mod tests { use super::*; #[test] fn test_has_ellipses() { // Tests for all args without ellipses. let test_cases = [ (1, vec!["64"], false), // Found flower braces, still attempt to parse and throw an error. (2, vec!["{1..64}"], true), (3, vec!["{1..2..}"], true), // Test for valid input. (4, vec!["1...64"], true), (5, vec!["{1...2O}"], true), (6, vec!["..."], true), (7, vec!["{-1...1}"], true), (8, vec!["{0...-1}"], true), (9, vec!["{1....4}"], true), (10, vec!["{1...64}"], true), (11, vec!["{...}"], true), (12, vec!["{1...64}", "{65...128}"], true), (13, vec!["http://rustfs{2...3}/export/set{1...64}"], true), ( 14, vec![ "http://rustfs{2...3}/export/set{1...64}", "http://rustfs{2...3}/export/set{65...128}", ], true, ), (15, vec!["mydisk-{a...z}{1...20}"], true), (16, vec!["mydisk-{1...4}{1..2.}"], true), ]; for (i, args, expected) in test_cases { let ret = has_ellipses(&args); assert_eq!(ret, expected, "Test{i}: Expected {expected}, got {ret}"); } } #[test] fn test_find_ellipses_patterns() { #[derive(Default)] struct TestCase<'a> { num: usize, pattern: &'a str, success: bool, want: Vec<Vec<&'a str>>, } let test_cases = [ TestCase { num: 1, pattern: "{1..64}", ..Default::default() }, TestCase { num: 2, pattern: "1...64", ..Default::default() }, TestCase { num: 2, pattern: "...", ..Default::default() }, TestCase { num: 3, pattern: "{1...", ..Default::default() }, TestCase { num: 4, pattern: "...64}", ..Default::default() }, TestCase { num: 5, pattern: "{...}", ..Default::default() }, TestCase { num: 6, pattern: "{-1...1}", ..Default::default() }, TestCase { num: 7, pattern: "{0...-1}", ..Default::default() }, TestCase { num: 8, pattern: "{1...2O}", ..Default::default() }, TestCase { num: 9, pattern: "{64...1}", ..Default::default() }, TestCase { num: 10, pattern: "{1....4}", ..Default::default() }, TestCase { num: 11, pattern: "mydisk-{a...z}{1...20}", ..Default::default() }, TestCase { num: 12, pattern: "mydisk-{1...4}{1..2.}", ..Default::default() }, TestCase { num: 13, pattern: "{1..2.}-mydisk-{1...4}", ..Default::default() }, TestCase { num: 14, pattern: "{{1...4}}", ..Default::default() }, TestCase { num: 16, pattern: "{4...02}", ..Default::default() }, TestCase { num: 17, pattern: "{f...z}", ..Default::default() }, // Test for valid input. TestCase { num: 18, pattern: "{1...64}", success: true, want: vec![ vec!["1"], vec!["2"], vec!["3"], vec!["4"], vec!["5"], vec!["6"], vec!["7"], vec!["8"], vec!["9"], vec!["10"], vec!["11"], vec!["12"], vec!["13"], vec!["14"], vec!["15"], vec!["16"], vec!["17"], vec!["18"], vec!["19"], vec!["20"], vec!["21"], vec!["22"], vec!["23"], vec!["24"], vec!["25"], vec!["26"], vec!["27"], vec!["28"], vec!["29"], vec!["30"], vec!["31"], vec!["32"], vec!["33"], vec!["34"], vec!["35"], vec!["36"], vec!["37"], vec!["38"], vec!["39"], vec!["40"], vec!["41"], vec!["42"], vec!["43"], vec!["44"], vec!["45"], vec!["46"], vec!["47"], vec!["48"], vec!["49"], vec!["50"], vec!["51"], vec!["52"], vec!["53"], vec!["54"], vec!["55"], vec!["56"], vec!["57"], vec!["58"], vec!["59"], vec!["60"], vec!["61"], vec!["62"], vec!["63"], vec!["64"], ], }, TestCase { num: 19, pattern: "{1...5} {65...70}", success: true, want: vec![ vec!["1 ", "65"], vec!["2 ", "65"], vec!["3 ", "65"], vec!["4 ", "65"], vec!["5 ", "65"], vec!["1 ", "66"], vec!["2 ", "66"], vec!["3 ", "66"], vec!["4 ", "66"], vec!["5 ", "66"], vec!["1 ", "67"], vec!["2 ", "67"], vec!["3 ", "67"], vec!["4 ", "67"], vec!["5 ", "67"], vec!["1 ", "68"], vec!["2 ", "68"], vec!["3 ", "68"], vec!["4 ", "68"], vec!["5 ", "68"], vec!["1 ", "69"], vec!["2 ", "69"], vec!["3 ", "69"], vec!["4 ", "69"], vec!["5 ", "69"], vec!["1 ", "70"], vec!["2 ", "70"], vec!["3 ", "70"], vec!["4 ", "70"], vec!["5 ", "70"], ], }, TestCase { num: 20, pattern: "{01...036}", success: true, want: vec![ vec!["001"], vec!["002"], vec!["003"], vec!["004"], vec!["005"], vec!["006"], vec!["007"], vec!["008"], vec!["009"], vec!["010"], vec!["011"], vec!["012"], vec!["013"], vec!["014"], vec!["015"], vec!["016"], vec!["017"], vec!["018"], vec!["019"], vec!["020"], vec!["021"], vec!["022"], vec!["023"], vec!["024"], vec!["025"], vec!["026"], vec!["027"], vec!["028"], vec!["029"], vec!["030"], vec!["031"], vec!["032"], vec!["033"], vec!["034"], vec!["035"], vec!["036"], ], }, TestCase { num: 21, pattern: "{001...036}", success: true, want: vec![ vec!["001"], vec!["002"], vec!["003"], vec!["004"], vec!["005"], vec!["006"], vec!["007"], vec!["008"], vec!["009"], vec!["010"], vec!["011"], vec!["012"], vec!["013"], vec!["014"], vec!["015"], vec!["016"], vec!["017"], vec!["018"], vec!["019"], vec!["020"], vec!["021"], vec!["022"], vec!["023"], vec!["024"], vec!["025"], vec!["026"], vec!["027"], vec!["028"], vec!["029"], vec!["030"], vec!["031"], vec!["032"], vec!["033"], vec!["034"], vec!["035"], vec!["036"], ], }, ]; for test_case in test_cases { let ret = find_ellipses_patterns(test_case.pattern); match ret { Ok(v) => { if !test_case.success { panic!("Test{}: Expected failure but passed instead", test_case.num); } let got = v.expand(); if got.len() != test_case.want.len() { panic!("Test{}: Expected {}, got {}", test_case.num, test_case.want.len(), got.len()); } assert_eq!(got, test_case.want, "Test{}: Expected {:?}, got {:?}", test_case.num, test_case.want, got); } Err(e) => { if test_case.success { panic!("Test{}: Expected success but failed instead {:?}", test_case.num, e); } } } } } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/io.rs
crates/utils/src/io.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use tokio::io::{AsyncRead, AsyncReadExt, AsyncWrite, AsyncWriteExt}; /// Write all bytes from buf to writer, returning the total number of bytes written. pub async fn write_all<W: AsyncWrite + Send + Sync + Unpin>(writer: &mut W, buf: &[u8]) -> std::io::Result<usize> { let mut total = 0; while total < buf.len() { match writer.write(&buf[total..]).await { Ok(0) => { break; } Ok(n) => total += n, Err(e) => return Err(e), } } Ok(total) } /// Read exactly buf.len() bytes into buf, or return an error if EOF is reached before. /// Like Go's io.ReadFull. #[allow(dead_code)] pub async fn read_full<R: AsyncRead + Send + Sync + Unpin>(mut reader: R, mut buf: &mut [u8]) -> std::io::Result<usize> { let mut total = 0; while !buf.is_empty() { let n = match reader.read(buf).await { Ok(n) => n, Err(e) => { if total == 0 { return Err(e); } // If the error is InvalidData (e.g., checksum mismatch), preserve it // instead of wrapping it as UnexpectedEof, so proper error handling can occur if e.kind() == std::io::ErrorKind::InvalidData { return Err(e); } return Err(std::io::Error::new( std::io::ErrorKind::UnexpectedEof, format!("read {total} bytes, error: {e}"), )); } }; if n == 0 { if total > 0 { return Ok(total); } return Err(std::io::Error::new(std::io::ErrorKind::UnexpectedEof, "early EOF")); } buf = &mut buf[n..]; total += n; } Ok(total) } /// Encodes a u64 into buf and returns the number of bytes written. /// Panics if buf is too small. pub fn put_uvarint(buf: &mut [u8], x: u64) -> usize { let mut i = 0; let mut x = x; while x >= 0x80 { buf[i] = (x as u8) | 0x80; x >>= 7; i += 1; } buf[i] = x as u8; i + 1 } pub fn put_uvarint_len(x: u64) -> usize { let mut i = 0; let mut x = x; while x >= 0x80 { x >>= 7; i += 1; } i + 1 } /// Decodes an u64 from buf and returns (value, number of bytes read). /// If buf is too small, returns (0, 0). /// If overflow, returns (0, -(n as isize)), where n is the number of bytes read. pub fn uvarint(buf: &[u8]) -> (u64, isize) { let mut x: u64 = 0; let mut s: u32 = 0; for (i, &b) in buf.iter().enumerate() { if i == 10 { // MaxVarintLen64 = 10 return (0, -((i + 1) as isize)); } if b < 0x80 { if i == 9 && b > 1 { return (0, -((i + 1) as isize)); } return (x | ((b as u64) << s), (i + 1) as isize); } x |= ((b & 0x7F) as u64) << s; s += 7; } (0, 0) } #[cfg(test)] mod tests { use super::*; use tokio::io::BufReader; use tracing::debug; #[tokio::test] async fn test_read_full_exact() { // let data = b"abcdef"; let data = b"channel async callback test data!"; let mut reader = BufReader::new(&data[..]); let size = data.len(); let mut total = 0; let mut rev = vec![0u8; size]; let mut count = 0; while total < size { let mut buf = [0u8; 8]; let n = read_full(&mut reader, &mut buf).await.unwrap(); total += n; rev[total - n..total].copy_from_slice(&buf[..n]); count += 1; debug!("Read progress - count: {}, total: {}, bytes read: {}", count, total, n); } assert_eq!(total, size); assert_eq!(&rev, data); } #[tokio::test] async fn test_read_full_short() { let data = b"abc"; let mut reader = BufReader::new(&data[..]); let mut buf = [0u8; 6]; let n = read_full(&mut reader, &mut buf).await.unwrap(); assert_eq!(n, 3); assert_eq!(&buf[..n], data); } #[tokio::test] async fn test_read_full_1m() { let size = 1024 * 1024; let data = vec![42u8; size]; let mut reader = BufReader::new(&data[..]); let mut buf = vec![0u8; size / 3]; read_full(&mut reader, &mut buf).await.unwrap(); assert_eq!(buf, data[..size / 3]); } #[test] fn test_put_uvarint_and_uvarint_zero() { let mut buf = [0u8; 16]; let n = put_uvarint(&mut buf, 0); let (decoded, m) = uvarint(&buf[..n]); assert_eq!(decoded, 0); assert_eq!(m as usize, n); } #[test] fn test_put_uvarint_and_uvarint_max() { let mut buf = [0u8; 16]; let n = put_uvarint(&mut buf, u64::MAX); let (decoded, m) = uvarint(&buf[..n]); assert_eq!(decoded, u64::MAX); assert_eq!(m as usize, n); } #[test] fn test_put_uvarint_and_uvarint_various() { let mut buf = [0u8; 16]; for &v in &[1u64, 127, 128, 255, 300, 16384, u32::MAX as u64] { let n = put_uvarint(&mut buf, v); let (decoded, m) = uvarint(&buf[..n]); assert_eq!(decoded, v, "decode mismatch for {v}"); assert_eq!(m as usize, n, "length mismatch for {v}"); } } #[test] fn test_uvarint_incomplete() { let buf = [0x80u8, 0x80, 0x80]; let (v, n) = uvarint(&buf); assert_eq!(v, 0); assert_eq!(n, 0); } #[test] fn test_uvarint_overflow_case() { let buf = [0xFFu8; 11]; let (v, n) = uvarint(&buf); assert_eq!(v, 0); assert!(n < 0); } #[tokio::test] async fn test_write_all_basic() { let data = b"hello world!"; let mut buf = Vec::new(); let n = write_all(&mut buf, data).await.unwrap(); assert_eq!(n, data.len()); assert_eq!(&buf, data); } #[tokio::test] async fn test_write_all_partial() { struct PartialWriter { inner: Vec<u8>, max_write: usize, } use std::pin::Pin; use std::task::{Context, Poll}; use tokio::io::AsyncWrite; impl AsyncWrite for PartialWriter { fn poll_write(mut self: Pin<&mut Self>, _cx: &mut Context<'_>, buf: &[u8]) -> Poll<std::io::Result<usize>> { let n = buf.len().min(self.max_write); self.inner.extend_from_slice(&buf[..n]); Poll::Ready(Ok(n)) } fn poll_flush(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<std::io::Result<()>> { Poll::Ready(Ok(())) } fn poll_shutdown(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<std::io::Result<()>> { Poll::Ready(Ok(())) } } let data = b"abcdefghijklmnopqrstuvwxyz"; let mut writer = PartialWriter { inner: Vec::new(), max_write: 5, }; let n = write_all(&mut writer, data).await.unwrap(); assert_eq!(n, data.len()); assert_eq!(&writer.inner, data); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/certs.rs
crates/utils/src/certs.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use crate::get_env_bool; use rustfs_config::{RUSTFS_TLS_CERT, RUSTFS_TLS_KEY}; use rustls::RootCertStore; use rustls::server::danger::ClientCertVerifier; use rustls::server::{ClientHello, ResolvesServerCert, ResolvesServerCertUsingSni, WebPkiClientVerifier}; use rustls::sign::CertifiedKey; use rustls_pemfile::{certs, private_key}; use rustls_pki_types::{CertificateDer, PrivateKeyDer}; use std::collections::HashMap; use std::io::Error; use std::path::Path; use std::sync::Arc; use std::{fs, io}; use tracing::{debug, warn}; /// Load public certificate from file. /// This function loads a public certificate from the specified file. /// /// # Arguments /// * `filename` - A string slice that holds the name of the file containing the public certificate. /// /// # Returns /// * An io::Result containing a vector of CertificateDer if successful, or an io::Error if an error occurs during loading. /// pub fn load_certs(filename: &str) -> io::Result<Vec<CertificateDer<'static>>> { // Open certificate file. let cert_file = fs::File::open(filename).map_err(|e| certs_error(format!("failed to open {filename}: {e}")))?; let mut reader = io::BufReader::new(cert_file); // Load and return certificate. let certs = certs(&mut reader) .collect::<Result<Vec<_>, _>>() .map_err(|e| certs_error(format!("certificate file {filename} format error:{e:?}")))?; if certs.is_empty() { return Err(certs_error(format!("No valid certificate was found in the certificate file {filename}"))); } Ok(certs) } /// Load a PEM certificate bundle and return each certificate as DER bytes. /// /// This is a low-level helper intended for TLS clients (reqwest/hyper-rustls) that /// need to add root certificates one-by-one. /// /// - Input: a PEM file that may contain multiple cert blocks. /// - Output: Vec of DER-encoded cert bytes, one per cert. /// /// NOTE: This intentionally returns raw bytes to avoid forcing downstream crates /// to depend on rustls types. pub fn load_cert_bundle_der_bytes(path: &str) -> io::Result<Vec<Vec<u8>>> { let pem = fs::read(path)?; let mut reader = io::BufReader::new(&pem[..]); let certs = certs(&mut reader) .collect::<Result<Vec<_>, _>>() .map_err(|e| certs_error(format!("Failed to parse PEM certs from {path}: {e}")))?; Ok(certs.into_iter().map(|c| c.to_vec()).collect()) } /// Builds a WebPkiClientVerifier for mTLS if enabled via environment variable. /// /// # Arguments /// * `tls_path` - Directory containing client CA certificates /// /// # Returns /// * `Ok(Some(verifier))` if mTLS is enabled and CA certs are found /// * `Ok(None)` if mTLS is disabled /// * `Err` if mTLS is enabled but configuration is invalid pub fn build_webpki_client_verifier(tls_path: &str) -> io::Result<Option<Arc<dyn ClientCertVerifier>>> { if !get_env_bool(rustfs_config::ENV_SERVER_MTLS_ENABLE, rustfs_config::DEFAULT_SERVER_MTLS_ENABLE) { return Ok(None); } let ca_path = mtls_ca_bundle_path(tls_path).ok_or_else(|| { Error::other(format!( "RUSTFS_SERVER_MTLS_ENABLE=true but missing {}/client_ca.crt (or fallback {}/ca.crt)", tls_path, tls_path )) })?; let der_list = load_cert_bundle_der_bytes(ca_path.to_str().unwrap_or_default())?; let mut store = RootCertStore::empty(); for der in der_list { store .add(der.into()) .map_err(|e| Error::other(format!("Invalid client CA cert: {e}")))?; } let verifier = WebPkiClientVerifier::builder(Arc::new(store)) .build() .map_err(|e| Error::other(format!("Build client cert verifier failed: {e}")))?; Ok(Some(verifier)) } /// Locate the mTLS client CA bundle in the specified TLS path fn mtls_ca_bundle_path(tls_path: &str) -> Option<std::path::PathBuf> { use std::path::Path; let p1 = Path::new(tls_path).join(rustfs_config::RUSTFS_CLIENT_CA_CERT_FILENAME); if p1.exists() { return Some(p1); } let p2 = Path::new(tls_path).join(rustfs_config::RUSTFS_CA_CERT); if p2.exists() { return Some(p2); } None } /// Load private key from file. /// This function loads a private key from the specified file. /// /// # Arguments /// * `filename` - A string slice that holds the name of the file containing the private key. /// /// # Returns /// * An io::Result containing the PrivateKeyDer if successful, or an io::Error if an error occurs during loading. /// pub fn load_private_key(filename: &str) -> io::Result<PrivateKeyDer<'static>> { // Open keyfile. let keyfile = fs::File::open(filename).map_err(|e| certs_error(format!("failed to open {filename}: {e}")))?; let mut reader = io::BufReader::new(keyfile); // Load and return a single private key. private_key(&mut reader)?.ok_or_else(|| certs_error(format!("no private key found in {filename}"))) } /// error function /// This function creates a new io::Error with the provided error message. /// /// # Arguments /// * `err` - A string containing the error message. /// /// # Returns /// * An io::Error instance with the specified error message. /// pub fn certs_error(err: String) -> Error { Error::other(err) } /// Load all certificates and private keys in the directory /// This function loads all certificate and private key pairs from the specified directory. /// It looks for files named `rustfs_cert.pem` and `rustfs_key.pem` in each subdirectory. /// The root directory can also contain a default certificate/private key pair. /// /// # Arguments /// * `dir_path` - A string slice that holds the path to the directory containing the certificates and private keys. /// /// # Returns /// * An io::Result containing a HashMap where the keys are domain names (or "default" for the root certificate) and the values are tuples of (Vec<CertificateDer>, PrivateKeyDer). If no valid certificate/private key pairs are found, an io::Error is returned. /// pub fn load_all_certs_from_directory( dir_path: &str, ) -> io::Result<HashMap<String, (Vec<CertificateDer<'static>>, PrivateKeyDer<'static>)>> { let mut cert_key_pairs = HashMap::new(); let dir = Path::new(dir_path); if !dir.exists() || !dir.is_dir() { return Err(certs_error(format!( "The certificate directory does not exist or is not a directory: {dir_path}" ))); } // 1. First check whether there is a certificate/private key pair in the root directory let root_cert_path = dir.join(RUSTFS_TLS_CERT); let root_key_path = dir.join(RUSTFS_TLS_KEY); if root_cert_path.exists() && root_key_path.exists() { debug!("find the root directory certificate: {:?}", root_cert_path); let root_cert_str = root_cert_path .to_str() .ok_or_else(|| certs_error(format!("Invalid UTF-8 in root certificate path: {root_cert_path:?}")))?; let root_key_str = root_key_path .to_str() .ok_or_else(|| certs_error(format!("Invalid UTF-8 in root key path: {root_key_path:?}")))?; match load_cert_key_pair(root_cert_str, root_key_str) { Ok((certs, key)) => { // The root directory certificate is used as the default certificate and is stored using special keys. cert_key_pairs.insert("default".to_string(), (certs, key)); } Err(e) => { warn!("unable to load root directory certificate: {}", e); } } } // 2.iterate through all folders in the directory for entry in fs::read_dir(dir)? { let entry = entry?; let path = entry.path(); if path.is_dir() { let domain_name: &str = path .file_name() .and_then(|name| name.to_str()) .ok_or_else(|| certs_error(format!("invalid domain name directory:{path:?}")))?; // find certificate and private key files let cert_path = path.join(RUSTFS_TLS_CERT); // e.g., rustfs_cert.pem let key_path = path.join(RUSTFS_TLS_KEY); // e.g., rustfs_key.pem if cert_path.exists() && key_path.exists() { debug!("find the domain name certificate: {} in {:?}", domain_name, cert_path); match load_cert_key_pair(cert_path.to_str().unwrap(), key_path.to_str().unwrap()) { Ok((certs, key)) => { cert_key_pairs.insert(domain_name.to_string(), (certs, key)); } Err(e) => { warn!("unable to load the certificate for {} domain name: {}", domain_name, e); } } } } } if cert_key_pairs.is_empty() { return Err(certs_error(format!( "No valid certificate/private key pair found in directory {dir_path}" ))); } Ok(cert_key_pairs) } /// loading a single certificate private key pair /// This function loads a certificate and private key from the specified paths. /// It returns a tuple containing the certificate and private key. /// /// # Arguments /// * `cert_path` - A string slice that holds the path to the certificate file. /// * `key_path` - A string slice that holds the path to the private key file /// /// # Returns /// * An io::Result containing a tuple of (Vec<CertificateDer>, PrivateKeyDer) if successful, or an io::Error if an error occurs during loading. /// fn load_cert_key_pair(cert_path: &str, key_path: &str) -> io::Result<(Vec<CertificateDer<'static>>, PrivateKeyDer<'static>)> { let certs = load_certs(cert_path)?; let key = load_private_key(key_path)?; Ok((certs, key)) } /// Create a multi-cert resolver /// This function loads all certificates and private keys from the specified directory. /// It uses the first certificate/private key pair found in the root directory as the default certificate. /// The rest of the certificates/private keys are used for SNI resolution. /// /// # Arguments /// * `cert_key_pairs` - A HashMap where the keys are domain names (or "default" for the root certificate) and the values are tuples of (Vec<CertificateDer>, PrivateKeyDer). /// /// # Returns /// * An io::Result containing an implementation of ResolvesServerCert if successful, or an io::Error if an error occurs during loading. /// pub fn create_multi_cert_resolver( cert_key_pairs: HashMap<String, (Vec<CertificateDer<'static>>, PrivateKeyDer<'static>)>, ) -> io::Result<impl ResolvesServerCert> { #[derive(Debug)] struct MultiCertResolver { cert_resolver: ResolvesServerCertUsingSni, default_cert: Option<Arc<CertifiedKey>>, } impl ResolvesServerCert for MultiCertResolver { fn resolve(&self, client_hello: ClientHello) -> Option<Arc<CertifiedKey>> { // try matching certificates with sni if let Some(cert) = self.cert_resolver.resolve(client_hello) { return Some(cert); } // If there is no matching SNI certificate, use the default certificate self.default_cert.clone() } } let mut resolver = ResolvesServerCertUsingSni::new(); let mut default_cert = None; for (domain, (certs, key)) in cert_key_pairs { // create a signature let signing_key = rustls::crypto::aws_lc_rs::sign::any_supported_type(&key) .map_err(|e| certs_error(format!("unsupported private key types:{domain}, err:{e:?}")))?; // create a CertifiedKey let certified_key = CertifiedKey::new(certs, signing_key); if domain == "default" { default_cert = Some(Arc::new(certified_key.clone())); } else { // add certificate to resolver resolver .add(&domain, certified_key) .map_err(|e| certs_error(format!("failed to add a domain name certificate:{domain},err: {e:?}")))?; } } Ok(MultiCertResolver { cert_resolver: resolver, default_cert, }) } /// Checks if TLS key logging is enabled. /// /// # Returns /// * A boolean indicating whether TLS key logging is enabled based on the `RUSTFS_TLS_KEYLOG` environment variable. /// pub fn tls_key_log() -> bool { crate::get_env_bool(rustfs_config::ENV_TLS_KEYLOG, rustfs_config::DEFAULT_TLS_KEYLOG) } #[cfg(test)] mod tests { use super::*; use std::fs; use tempfile::TempDir; #[test] fn test_certs_error_function() { let error_msg = "Test error message"; let error = certs_error(error_msg.to_string()); assert_eq!(error.kind(), std::io::ErrorKind::Other); assert_eq!(error.to_string(), error_msg); } #[test] fn test_load_certs_file_not_found() { let result = load_certs("non_existent_file.pem"); assert!(result.is_err()); let error = result.unwrap_err(); assert_eq!(error.kind(), std::io::ErrorKind::Other); assert!(error.to_string().contains("failed to open")); } #[test] fn test_load_private_key_file_not_found() { let result = load_private_key("non_existent_key.pem"); assert!(result.is_err()); let error = result.unwrap_err(); assert_eq!(error.kind(), std::io::ErrorKind::Other); assert!(error.to_string().contains("failed to open")); } #[test] fn test_load_certs_empty_file() { let temp_dir = TempDir::new().unwrap(); let cert_path = temp_dir.path().join("empty.pem"); fs::write(&cert_path, "").unwrap(); let result = load_certs(cert_path.to_str().unwrap()); assert!(result.is_err()); let error = result.unwrap_err(); assert!(error.to_string().contains("No valid certificate was found")); } #[test] fn test_load_certs_invalid_format() { let temp_dir = TempDir::new().unwrap(); let cert_path = temp_dir.path().join("invalid.pem"); fs::write(&cert_path, "invalid certificate content").unwrap(); let result = load_certs(cert_path.to_str().unwrap()); assert!(result.is_err()); let error = result.unwrap_err(); assert!(error.to_string().contains("No valid certificate was found")); } #[test] fn test_load_private_key_empty_file() { let temp_dir = TempDir::new().unwrap(); let key_path = temp_dir.path().join("empty_key.pem"); fs::write(&key_path, "").unwrap(); let result = load_private_key(key_path.to_str().unwrap()); assert!(result.is_err()); let error = result.unwrap_err(); assert!(error.to_string().contains("no private key found")); } #[test] fn test_load_private_key_invalid_format() { let temp_dir = TempDir::new().unwrap(); let key_path = temp_dir.path().join("invalid_key.pem"); fs::write(&key_path, "invalid private key content").unwrap(); let result = load_private_key(key_path.to_str().unwrap()); assert!(result.is_err()); let error = result.unwrap_err(); assert!(error.to_string().contains("no private key found")); } #[test] fn test_load_all_certs_from_directory_not_exists() { let result = load_all_certs_from_directory("/non/existent/directory"); assert!(result.is_err()); let error = result.unwrap_err(); assert!(error.to_string().contains("does not exist or is not a directory")); } #[test] fn test_load_all_certs_from_directory_empty() { let temp_dir = TempDir::new().unwrap(); let result = load_all_certs_from_directory(temp_dir.path().to_str().unwrap()); assert!(result.is_err()); let error = result.unwrap_err(); assert!(error.to_string().contains("No valid certificate/private key pair found")); } #[test] fn test_load_all_certs_from_directory_file_instead_of_dir() { let temp_dir = TempDir::new().unwrap(); let file_path = temp_dir.path().join("not_a_directory.txt"); fs::write(&file_path, "content").unwrap(); let result = load_all_certs_from_directory(file_path.to_str().unwrap()); assert!(result.is_err()); let error = result.unwrap_err(); assert!(error.to_string().contains("does not exist or is not a directory")); } #[test] fn test_load_cert_key_pair_missing_cert() { let temp_dir = TempDir::new().unwrap(); let key_path = temp_dir.path().join("test_key.pem"); fs::write(&key_path, "dummy key content").unwrap(); let result = load_cert_key_pair("non_existent_cert.pem", key_path.to_str().unwrap()); assert!(result.is_err()); } #[test] fn test_load_cert_key_pair_missing_key() { let temp_dir = TempDir::new().unwrap(); let cert_path = temp_dir.path().join("test_cert.pem"); fs::write(&cert_path, "dummy cert content").unwrap(); let result = load_cert_key_pair(cert_path.to_str().unwrap(), "non_existent_key.pem"); assert!(result.is_err()); } #[test] fn test_create_multi_cert_resolver_empty_map() { let empty_map = HashMap::new(); let result = create_multi_cert_resolver(empty_map); // Should succeed even with empty map assert!(result.is_ok()); } #[test] fn test_error_message_formatting() { let test_cases = vec![ ("file not found", "failed to open test.pem: file not found"), ("permission denied", "failed to open key.pem: permission denied"), ("invalid format", "certificate file cert.pem format error:invalid format"), ]; for (input, _expected_pattern) in test_cases { let error1 = certs_error(format!("failed to open test.pem: {input}")); assert!(error1.to_string().contains(input)); let error2 = certs_error(format!("failed to open key.pem: {input}")); assert!(error2.to_string().contains(input)); } } #[test] fn test_path_handling_edge_cases() { // Test with various path formats let path_cases = vec![ "", // Empty path ".", // Current directory "..", // Parent directory "/", // Root directory (Unix) "relative/path", // Relative path "/absolute/path", // Absolute path ]; for path in path_cases { let result = load_all_certs_from_directory(path); // All should fail since these are not valid cert directories assert!(result.is_err()); } } #[test] fn test_filename_constants_consistency() { // Test that the constants match expected values assert_eq!(RUSTFS_TLS_CERT, "rustfs_cert.pem"); assert_eq!(RUSTFS_TLS_KEY, "rustfs_key.pem"); // Test that constants are not empty assert!(!RUSTFS_TLS_CERT.is_empty()); assert!(!RUSTFS_TLS_KEY.is_empty()); // Test that constants have proper extensions assert!(RUSTFS_TLS_CERT.ends_with(".pem")); assert!(RUSTFS_TLS_KEY.ends_with(".pem")); } #[test] fn test_directory_structure_validation() { let temp_dir = TempDir::new().unwrap(); // Create a subdirectory without certificates let sub_dir = temp_dir.path().join("example.com"); fs::create_dir(&sub_dir).unwrap(); // Should fail because no certificates found let result = load_all_certs_from_directory(temp_dir.path().to_str().unwrap()); assert!(result.is_err()); assert!( result .unwrap_err() .to_string() .contains("No valid certificate/private key pair found") ); } #[test] fn test_unicode_path_handling() { let temp_dir = TempDir::new().unwrap(); // Create directory with Unicode characters let unicode_dir = temp_dir.path().join("test_directory"); fs::create_dir(&unicode_dir).unwrap(); let result = load_all_certs_from_directory(unicode_dir.to_str().unwrap()); assert!(result.is_err()); assert!( result .unwrap_err() .to_string() .contains("No valid certificate/private key pair found") ); } #[test] fn test_concurrent_access_safety() { use std::sync::Arc; use std::thread; let temp_dir = TempDir::new().unwrap(); let dir_path = Arc::new(temp_dir.path().to_string_lossy().to_string()); let handles: Vec<_> = (0..5) .map(|_| { let path = Arc::clone(&dir_path); thread::spawn(move || { let result = load_all_certs_from_directory(&path); // All should fail since directory is empty assert!(result.is_err()); }) }) .collect(); for handle in handles { handle.join().expect("Thread should complete successfully"); } } #[test] fn test_memory_efficiency() { // Test that error types are reasonably sized use std::mem; let error = certs_error("test".to_string()); let error_size = mem::size_of_val(&error); // Error should not be excessively large assert!(error_size < 1024, "Error size should be reasonable, got {error_size} bytes"); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/dirs.rs
crates/utils/src/dirs.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use rustfs_config::{DEFAULT_LOG_DIR, DEFAULT_LOG_FILENAME}; use std::env; use std::fs; use std::path::{Path, PathBuf}; use tracing::debug; /// Get the absolute path to the current project /// /// This function will try the following method to get the project path: /// 1. Use the `CARGO_MANIFEST_DIR` environment variable to get the project root directory. /// 2. Use `std::env::current_exe()` to get the executable file path and deduce the project root directory. /// 3. Use `std::env::current_dir()` to get the current working directory and try to deduce the project root directory. /// /// If all methods fail, an error is returned. /// /// # Returns /// - `Ok(PathBuf)`: The absolute path of the project that was successfully obtained. /// - `Err(String)`: Error message for the failed path. /// pub fn get_project_root() -> Result<PathBuf, String> { // Try to get the project root directory through the CARGO_MANIFEST_DIR environment variable if let Ok(manifest_dir) = env::var("CARGO_MANIFEST_DIR") { let project_root = Path::new(&manifest_dir).to_path_buf(); debug!("Get the project root directory with CARGO_MANIFEST_DIR:{}", project_root.display()); return Ok(project_root); } // Try to deduce the project root directory through the current executable file path if let Ok(current_exe) = env::current_exe() { let mut project_root = current_exe; // Assume that the project root directory is in the parent directory of the parent directory of the executable path (usually target/debug or target/release) project_root.pop(); // Remove the executable file name project_root.pop(); // Remove target/debug or target/release debug!("Deduce the project root directory through current_exe:{}", project_root.display()); return Ok(project_root); } // Try to deduce the project root directory from the current working directory if let Ok(mut current_dir) = env::current_dir() { // Assume that the project root directory is in the parent directory of the current working directory current_dir.pop(); debug!("Deduce the project root directory through current_dir:{}", current_dir.display()); return Ok(current_dir); } // If all methods fail, return an error Err("The project root directory cannot be obtained. Please check the running environment and project structure.".to_string()) } /// Get the log directory as a string /// This function will try to find a writable log directory in the following order: /// /// 1. Environment variables are specified /// 2. System temporary directory /// 3. User home directory /// 4. Current working directory /// 5. Relative path /// /// # Arguments /// * `key` - The environment variable key to check for log directory /// /// # Returns /// * `String` - The log directory path as a string /// pub fn get_log_directory_to_string(key: &str) -> String { get_log_directory(key).to_string_lossy().to_string() } /// Get the log directory /// This function will try to find a writable log directory in the following order: /// /// 1. Environment variables are specified /// 2. System temporary directory /// 3. User home directory /// 4. Current working directory /// 5. Relative path /// /// # Arguments /// * `key` - The environment variable key to check for log directory /// /// # Returns /// * `PathBuf` - The log directory path /// pub fn get_log_directory(key: &str) -> PathBuf { // Environment variables are specified if let Ok(log_dir) = env::var(key) { let path = PathBuf::from(log_dir); if ensure_directory_writable(&path) { return path; } } // System temporary directory if let Ok(mut temp_dir) = env::temp_dir().canonicalize() { temp_dir.push(DEFAULT_LOG_FILENAME); temp_dir.push(DEFAULT_LOG_DIR); if ensure_directory_writable(&temp_dir) { return temp_dir; } } // User home directory if let Ok(home_dir) = env::var("HOME").or_else(|_| env::var("USERPROFILE")) { let mut path = PathBuf::from(home_dir); path.push(format!(".{DEFAULT_LOG_FILENAME}")); path.push(DEFAULT_LOG_DIR); if ensure_directory_writable(&path) { return path; } } // Current working directory if let Ok(current_dir) = env::current_dir() { let mut path = current_dir; path.push(DEFAULT_LOG_DIR); if ensure_directory_writable(&path) { return path; } } // Relative path PathBuf::from(DEFAULT_LOG_DIR) } fn ensure_directory_writable(path: &PathBuf) -> bool { // Try creating a catalog if fs::create_dir_all(path).is_err() { return false; } // Check write permissions let test_file = path.join(".write_test"); match fs::write(&test_file, "test") { Ok(_) => { let _ = fs::remove_file(&test_file); true } Err(_) => false, } } #[cfg(test)] mod tests { use super::*; #[test] fn test_get_project_root() { match get_project_root() { Ok(path) => { assert!(path.exists(), "The project root directory does not exist:{}", path.display()); println!("The test is passed, the project root directory:{}", path.display()); } Err(e) => panic!("Failed to get the project root directory:{e}"), } } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/hash.rs
crates/utils/src/hash.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use highway::{HighwayHash, HighwayHasher, Key}; use md5::{Digest, Md5}; use serde::{Deserialize, Serialize}; use sha2::Sha256; /// The fixed key for HighwayHash256. DO NOT change for compatibility. const HIGHWAY_HASH256_KEY: [u64; 4] = [3, 4, 2, 1]; #[derive(Serialize, Deserialize, Debug, PartialEq, Default, Clone, Eq, Hash)] /// Supported hash algorithms for bitrot protection. pub enum HashAlgorithm { // SHA256 represents the SHA-256 hash function SHA256, // HighwayHash256 represents the HighwayHash-256 hash function HighwayHash256, // HighwayHash256S represents the Streaming HighwayHash-256 hash function #[default] HighwayHash256S, // BLAKE2b512 represents the BLAKE2b-512 hash function BLAKE2b512, /// MD5 (128-bit) Md5, /// No hash (for testing or unprotected data) None, } enum HashEncoded { Md5([u8; 16]), Sha256([u8; 32]), HighwayHash256([u8; 32]), HighwayHash256S([u8; 32]), Blake2b512(blake3::Hash), None, } impl AsRef<[u8]> for HashEncoded { #[inline] fn as_ref(&self) -> &[u8] { match self { HashEncoded::Md5(hash) => hash.as_ref(), HashEncoded::Sha256(hash) => hash.as_ref(), HashEncoded::HighwayHash256(hash) => hash.as_ref(), HashEncoded::HighwayHash256S(hash) => hash.as_ref(), HashEncoded::Blake2b512(hash) => hash.as_bytes(), HashEncoded::None => &[], } } } #[inline] fn u8x32_from_u64x4(input: [u64; 4]) -> [u8; 32] { let mut output = [0u8; 32]; for (i, &n) in input.iter().enumerate() { output[i * 8..(i + 1) * 8].copy_from_slice(&n.to_le_bytes()); } output } impl HashAlgorithm { /// Hash the input data and return the hash result as Vec<u8>. /// /// # Arguments /// * `data` - A byte slice representing the data to be hashed /// /// # Returns /// A byte slice containing the hash of the input data /// pub fn hash_encode(&self, data: &[u8]) -> impl AsRef<[u8]> { match self { HashAlgorithm::Md5 => HashEncoded::Md5(Md5::digest(data).into()), HashAlgorithm::HighwayHash256 => { let mut hasher = HighwayHasher::new(Key(HIGHWAY_HASH256_KEY)); hasher.append(data); HashEncoded::HighwayHash256(u8x32_from_u64x4(hasher.finalize256())) } HashAlgorithm::SHA256 => HashEncoded::Sha256(Sha256::digest(data).into()), HashAlgorithm::HighwayHash256S => { let mut hasher = HighwayHasher::new(Key(HIGHWAY_HASH256_KEY)); hasher.append(data); HashEncoded::HighwayHash256S(u8x32_from_u64x4(hasher.finalize256())) } HashAlgorithm::BLAKE2b512 => HashEncoded::Blake2b512(blake3::hash(data)), HashAlgorithm::None => HashEncoded::None, } } /// Return the output size in bytes for the hash algorithm. /// /// # Returns /// The size in bytes of the hash output /// pub fn size(&self) -> usize { match self { HashAlgorithm::SHA256 => 32, HashAlgorithm::HighwayHash256 => 32, HashAlgorithm::HighwayHash256S => 32, HashAlgorithm::BLAKE2b512 => 32, // blake3 outputs 32 bytes by default HashAlgorithm::Md5 => 16, HashAlgorithm::None => 0, } } } use siphasher::sip::SipHasher; pub const EMPTY_STRING_SHA256_HASH: &str = "e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"; pub const DEFAULT_SIP_HASH_KEY: [u8; 16] = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]; /// SipHash function to hash a string key into a bucket index. /// /// # Arguments /// * `key` - The input string to be hashed /// * `cardinality` - The number of buckets /// * `id` - A 16-byte array used as the SipHash key /// /// # Returns /// A usize representing the bucket index /// pub fn sip_hash(key: &str, cardinality: usize, id: &[u8; 16]) -> usize { // Your key, must be 16 bytes // Calculate SipHash value of the string let result = SipHasher::new_with_key(id).hash(key.as_bytes()); (result as usize) % cardinality } /// CRC32 hash function to hash a string key into a bucket index. /// /// # Arguments /// * `key` - The input string to be hashed /// * `cardinality` - The number of buckets /// /// # Returns /// A usize representing the bucket index /// pub fn crc_hash(key: &str, cardinality: usize) -> usize { let mut hasher = crc_fast::Digest::new(crc_fast::CrcAlgorithm::Crc32IsoHdlc); hasher.update(key.as_bytes()); let checksum = hasher.finalize() as u32; checksum as usize % cardinality } #[cfg(test)] mod tests { use super::*; #[test] fn test_hash_algorithm_sizes() { assert_eq!(HashAlgorithm::Md5.size(), 16); assert_eq!(HashAlgorithm::HighwayHash256.size(), 32); assert_eq!(HashAlgorithm::HighwayHash256S.size(), 32); assert_eq!(HashAlgorithm::SHA256.size(), 32); assert_eq!(HashAlgorithm::BLAKE2b512.size(), 32); assert_eq!(HashAlgorithm::None.size(), 0); } #[test] fn test_hash_encode_none() { let data = b"test data"; let hash = HashAlgorithm::None.hash_encode(data); let hash = hash.as_ref(); assert_eq!(hash.len(), 0); } #[test] fn test_hash_encode_md5() { let data = b"test data"; let hash = HashAlgorithm::Md5.hash_encode(data); let hash = hash.as_ref(); assert_eq!(hash.len(), 16); // MD5 should be deterministic let hash2 = HashAlgorithm::Md5.hash_encode(data); let hash2 = hash2.as_ref(); assert_eq!(hash, hash2); } #[test] fn test_hash_encode_highway() { let data = b"test data"; let hash = HashAlgorithm::HighwayHash256.hash_encode(data); let hash = hash.as_ref(); assert_eq!(hash.len(), 32); // HighwayHash should be deterministic let hash2 = HashAlgorithm::HighwayHash256.hash_encode(data); let hash2 = hash2.as_ref(); assert_eq!(hash, hash2); } #[test] fn test_hash_encode_sha256() { let data = b"test data"; let hash = HashAlgorithm::SHA256.hash_encode(data); let hash = hash.as_ref(); assert_eq!(hash.len(), 32); // SHA256 should be deterministic let hash2 = HashAlgorithm::SHA256.hash_encode(data); let hash2 = hash2.as_ref(); assert_eq!(hash, hash2); } #[test] fn test_hash_encode_blake2b512() { let data = b"test data"; let hash = HashAlgorithm::BLAKE2b512.hash_encode(data); let hash = hash.as_ref(); assert_eq!(hash.len(), 32); // blake3 outputs 32 bytes by default // BLAKE2b512 should be deterministic let hash2 = HashAlgorithm::BLAKE2b512.hash_encode(data); let hash2 = hash2.as_ref(); assert_eq!(hash, hash2); } #[test] fn test_different_data_different_hashes() { let data1 = b"test data 1"; let data2 = b"test data 2"; let md5_hash1 = HashAlgorithm::Md5.hash_encode(data1); let md5_hash2 = HashAlgorithm::Md5.hash_encode(data2); assert_ne!(md5_hash1.as_ref(), md5_hash2.as_ref()); let highway_hash1 = HashAlgorithm::HighwayHash256.hash_encode(data1); let highway_hash2 = HashAlgorithm::HighwayHash256.hash_encode(data2); assert_ne!(highway_hash1.as_ref(), highway_hash2.as_ref()); let sha256_hash1 = HashAlgorithm::SHA256.hash_encode(data1); let sha256_hash2 = HashAlgorithm::SHA256.hash_encode(data2); assert_ne!(sha256_hash1.as_ref(), sha256_hash2.as_ref()); let blake_hash1 = HashAlgorithm::BLAKE2b512.hash_encode(data1); let blake_hash2 = HashAlgorithm::BLAKE2b512.hash_encode(data2); assert_ne!(blake_hash1.as_ref(), blake_hash2.as_ref()); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/ip.rs
crates/utils/src/ip.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::net::{IpAddr, Ipv4Addr}; /// Get the IP address of the machine /// /// Priority is given to trying to get the IPv4 address, and if it fails, try to get the IPv6 address. /// If both fail to retrieve, None is returned. /// /// # Returns /// * `Some(IpAddr)` - Native IP address (IPv4 or IPv6) /// * `None` - Unable to obtain any native IP address /// pub fn get_local_ip() -> Option<IpAddr> { local_ip_address::local_ip() .ok() .or_else(|| local_ip_address::local_ipv6().ok()) } /// Get the IP address of the machine as a string /// /// If the IP address cannot be obtained, returns "127.0.0.1" as the default value. /// /// # Returns /// * `String` - Native IP address (IPv4 or IPv6) as a string, or the default value /// pub fn get_local_ip_with_default() -> String { get_local_ip() .unwrap_or_else(|| IpAddr::V4(Ipv4Addr::new(127, 0, 0, 1))) // Provide a safe default value .to_string() } #[cfg(test)] mod tests { use super::*; use std::net::Ipv4Addr; #[test] fn test_get_local_ip_returns_some_ip() { // Test getting local IP address, should return Some value let ip = get_local_ip(); assert!(ip.is_some(), "Should be able to get local IP address"); if let Some(ip_addr) = ip { println!("Local IP address: {ip_addr}"); // Verify that the returned IP address is valid match ip_addr { IpAddr::V4(ipv4) => { assert!(!ipv4.is_unspecified(), "IPv4 should not be unspecified (0.0.0.0)"); println!("Got IPv4 address: {ipv4}"); } IpAddr::V6(ipv6) => { assert!(!ipv6.is_unspecified(), "IPv6 should not be unspecified (::)"); println!("Got IPv6 address: {ipv6}"); } } } } #[test] fn test_get_local_ip_with_default_never_empty() { // Test that function with default value never returns empty string let ip_string = get_local_ip_with_default(); assert!(!ip_string.is_empty(), "IP string should never be empty"); // Verify that the returned string can be parsed as a valid IP address let parsed_ip: Result<IpAddr, _> = ip_string.parse(); assert!(parsed_ip.is_ok(), "Returned string should be a valid IP address: {ip_string}"); println!("Local IP with default: {ip_string}"); } #[test] fn test_get_local_ip_with_default_fallback() { // Test whether the default value is 127.0.0.1 let default_ip = IpAddr::V4(Ipv4Addr::new(127, 0, 0, 1)); let ip_string = get_local_ip_with_default(); // If unable to get real IP, should return default value if get_local_ip().is_none() { assert_eq!(ip_string, default_ip.to_string()); } // In any case, should always return a valid IP address string let parsed: Result<IpAddr, _> = ip_string.parse(); assert!(parsed.is_ok(), "Should always return a valid IP string"); } #[test] fn test_ip_address_types() { // Test IP address type recognition if let Some(ip) = get_local_ip() { match ip { IpAddr::V4(ipv4) => { // Test IPv4 address properties println!("IPv4 address: {ipv4}"); assert!(!ipv4.is_multicast(), "Local IP should not be multicast"); assert!(!ipv4.is_broadcast(), "Local IP should not be broadcast"); // Check if it's a private address (usually local IP is private) let is_private = ipv4.is_private(); let is_loopback = ipv4.is_loopback(); println!("IPv4 is private: {is_private}, is loopback: {is_loopback}"); } IpAddr::V6(ipv6) => { // Test IPv6 address properties println!("IPv6 address: {ipv6}"); assert!(!ipv6.is_multicast(), "Local IP should not be multicast"); let is_loopback = ipv6.is_loopback(); println!("IPv6 is loopback: {is_loopback}"); } } } } #[test] fn test_ip_string_format() { // Test IP address string format let ip_string = get_local_ip_with_default(); // Verify string format assert!(!ip_string.contains(' '), "IP string should not contain spaces"); assert!(!ip_string.is_empty(), "IP string should not be empty"); // Verify round-trip conversion let parsed_ip: IpAddr = ip_string.parse().expect("Should parse as valid IP"); let back_to_string = parsed_ip.to_string(); // For standard IP addresses, round-trip conversion should be consistent println!("Original: {ip_string}, Parsed back: {back_to_string}"); } #[test] fn test_default_fallback_value() { // Test correctness of default fallback value let default_ip = IpAddr::V4(Ipv4Addr::new(127, 0, 0, 1)); assert_eq!(default_ip.to_string(), "127.0.0.1"); // Verify default IP properties if let IpAddr::V4(ipv4) = default_ip { assert!(ipv4.is_loopback(), "Default IP should be loopback"); assert!(!ipv4.is_unspecified(), "Default IP should not be unspecified"); assert!(!ipv4.is_multicast(), "Default IP should not be multicast"); } } #[test] fn test_consistency_between_functions() { // Test consistency between the two functions let ip_option = get_local_ip(); let ip_string = get_local_ip_with_default(); match ip_option { Some(ip) => { // If get_local_ip returns Some, then get_local_ip_with_default should return the same IP assert_eq!(ip.to_string(), ip_string, "Both functions should return the same IP when available"); } None => { // If get_local_ip returns None, then get_local_ip_with_default should return default value assert_eq!(ip_string, "127.0.0.1", "Should return default value when no IP is available"); } } } #[test] fn test_multiple_calls_consistency() { // Test consistency of multiple calls let ip1 = get_local_ip(); let ip2 = get_local_ip(); let ip_str1 = get_local_ip_with_default(); let ip_str2 = get_local_ip_with_default(); // Multiple calls should return the same result assert_eq!(ip1, ip2, "Multiple calls to get_local_ip should return same result"); assert_eq!(ip_str1, ip_str2, "Multiple calls to get_local_ip_with_default should return same result"); } #[cfg(feature = "integration")] #[test] fn test_network_connectivity() { // Integration test: verify that the obtained IP address can be used for network connections if let Some(ip) = get_local_ip() { match ip { IpAddr::V4(ipv4) => { // For IPv4, check if it's a valid network address assert!(!ipv4.is_unspecified(), "Should not be 0.0.0.0"); // If it's not a loopback address, it should be routable if !ipv4.is_loopback() { println!("Got routable IPv4: {ipv4}"); } } IpAddr::V6(ipv6) => { // For IPv6, check if it's a valid network address assert!(!ipv6.is_unspecified(), "Should not be ::"); if !ipv6.is_loopback() { println!("Got routable IPv6: {ipv6}"); } } } } } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/crypto.rs
crates/utils/src/crypto.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use hex_simd::{AsOut, AsciiCase}; use hmac::{Hmac, KeyInit, Mac}; use hyper::body::Bytes; use sha1::Sha1; use sha2::{Digest, Sha256}; use std::mem::MaybeUninit; /// Base64 URL safe encoding without padding /// `base64_encode_url_safe_no_pad(input)` /// /// # Arguments /// * `input` - A byte slice to be encoded /// /// # Returns /// A `String` containing the Base64 URL safe encoded representation of the input data pub fn base64_encode_url_safe_no_pad(input: &[u8]) -> String { base64_simd::URL_SAFE_NO_PAD.encode_to_string(input) } /// Base64 URL safe decoding without padding /// `base64_decode_url_safe_no_pad(input)` /// /// # Arguments /// * `input` - A byte slice containing the Base64 URL safe encoded data /// /// # Returns /// A `Result` containing a `Vec<u8>` with the decoded data or a `base64_simd::Error` if decoding fails /// /// # Errors /// This function will return an error if the input data is not valid Base64 URL safe encoded data /// pub fn base64_decode_url_safe_no_pad(input: &[u8]) -> Result<Vec<u8>, base64_simd::Error> { base64_simd::URL_SAFE_NO_PAD.decode_to_vec(input) } /// encode to hex string (lowercase) /// `hex(data)` /// /// # Arguments /// * `data` - A byte slice to be encoded /// /// # Returns /// A `String` containing the hexadecimal representation of the input data in lowercase /// pub fn hex(data: impl AsRef<[u8]>) -> String { hex_simd::encode_to_string(data, hex_simd::AsciiCase::Lower) } /// verify sha256 checksum string /// /// # Arguments /// * `s` - A string slice to be verified /// /// # Returns /// A `bool` indicating whether the input string is a valid SHA-256 checksum (64 /// pub fn is_sha256_checksum(s: &str) -> bool { // TODO: optimize let is_lowercase_hex = |c: u8| matches!(c, b'0'..=b'9' | b'a'..=b'f'); s.len() == 64 && s.as_bytes().iter().copied().all(is_lowercase_hex) } /// HMAC-SHA1 hashing /// `hmac_sha1(key, data)` /// /// # Arguments /// * `key` - A byte slice representing the HMAC key /// * `data` - A byte slice representing the data to be hashed /// /// # Returns /// A 20-byte array containing the HMAC-SHA1 hash of the input data using the provided key /// pub fn hmac_sha1(key: impl AsRef<[u8]>, data: impl AsRef<[u8]>) -> [u8; 20] { let mut m = <Hmac<Sha1>>::new_from_slice(key.as_ref()).unwrap(); m.update(data.as_ref()); m.finalize().into_bytes().into() } /// HMAC-SHA256 hashing /// `hmac_sha256(key, data)` /// /// # Arguments /// * `key` - A byte slice representing the HMAC key /// * `data` - A byte slice representing the data to be hashed /// /// # Returns /// A 32-byte array containing the HMAC-SHA256 hash of the input data using the provided key /// pub fn hmac_sha256(key: impl AsRef<[u8]>, data: impl AsRef<[u8]>) -> [u8; 32] { let mut m = Hmac::<Sha256>::new_from_slice(key.as_ref()).unwrap(); m.update(data.as_ref()); m.finalize().into_bytes().into() } /// `f(hex(src))` fn hex_bytes32<R>(src: impl AsRef<[u8]>, f: impl FnOnce(&str) -> R) -> R { let buf: &mut [_] = &mut [MaybeUninit::uninit(); 64]; let ans = hex_simd::encode_as_str(src.as_ref(), buf.as_out(), AsciiCase::Lower); f(ans) } fn sha256(data: &[u8]) -> impl AsRef<[u8; 32]> + use<> { <Sha256 as Digest>::digest(data) } fn sha256_chunk(chunk: &[Bytes]) -> impl AsRef<[u8; 32]> + use<> { let mut h = <Sha256 as Digest>::new(); chunk.iter().for_each(|data| h.update(data)); h.finalize() } /// hex of sha256 `f(hex(sha256(data)))` /// /// # Arguments /// * `data` - A byte slice representing the data to be hashed /// * `f` - A closure that takes a string slice and returns a value of type `R` /// /// # Returns /// A value of type `R` returned by the closure `f` after processing the hexadecimal /// representation of the SHA-256 hash of the input data /// pub fn hex_sha256<R>(data: &[u8], f: impl FnOnce(&str) -> R) -> R { hex_bytes32(sha256(data).as_ref(), f) } /// `f(hex(sha256(chunk)))` pub fn hex_sha256_chunk<R>(chunk: &[Bytes], f: impl FnOnce(&str) -> R) -> R { hex_bytes32(sha256_chunk(chunk).as_ref(), f) } #[test] fn test_base64_encoding_decoding() { let original_uuid_timestamp = "c0194290-d911-45cb-8e12-79ec563f46a8x1735460504394878000"; let encoded_string = base64_encode_url_safe_no_pad(original_uuid_timestamp.as_bytes()); println!("Encoded: {}", &encoded_string); let decoded_bytes = base64_decode_url_safe_no_pad(encoded_string.clone().as_bytes()).unwrap(); let decoded_string = String::from_utf8(decoded_bytes).unwrap(); assert_eq!(decoded_string, original_uuid_timestamp) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/retry.rs
crates/utils/src/retry.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use futures::Stream; use hyper::http; use std::{ pin::Pin, sync::LazyLock, task::{Context, Poll}, time::Duration, }; use tokio::time::{Interval, MissedTickBehavior, interval}; pub const MAX_RETRY: i64 = 10; pub const MAX_JITTER: f64 = 1.0; pub const NO_JITTER: f64 = 0.0; pub const DEFAULT_RETRY_UNIT: Duration = Duration::from_millis(200); pub const DEFAULT_RETRY_CAP: Duration = Duration::from_secs(1); #[derive(Debug)] pub struct RetryTimer { base_sleep: Duration, max_sleep: Duration, jitter: f64, random: u64, max_retry: i64, rem: i64, timer: Option<Interval>, } impl RetryTimer { pub fn new(max_retry: i64, base_sleep: Duration, max_sleep: Duration, jitter: f64, random: u64) -> Self { Self { base_sleep, max_sleep, jitter, random, max_retry, rem: max_retry, timer: None, } } } impl Stream for RetryTimer { type Item = (); fn poll_next(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Option<()>> { if self.rem == 0 { return Poll::Ready(None); } let jitter = self.jitter.clamp(NO_JITTER, MAX_JITTER); let attempt = self.max_retry - self.rem; let mut sleep = self.base_sleep * (1 << attempt); if sleep > self.max_sleep { sleep = self.max_sleep; } if (jitter - NO_JITTER).abs() > 1e-9 { let sleep_ms = sleep.as_millis(); let reduction = ((sleep_ms as f64) * (self.random as f64 * jitter / 100_f64)).round() as u128; let jittered_ms = sleep_ms.saturating_sub(reduction); let clamped_ms = std::cmp::min(jittered_ms.max(1), u128::from(u64::MAX)); sleep = Duration::from_millis(clamped_ms as u64); } //println!("sleep: {sleep:?}"); if self.timer.is_none() { let mut timer = interval(sleep); timer.set_missed_tick_behavior(MissedTickBehavior::Delay); self.timer = Some(timer); } let mut timer = self.timer.as_mut().unwrap(); match Pin::new(&mut timer).poll_tick(cx) { Poll::Ready(_) => { self.rem -= 1; if self.rem > 0 { let mut new_timer = interval(sleep); new_timer.set_missed_tick_behavior(MissedTickBehavior::Delay); new_timer.reset(); self.timer = Some(new_timer); } Poll::Ready(Some(())) } Poll::Pending => Poll::Pending, } } } static RETRYABLE_S3CODES: LazyLock<Vec<String>> = LazyLock::new(|| { vec![ "RequestError".to_string(), "RequestTimeout".to_string(), "Throttling".to_string(), "ThrottlingException".to_string(), "RequestLimitExceeded".to_string(), "RequestThrottled".to_string(), "InternalError".to_string(), "ExpiredToken".to_string(), "ExpiredTokenException".to_string(), "SlowDown".to_string(), ] }); static RETRYABLE_HTTP_STATUSCODES: LazyLock<Vec<http::StatusCode>> = LazyLock::new(|| { vec![ http::StatusCode::REQUEST_TIMEOUT, http::StatusCode::TOO_MANY_REQUESTS, //499, http::StatusCode::INTERNAL_SERVER_ERROR, http::StatusCode::BAD_GATEWAY, http::StatusCode::SERVICE_UNAVAILABLE, http::StatusCode::GATEWAY_TIMEOUT, //520, ] }); pub fn is_s3code_retryable(s3code: &str) -> bool { RETRYABLE_S3CODES.contains(&s3code.to_string()) } pub fn is_http_status_retryable(http_statuscode: &http::StatusCode) -> bool { RETRYABLE_HTTP_STATUSCODES.contains(http_statuscode) } pub fn is_request_error_retryable(_err: std::io::Error) -> bool { /*if err == Err::Canceled || err == Err::DeadlineExceeded { return err() == nil; } let uerr = err.(*url.Error); if uerr.is_ok() { let e = uerr.unwrap(); return match e.type { x509.UnknownAuthorityError => { false } _ => true, }; return match e.error() { "http: server gave HTTP response to HTTPS client" => { false } _ => rue, }; } true*/ todo!(); } #[cfg(test)] #[allow(unused_imports)] mod tests { use super::*; use futures::StreamExt; use rand::Rng; use std::time::UNIX_EPOCH; #[tokio::test] async fn test_retry() { let req_retry = 10; let random = rand::rng().random_range(0..=100); let mut retry_timer = RetryTimer::new(req_retry, DEFAULT_RETRY_UNIT, DEFAULT_RETRY_CAP, MAX_JITTER, random); println!("retry_timer: {retry_timer:?}"); while retry_timer.next().await.is_some() { println!( "\ntime: {:?}", std::time::SystemTime::now().duration_since(UNIX_EPOCH).unwrap().as_millis() ); } } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/net.rs
crates/utils/src/net.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use bytes::Bytes; use futures::{Stream, StreamExt, pin_mut}; #[cfg(test)] use std::sync::MutexGuard; use std::{ collections::{HashMap, HashSet}, fmt::Display, io::Error, net::{IpAddr, Ipv6Addr, SocketAddr, TcpListener, ToSocketAddrs}, sync::{Arc, LazyLock, Mutex, RwLock}, time::{Duration, Instant}, }; use tracing::{error, info}; use transform_stream::AsyncTryStream; use url::{Host, Url}; static LOCAL_IPS: LazyLock<Vec<IpAddr>> = LazyLock::new(|| must_get_local_ips().unwrap()); #[derive(Debug, Clone)] struct DnsCacheEntry { ips: HashSet<IpAddr>, cached_at: Instant, } impl DnsCacheEntry { fn new(ips: HashSet<IpAddr>) -> Self { Self { ips, cached_at: Instant::now(), } } fn is_expired(&self, ttl: Duration) -> bool { self.cached_at.elapsed() > ttl } } static DNS_CACHE: LazyLock<Mutex<HashMap<String, DnsCacheEntry>>> = LazyLock::new(|| Mutex::new(HashMap::new())); const DNS_CACHE_TTL: Duration = Duration::from_secs(300); // 5 minutes type DynDnsResolver = dyn Fn(&str) -> std::io::Result<HashSet<IpAddr>> + Send + Sync + 'static; static CUSTOM_DNS_RESOLVER: LazyLock<RwLock<Option<Arc<DynDnsResolver>>>> = LazyLock::new(|| RwLock::new(None)); fn resolve_domain(domain: &str) -> std::io::Result<HashSet<IpAddr>> { if let Some(resolver) = CUSTOM_DNS_RESOLVER.read().unwrap().clone() { return resolver(domain); } (domain, 0) .to_socket_addrs() .map(|v| v.map(|v| v.ip()).collect::<HashSet<_>>()) .map_err(Error::other) } #[cfg(test)] fn clear_dns_cache() { if let Ok(mut cache) = DNS_CACHE.lock() { cache.clear(); } } #[cfg(test)] static DNS_RESOLVER_TEST_LOCK: LazyLock<Mutex<()>> = LazyLock::new(|| Mutex::new(())); #[cfg(test)] fn reset_dns_resolver_inner() { *CUSTOM_DNS_RESOLVER.write().unwrap() = None; clear_dns_cache(); } #[cfg(test)] pub struct MockResolverGuard { _lock: MutexGuard<'static, ()>, } #[cfg(test)] impl Drop for MockResolverGuard { fn drop(&mut self) { reset_dns_resolver_inner(); } } #[cfg(test)] pub fn set_mock_dns_resolver<F>(resolver: F) -> MockResolverGuard where F: Fn(&str) -> std::io::Result<HashSet<IpAddr>> + Send + Sync + 'static, { let lock = DNS_RESOLVER_TEST_LOCK.lock().unwrap(); *CUSTOM_DNS_RESOLVER.write().unwrap() = Some(Arc::new(resolver)); clear_dns_cache(); MockResolverGuard { _lock: lock } } #[cfg(test)] pub fn reset_dns_resolver() { let _lock = DNS_RESOLVER_TEST_LOCK.lock().unwrap(); reset_dns_resolver_inner(); } /// helper for validating if the provided arg is an ip address. pub fn is_socket_addr(addr: &str) -> bool { // TODO IPv6 zone information? addr.parse::<SocketAddr>().is_ok() || addr.parse::<IpAddr>().is_ok() } /// checks if server_addr is valid and local host. pub fn check_local_server_addr(server_addr: &str) -> std::io::Result<SocketAddr> { let addr: Vec<SocketAddr> = match server_addr.to_socket_addrs() { Ok(addr) => addr.collect(), Err(err) => return Err(Error::other(err)), }; // 0.0.0.0 is a wildcard address and refers to local network // addresses. I.e, 0.0.0.0:9000 like ":9000" refers to port // 9000 on localhost. for a in addr { if a.ip().is_unspecified() { return Ok(a); } let host = match a { SocketAddr::V4(a) => Host::<&str>::Ipv4(*a.ip()), SocketAddr::V6(a) => Host::Ipv6(*a.ip()), }; if is_local_host(host, 0, 0)? { return Ok(a); } } Err(Error::other("host in server address should be this server")) } /// checks if the given parameter correspond to one of /// the local IP of the current machine pub fn is_local_host(host: Host<&str>, port: u16, local_port: u16) -> std::io::Result<bool> { let local_set: HashSet<IpAddr> = LOCAL_IPS.iter().copied().collect(); let is_local_host = match host { Host::Domain(domain) => { let ips = resolve_domain(domain)?.into_iter().collect::<Vec<_>>(); ips.iter().any(|ip| local_set.contains(ip)) } Host::Ipv4(ip) => local_set.contains(&IpAddr::V4(ip)), Host::Ipv6(ip) => local_set.contains(&IpAddr::V6(ip)), }; if port > 0 { return Ok(is_local_host && port == local_port); } Ok(is_local_host) } /// returns IP address of given host using layered DNS resolution. /// /// This is the async version of `get_host_ip()` that provides enhanced DNS resolution /// with Kubernetes support when the "net" feature is enabled. pub async fn get_host_ip(host: Host<&str>) -> std::io::Result<HashSet<IpAddr>> { match host { Host::Domain(domain) => { // Check cache first if CUSTOM_DNS_RESOLVER.read().unwrap().is_none() && let Ok(mut cache) = DNS_CACHE.lock() && let Some(entry) = cache.get(domain) { if !entry.is_expired(DNS_CACHE_TTL) { return Ok(entry.ips.clone()); } // Remove expired entry cache.remove(domain); } info!("Cache miss for domain {domain}, querying system resolver."); // Fallback to standard resolution when DNS resolver is not available match resolve_domain(domain) { Ok(ips) => { if CUSTOM_DNS_RESOLVER.read().unwrap().is_none() { // Cache the result if let Ok(mut cache) = DNS_CACHE.lock() { cache.insert(domain.to_string(), DnsCacheEntry::new(ips.clone())); // Limit cache size to prevent memory bloat if cache.len() > 1000 { cache.retain(|_, v| !v.is_expired(DNS_CACHE_TTL)); } } } info!("System query for domain {domain}: {:?}", ips); Ok(ips) } Err(err) => { error!("Failed to resolve domain {domain} using system resolver, err: {err}"); Err(Error::other(err)) } } } Host::Ipv4(ip) => Ok([IpAddr::V4(ip)].into_iter().collect()), Host::Ipv6(ip) => Ok([IpAddr::V6(ip)].into_iter().collect()), } } pub fn get_available_port() -> u16 { TcpListener::bind("0.0.0.0:0").unwrap().local_addr().unwrap().port() } /// returns IPs of local interface pub fn must_get_local_ips() -> std::io::Result<Vec<IpAddr>> { match netif::up() { Ok(up) => Ok(up.map(|x| x.address().to_owned()).collect()), Err(err) => Err(Error::other(format!("Unable to get IP addresses of this host: {err}"))), } } pub fn get_default_location(_u: Url, _region_override: &str) -> String { todo!(); } pub fn get_endpoint_url(endpoint: &str, secure: bool) -> Result<Url, Error> { let mut scheme = "https"; if !secure { scheme = "http"; } let endpoint_url_str = format!("{scheme}://{endpoint}"); let Ok(endpoint_url) = Url::parse(&endpoint_url_str) else { return Err(Error::other("url parse error.")); }; //is_valid_endpoint_url(endpoint_url)?; Ok(endpoint_url) } pub const DEFAULT_DIAL_TIMEOUT: i64 = 5; const ALLOWED_CUSTOM_QUERY_PREFIX: &str = "x-"; pub fn is_custom_query_value(qs_key: &str) -> bool { qs_key.starts_with(ALLOWED_CUSTOM_QUERY_PREFIX) } #[derive(Debug, Clone)] pub struct XHost { pub name: String, pub port: u16, pub is_port_set: bool, } impl Display for XHost { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { if !self.is_port_set { write!(f, "{}", self.name) } else if self.name.contains(':') { write!(f, "[{}]:{}", self.name, self.port) } else { write!(f, "{}:{}", self.name, self.port) } } } impl TryFrom<String> for XHost { type Error = Error; fn try_from(value: String) -> Result<Self, Self::Error> { if let Some(addr) = value.to_socket_addrs()?.next() { Ok(Self { name: addr.ip().to_string(), port: addr.port(), is_port_set: addr.port() > 0, }) } else { Err(Error::new(std::io::ErrorKind::InvalidData, "value invalid")) } } } pub fn parse_and_resolve_address(addr_str: &str) -> std::io::Result<SocketAddr> { let resolved_addr: SocketAddr = if let Some(port) = addr_str.strip_prefix(":") { let port_str = port; let port: u16 = port_str .parse() .map_err(|e| Error::other(format!("Invalid port format: {addr_str}, err:{e:?}")))?; let final_port = if port == 0 { get_available_port() // assume get_available_port is available here } else { port }; SocketAddr::new(IpAddr::V6(Ipv6Addr::UNSPECIFIED), final_port) } else { let mut addr = check_local_server_addr(addr_str)?; // assume check_local_server_addr is available here if addr.port() == 0 { addr.set_port(get_available_port()); } addr }; Ok(resolved_addr) } #[allow(dead_code)] pub fn bytes_stream<S, E>(stream: S, content_length: usize) -> impl Stream<Item = Result<Bytes, E>> + Send + 'static where S: Stream<Item = Result<Bytes, E>> + Send + 'static, E: Send + 'static, { AsyncTryStream::<Bytes, E, _>::new(|mut y| async move { pin_mut!(stream); let mut remaining: usize = content_length; while let Some(result) = stream.next().await { let mut bytes = result?; if bytes.len() > remaining { bytes.truncate(remaining); } remaining -= bytes.len(); y.yield_ok(bytes).await; } Ok(()) }) } #[cfg(test)] mod test { use super::*; use std::net::{Ipv4Addr, Ipv6Addr}; use std::{collections::HashSet, io::Error as IoError}; fn mock_resolver(domain: &str) -> std::io::Result<HashSet<IpAddr>> { match domain { "localhost" => Ok([ IpAddr::V4(Ipv4Addr::new(127, 0, 0, 1)), IpAddr::V6(Ipv6Addr::new(0, 0, 0, 0, 0, 0, 0, 1)), ] .into_iter() .collect()), "example.org" => Ok([IpAddr::V4(Ipv4Addr::new(192, 0, 2, 10))].into_iter().collect()), "invalid.nonexistent.domain.example" => Err(IoError::other("mock DNS failure")), _ => Ok(HashSet::new()), } } #[test] fn test_is_socket_addr() { let test_cases = [ // Valid IP addresses ("192.168.1.0", true), ("127.0.0.1", true), ("10.0.0.1", true), ("0.0.0.0", true), ("255.255.255.255", true), // Valid IPv6 addresses ("2001:db8::1", true), ("::1", true), ("::", true), ("fe80::1", true), // Valid socket addresses ("192.168.1.0:8080", true), ("127.0.0.1:9000", true), ("[2001:db8::1]:9000", true), ("[::1]:8080", true), ("0.0.0.0:0", true), // Invalid addresses ("localhost", false), ("localhost:9000", false), ("example.com", false), ("example.com:8080", false), ("http://192.168.1.0", false), ("http://192.168.1.0:9000", false), ("256.256.256.256", false), ("192.168.1", false), ("192.168.1.0.1", false), ("", false), (":", false), (":::", false), ("invalid_ip", false), ]; for (addr, expected) in test_cases { let result = is_socket_addr(addr); assert_eq!(expected, result, "addr: '{addr}', expected: {expected}, got: {result}"); } } #[test] fn test_check_local_server_addr() { // Test valid local addresses let valid_cases = ["localhost:54321", "127.0.0.1:9000", "0.0.0.0:9000", "[::1]:8080", "::1:8080"]; for addr in valid_cases { let result = check_local_server_addr(addr); assert!(result.is_ok(), "Expected '{addr}' to be valid, but got error: {result:?}"); } // Test invalid addresses let invalid_cases = [ ("localhost", "invalid socket address"), ("", "invalid socket address"), ("203.0.113.1:54321", "host in server address should be this server"), ("8.8.8.8:53", "host in server address should be this server"), (":-10", "invalid port value"), ("invalid:port", "invalid port value"), ]; for (addr, expected_error_pattern) in invalid_cases { let result = check_local_server_addr(addr); assert!(result.is_err(), "Expected '{addr}' to be invalid, but it was accepted: {result:?}"); let error_msg = result.unwrap_err().to_string(); assert!( error_msg.contains(expected_error_pattern) || error_msg.contains("invalid socket address"), "Error message '{error_msg}' doesn't contain expected pattern '{expected_error_pattern}' for address '{addr}'" ); } } #[test] fn test_is_local_host() { let _resolver_guard = set_mock_dns_resolver(mock_resolver); // Test localhost domain let localhost_host = Host::Domain("localhost"); assert!(is_local_host(localhost_host, 0, 0).unwrap()); // Test loopback IP addresses let ipv4_loopback = Host::Ipv4(Ipv4Addr::new(127, 0, 0, 1)); assert!(is_local_host(ipv4_loopback, 0, 0).unwrap()); let ipv6_loopback = Host::Ipv6(Ipv6Addr::new(0, 0, 0, 0, 0, 0, 0, 1)); assert!(is_local_host(ipv6_loopback, 0, 0).unwrap()); // Test port matching let localhost_with_port1 = Host::Domain("localhost"); assert!(is_local_host(localhost_with_port1, 8080, 8080).unwrap()); let localhost_with_port2 = Host::Domain("localhost"); assert!(!is_local_host(localhost_with_port2, 8080, 9000).unwrap()); // Test non-local host let external_host = Host::Ipv4(Ipv4Addr::new(8, 8, 8, 8)); assert!(!is_local_host(external_host, 0, 0).unwrap()); // Test invalid domain should return error let invalid_host = Host::Domain("invalid.nonexistent.domain.example"); assert!(is_local_host(invalid_host, 0, 0).is_err()); } #[tokio::test] async fn test_get_host_ip() { let _resolver_guard = set_mock_dns_resolver(mock_resolver); // Test IPv4 address let ipv4_host = Host::Ipv4(Ipv4Addr::new(192, 168, 1, 1)); let ipv4_result = get_host_ip(ipv4_host).await.unwrap(); assert_eq!(ipv4_result.len(), 1); assert!(ipv4_result.contains(&IpAddr::V4(Ipv4Addr::new(192, 168, 1, 1)))); // Test IPv6 address let ipv6_host = Host::Ipv6(Ipv6Addr::new(0x2001, 0xdb8, 0, 0, 0, 0, 0, 1)); let ipv6_result = get_host_ip(ipv6_host).await.unwrap(); assert_eq!(ipv6_result.len(), 1); assert!(ipv6_result.contains(&IpAddr::V6(Ipv6Addr::new(0x2001, 0xdb8, 0, 0, 0, 0, 0, 1)))); // Test localhost domain let localhost_host = Host::Domain("localhost"); let localhost_result = get_host_ip(localhost_host).await.unwrap(); assert!(!localhost_result.is_empty()); // Should contain at least loopback address assert!( localhost_result.contains(&IpAddr::V4(Ipv4Addr::new(127, 0, 0, 1))) || localhost_result.contains(&IpAddr::V6(Ipv6Addr::new(0, 0, 0, 0, 0, 0, 0, 1))) ); // Test invalid domain let invalid_host = Host::Domain("invalid.nonexistent.domain.example"); assert!(get_host_ip(invalid_host).await.is_err()); } #[test] fn test_get_available_port() { let port1 = get_available_port(); let port2 = get_available_port(); // Port should be in valid range (u16 max is always <= 65535) assert!(port1 > 0); assert!(port2 > 0); // Different calls should typically return different ports assert_ne!(port1, port2); } #[test] fn test_must_get_local_ips() { let local_ips = must_get_local_ips().unwrap(); let local_set: HashSet<IpAddr> = local_ips.into_iter().collect(); // Should contain loopback addresses assert!(local_set.contains(&IpAddr::V4(Ipv4Addr::new(127, 0, 0, 1)))); // Should not be empty assert!(!local_set.is_empty()); // All IPs should be valid for ip in &local_set { match ip { IpAddr::V4(_) | IpAddr::V6(_) => {} // Valid } } } #[test] fn test_xhost_display() { // Test without port let host_no_port = XHost { name: "example.com".to_string(), port: 0, is_port_set: false, }; assert_eq!(host_no_port.to_string(), "example.com"); // Test with port (IPv4-like name) let host_with_port = XHost { name: "192.168.1.1".to_string(), port: 8080, is_port_set: true, }; assert_eq!(host_with_port.to_string(), "192.168.1.1:8080"); // Test with port (IPv6-like name) let host_ipv6_with_port = XHost { name: "2001:db8::1".to_string(), port: 9000, is_port_set: true, }; assert_eq!(host_ipv6_with_port.to_string(), "[2001:db8::1]:9000"); // Test domain name with port let host_domain_with_port = XHost { name: "example.com".to_string(), port: 443, is_port_set: true, }; assert_eq!(host_domain_with_port.to_string(), "example.com:443"); } #[test] fn test_xhost_try_from() { // Test valid IPv4 address with port let result = XHost::try_from("192.168.1.1:8080".to_string()).unwrap(); assert_eq!(result.name, "192.168.1.1"); assert_eq!(result.port, 8080); assert!(result.is_port_set); // Test valid IPv4 address without port let result = XHost::try_from("192.168.1.1:0".to_string()).unwrap(); assert_eq!(result.name, "192.168.1.1"); assert_eq!(result.port, 0); assert!(!result.is_port_set); // Test valid IPv6 address with port let result = XHost::try_from("[2001:db8::1]:9000".to_string()).unwrap(); assert_eq!(result.name, "2001:db8::1"); assert_eq!(result.port, 9000); assert!(result.is_port_set); // Test localhost with port (localhost may resolve to either IPv4 or IPv6) let result = XHost::try_from("localhost:3000".to_string()).unwrap(); // localhost can resolve to either 127.0.0.1 or ::1 depending on system configuration assert!(result.name == "127.0.0.1" || result.name == "::1"); assert_eq!(result.port, 3000); assert!(result.is_port_set); // Test invalid format let result = XHost::try_from("invalid_format".to_string()); assert!(result.is_err()); // Test empty string let result = XHost::try_from("".to_string()); assert!(result.is_err()); } #[test] fn test_parse_and_resolve_address() { // Test port-only format let result = parse_and_resolve_address(":8080").unwrap(); assert_eq!(result.ip(), IpAddr::V6(Ipv6Addr::UNSPECIFIED)); assert_eq!(result.port(), 8080); // Test port-only format with port 0 (should get available port) let result = parse_and_resolve_address(":0").unwrap(); assert_eq!(result.ip(), IpAddr::V6(Ipv6Addr::UNSPECIFIED)); assert!(result.port() > 0); // Test localhost with port let result = parse_and_resolve_address("localhost:9000").unwrap(); assert_eq!(result.port(), 9000); // Test localhost with port 0 (should get available port) let result = parse_and_resolve_address("localhost:0").unwrap(); assert!(result.port() > 0); // Test 0.0.0.0 with port let result = parse_and_resolve_address("0.0.0.0:7000").unwrap(); assert_eq!(result.ip(), IpAddr::V4(Ipv4Addr::new(0, 0, 0, 0))); assert_eq!(result.port(), 7000); // Test invalid port format let result = parse_and_resolve_address(":invalid_port"); assert!(result.is_err()); // Test invalid address let result = parse_and_resolve_address("example.org:8080"); assert!(result.is_err()); } #[test] fn test_edge_cases() { // Test empty string for is_socket_addr assert!(!is_socket_addr("")); // Test single colon for is_socket_addr assert!(!is_socket_addr(":")); // Test malformed IPv6 for is_socket_addr assert!(!is_socket_addr("[::]")); assert!(!is_socket_addr("[::1")); // Test very long strings let long_string = "a".repeat(1000); assert!(!is_socket_addr(&long_string)); // Test unicode characters assert!(!is_socket_addr("ΠΏΡ€ΠΈΠΌΠ΅Ρ€.example.com")); // Test special characters assert!(!is_socket_addr("test@example.com:8080")); assert!(!is_socket_addr("http://example.com:8080")); } #[test] fn test_boundary_values() { // Test port boundaries assert!(is_socket_addr("127.0.0.1:0")); assert!(is_socket_addr("127.0.0.1:65535")); assert!(!is_socket_addr("127.0.0.1:65536")); // Test IPv4 boundaries assert!(is_socket_addr("0.0.0.0")); assert!(is_socket_addr("255.255.255.255")); assert!(!is_socket_addr("256.0.0.0")); assert!(!is_socket_addr("0.0.0.256")); // Test XHost with boundary ports let host_max_port = XHost { name: "example.com".to_string(), port: 65535, is_port_set: true, }; assert_eq!(host_max_port.to_string(), "example.com:65535"); let host_zero_port = XHost { name: "example.com".to_string(), port: 0, is_port_set: true, }; assert_eq!(host_zero_port.to_string(), "example.com:0"); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/envs.rs
crates/utils/src/envs.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::env; /// Retrieve an environment variable as a specific type, with a default value if not set or parsing fails. /// 8-bit type: signed i8 /// /// #Parameters /// - `key`: The environment variable key to look up. /// - `default`: The default value to return if the environment variable is not set or parsing fails. /// /// #Returns /// - `i8`: The parsed value as i8 if successful, otherwise the default value. /// pub fn get_env_i8(key: &str, default: i8) -> i8 { env::var(key).ok().and_then(|v| v.parse().ok()).unwrap_or(default) } /// Retrieve an environment variable as a specific type, returning None if not set or parsing fails. /// 8-bit type: signed i8 /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `Option<i8>`: The parsed value as i8 if successful, otherwise None /// pub fn get_env_opt_i8(key: &str) -> Option<i8> { env::var(key).ok().and_then(|v| v.parse().ok()) } /// Retrieve an environment variable as a specific type, with a default value if not set or parsing fails. /// 8-bit type: unsigned u8 /// /// #Parameters /// - `key`: The environment variable key to look up. /// - `default`: The default value to return if the environment variable is not set or parsing fails. /// /// #Returns /// - `u8`: The parsed value as u8 if successful, otherwise the default value. /// pub fn get_env_u8(key: &str, default: u8) -> u8 { env::var(key).ok().and_then(|v| v.parse().ok()).unwrap_or(default) } /// Retrieve an environment variable as a specific type, returning None if not set or parsing fails. /// 8-bit type: unsigned u8 /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `Option<u8>`: The parsed value as u8 if successful, otherwise None /// pub fn get_env_opt_u8(key: &str) -> Option<u8> { env::var(key).ok().and_then(|v| v.parse().ok()) } /// Retrieve an environment variable as a specific type, with a default value if not set or parsing fails. /// 16-bit type: signed i16 /// /// #Parameters /// - `key`: The environment variable key to look up. /// - `default`: The default value to return if the environment variable is not set or parsing fails. /// /// #Returns /// - `i16`: The parsed value as i16 if successful, otherwise the default value. /// pub fn get_env_i16(key: &str, default: i16) -> i16 { env::var(key).ok().and_then(|v| v.parse().ok()).unwrap_or(default) } /// Retrieve an environment variable as a specific type, returning None if not set or parsing fails. /// 16-bit type: signed i16 /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `Option<i16>`: The parsed value as i16 if successful, otherwise None /// pub fn get_env_opt_i16(key: &str) -> Option<i16> { env::var(key).ok().and_then(|v| v.parse().ok()) } /// Retrieve an environment variable as a specific type, with a default value if not set or parsing fails. /// 16-bit type: unsigned u16 /// /// #Parameters /// - `key`: The environment variable key to look up. /// - `default`: The default value to return if the environment variable is not set or parsing fails. /// /// #Returns /// - `u16`: The parsed value as u16 if successful, otherwise the default value. /// pub fn get_env_u16(key: &str, default: u16) -> u16 { env::var(key).ok().and_then(|v| v.parse().ok()).unwrap_or(default) } /// Retrieve an environment variable as a specific type, returning None if not set or parsing fails. /// 16-bit type: unsigned u16 /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `Option<u16>`: The parsed value as u16 if successful, otherwise None /// pub fn get_env_u16_opt(key: &str) -> Option<u16> { env::var(key).ok().and_then(|v| v.parse().ok()) } /// Retrieve an environment variable as a specific type, returning None if not set or parsing fails. /// 16-bit type: unsigned u16 /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `Option<u16>`: The parsed value as u16 if successful, otherwise None /// pub fn get_env_opt_u16(key: &str) -> Option<u16> { get_env_u16_opt(key) } /// Retrieve an environment variable as a specific type, with a default value if not set or parsing fails. /// 32-bit type: signed i32 /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `i32`: The parsed value as i32 if successful, otherwise the default value. /// pub fn get_env_i32(key: &str, default: i32) -> i32 { env::var(key).ok().and_then(|v| v.parse().ok()).unwrap_or(default) } /// Retrieve an environment variable as a specific type, returning None if not set or parsing fails. /// 32-bit type: signed i32 /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `Option<i32>`: The parsed value as i32 if successful, otherwise None /// pub fn get_env_opt_i32(key: &str) -> Option<i32> { env::var(key).ok().and_then(|v| v.parse().ok()) } /// Retrieve an environment variable as a specific type, with a default value if not set or parsing fails. /// 32-bit type: unsigned u32 /// /// #Parameters /// - `key`: The environment variable key to look up. /// - `default`: The default value to return if the environment variable is not set or parsing fails. /// /// #Returns /// - `u32`: The parsed value as u32 if successful, otherwise the default value. /// pub fn get_env_u32(key: &str, default: u32) -> u32 { env::var(key).ok().and_then(|v| v.parse().ok()).unwrap_or(default) } /// Retrieve an environment variable as a specific type, returning None if not set or parsing fails. /// 32-bit type: unsigned u32 /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `Option<u32>`: The parsed value as u32 if successful, otherwise None /// pub fn get_env_opt_u32(key: &str) -> Option<u32> { env::var(key).ok().and_then(|v| v.parse().ok()) } /// Retrieve an environment variable as a specific type, with a default value if not set or parsing fails. /// /// #Parameters /// - `key`: The environment variable key to look up. /// - `default`: The default value to return if the environment variable is not set or parsing /// /// #Returns /// - `f32`: The parsed value as f32 if successful, otherwise the default value /// pub fn get_env_f32(key: &str, default: f32) -> f32 { env::var(key).ok().and_then(|v| v.parse().ok()).unwrap_or(default) } /// Retrieve an environment variable as a specific type, returning None if not set or parsing fails. /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `Option<f32>`: The parsed value as f32 if successful, otherwise None /// pub fn get_env_opt_f32(key: &str) -> Option<f32> { env::var(key).ok().and_then(|v| v.parse().ok()) } /// Retrieve an environment variable as a specific type, with a default value if not set or parsing fails. /// /// #Parameters /// - `key`: The environment variable key to look up. /// - `default`: The default value to return if the environment variable is not set or parsing /// /// #Returns /// - `i64`: The parsed value as i64 if successful, otherwise the default value /// pub fn get_env_i64(key: &str, default: i64) -> i64 { env::var(key).ok().and_then(|v| v.parse().ok()).unwrap_or(default) } /// Retrieve an environment variable as a specific type, returning None if not set or parsing fails. /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `Option<i64>`: The parsed value as i64 if successful, otherwise None /// pub fn get_env_opt_i64(key: &str) -> Option<i64> { env::var(key).ok().and_then(|v| v.parse().ok()) } /// Retrieve an environment variable as a specific type, returning Option<Option<i64>> if not set or parsing fails. /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `Option<Option<i64>>`: The parsed value as i64 if successful, otherwise None /// pub fn get_env_opt_opt_i64(key: &str) -> Option<Option<i64>> { env::var(key).ok().map(|v| v.parse().ok()) } /// Retrieve an environment variable as a specific type, with a default value if not set or parsing fails. /// /// #Parameters /// - `key`: The environment variable key to look up. /// - `default`: The default value to return if the environment variable is not set or parsing /// /// #Returns /// - `u64`: The parsed value as u64 if successful, otherwise the default value. /// pub fn get_env_u64(key: &str, default: u64) -> u64 { env::var(key).ok().and_then(|v| v.parse().ok()).unwrap_or(default) } /// Retrieve an environment variable as a specific type, returning None if not set or parsing fails. /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `Option<u64>`: The parsed value as u64 if successful, otherwise None /// pub fn get_env_opt_u64(key: &str) -> Option<u64> { env::var(key).ok().and_then(|v| v.parse().ok()) } /// Retrieve an environment variable as a specific type, with a default value if not set or parsing fails. /// /// #Parameters /// - `key`: The environment variable key to look up. /// - `default`: The default value to return if the environment variable is not set or parsing /// /// #Returns /// - `f64`: The parsed value as f64 if successful, otherwise the default value. /// pub fn get_env_f64(key: &str, default: f64) -> f64 { env::var(key).ok().and_then(|v| v.parse().ok()).unwrap_or(default) } /// Retrieve an environment variable as a specific type, returning None if not set or parsing fails. /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `Option<f64>`: The parsed value as f64 if successful, otherwise None /// pub fn get_env_opt_f64(key: &str) -> Option<f64> { env::var(key).ok().and_then(|v| v.parse().ok()) } /// Retrieve an environment variable as a specific type, with a default value if not set or parsing fails. /// /// #Parameters /// - `key`: The environment variable key to look up. /// - `default`: The default value to return if the environment variable is not set or parsing /// /// #Returns /// - `usize`: The parsed value as usize if successful, otherwise the default value. /// pub fn get_env_usize(key: &str, default: usize) -> usize { env::var(key).ok().and_then(|v| v.parse().ok()).unwrap_or(default) } /// Retrieve an environment variable as a specific type, returning None if not set or parsing fails. /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `Option<usize>`: The parsed value as usize if successful, otherwise None /// pub fn get_env_usize_opt(key: &str) -> Option<usize> { env::var(key).ok().and_then(|v| v.parse().ok()) } /// Retrieve an environment variable as a specific type, returning None if not set or parsing fails. /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `Option<usize>`: The parsed value as usize if successful, otherwise None /// pub fn get_env_opt_usize(key: &str) -> Option<usize> { get_env_usize_opt(key) } /// Retrieve an environment variable as a String, with a default value if not set. /// /// #Parameters /// - `key`: The environment variable key to look up. /// - `default`: The default string value to return if the environment variable is not set. /// /// #Returns /// - `String`: The environment variable value if set, otherwise the default value. /// pub fn get_env_str(key: &str, default: &str) -> String { env::var(key).unwrap_or_else(|_| default.to_string()) } /// Retrieve an environment variable as a String, returning None if not set. /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `Option<String>`: The environment variable value if set, otherwise None. /// pub fn get_env_opt_str(key: &str) -> Option<String> { env::var(key).ok() } /// Retrieve an environment variable as a boolean, with a default value if not set or parsing fails. /// /// #Parameters /// - `key`: The environment variable key to look up. /// - `default`: The default boolean value to return if the environment variable is not set or cannot be parsed. /// /// #Returns /// - `bool`: The parsed boolean value if successful, otherwise the default value. /// pub fn get_env_bool(key: &str, default: bool) -> bool { env::var(key) .ok() .and_then(|v| match v.to_lowercase().as_str() { "1" | "t" | "T" | "true" | "TRUE" | "True" | "on" | "ON" | "On" | "enabled" => Some(true), "0" | "f" | "F" | "false" | "FALSE" | "False" | "off" | "OFF" | "Off" | "disabled" => Some(false), _ => None, }) .unwrap_or(default) } /// Retrieve an environment variable as a boolean, returning None if not set or parsing fails. /// /// #Parameters /// - `key`: The environment variable key to look up. /// /// #Returns /// - `Option<bool>`: The parsed boolean value if successful, otherwise None. /// pub fn get_env_opt_bool(key: &str) -> Option<bool> { env::var(key).ok().and_then(|v| match v.to_lowercase().as_str() { "1" | "t" | "T" | "true" | "TRUE" | "True" | "on" | "ON" | "On" | "enabled" => Some(true), "0" | "f" | "F" | "false" | "FALSE" | "False" | "off" | "OFF" | "Off" | "disabled" => Some(false), _ => None, }) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/obj/mod.rs
crates/utils/src/obj/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. mod metadata; pub use metadata::*;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/obj/metadata.rs
crates/utils/src/obj/metadata.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use crate::http::{RESERVED_METADATA_PREFIX_LOWER, is_minio_header, is_rustfs_header}; use std::collections::HashMap; /// Extract user-defined metadata keys from object metadata. /// /// This function filters out system-level metadata and returns only user-defined keys. /// /// Excluded keys include: /// - S3 standard headers: content-type, cache-control, content-encoding, content-disposition, /// content-language, expires /// - x-amz-* headers (except user metadata with x-amz-meta- prefix which are stripped) /// - x-rustfs-internal-* headers (system internal metadata) /// - Storage/replication system keys: x-amz-storage-class, x-amz-tagging, x-amz-replication-status /// - Object metadata: etag, md5Sum, last-modified /// /// # Arguments /// * `metadata` - The complete metadata HashMap from ObjectInfo.user_defined /// /// # Returns /// A new HashMap containing only user-defined metadata entries. Keys that use /// the user-metadata prefix (for example `x-amz-meta-`) are returned with that /// prefix stripped. /// /// Note: The keys in the returned map may therefore differ from the keys in /// the input `metadata` map and cannot be used directly to remove entries /// from `metadata`. If you need to identify which original keys to remove, /// consider using an in-place filtering approach or returning the original /// keys instead. /// /// # Example /// ``` /// use std::collections::HashMap; /// use rustfs_utils::obj::extract_user_defined_metadata; /// /// let mut metadata = HashMap::new(); /// metadata.insert("content-type".to_string(), "application/json".to_string()); /// metadata.insert("x-minio-key".to_string(), "application/json".to_string()); /// metadata.insert("x-amz-grant-sse".to_string(), "application/json".to_string()); /// metadata.insert("x-amz-meta-user-key".to_string(), "user-value".to_string()); /// metadata.insert("my-custom-key".to_string(), "custom-value".to_string()); /// /// let user_keys = extract_user_defined_metadata(&metadata); /// assert_eq!(user_keys.len(), 2); /// assert_eq!(user_keys.get("user-key"), Some(&"user-value".to_string())); /// assert_eq!(user_keys.get("my-custom-key"), Some(&"custom-value".to_string())); /// ``` pub fn extract_user_defined_metadata(metadata: &HashMap<String, String>) -> HashMap<String, String> { let mut user_metadata = HashMap::new(); let system_headers = [ "content-type", "cache-control", "content-encoding", "content-disposition", "content-language", "expires", "content-length", "content-md5", "content-range", "last-modified", "etag", "md5sum", "date", ]; for (key, value) in metadata { let lower_key = key.to_ascii_lowercase(); if lower_key.starts_with(RESERVED_METADATA_PREFIX_LOWER) { continue; } if system_headers.contains(&lower_key.as_str()) { continue; } if let Some(user_key) = lower_key.strip_prefix("x-amz-meta-") { if !user_key.is_empty() { user_metadata.insert(user_key.to_string(), value.clone()); } continue; } // Check if it's x-rustfs-meta-* and extract user key if let Some(user_key) = lower_key.strip_prefix("x-rustfs-meta-") { if !user_key.is_empty() { user_metadata.insert(user_key.to_string(), value.clone()); } continue; } // Skip other x-amz-* headers if lower_key.starts_with("x-amz-") { continue; } // Skip other RustFS headers (x-rustfs-replication-*, etc.) if is_rustfs_header(key) { continue; } // Skip MinIO headers (compatibility) if is_minio_header(key) { continue; } // All other keys are considered user-defined user_metadata.insert(key.clone(), value.clone()); } user_metadata } #[cfg(test)] mod tests { use super::*; #[test] fn test_extract_user_defined_metadata_basic() { let mut metadata = HashMap::new(); metadata.insert("my-key".to_string(), "my-value".to_string()); metadata.insert("custom-header".to_string(), "custom-value".to_string()); let user_metadata = extract_user_defined_metadata(&metadata); assert_eq!(user_metadata.len(), 2); assert_eq!(user_metadata.get("my-key"), Some(&"my-value".to_string())); assert_eq!(user_metadata.get("custom-header"), Some(&"custom-value".to_string())); } #[test] fn test_extract_user_defined_metadata_exclude_system_headers() { let mut metadata = HashMap::new(); metadata.insert("content-type".to_string(), "application/json".to_string()); metadata.insert("cache-control".to_string(), "no-cache".to_string()); metadata.insert("content-encoding".to_string(), "gzip".to_string()); metadata.insert("content-disposition".to_string(), "attachment".to_string()); metadata.insert("content-language".to_string(), "en-US".to_string()); metadata.insert("expires".to_string(), "Wed, 21 Oct 2015 07:28:00 GMT".to_string()); metadata.insert("etag".to_string(), "abc123".to_string()); metadata.insert("last-modified".to_string(), "Tue, 20 Oct 2015 07:28:00 GMT".to_string()); metadata.insert("my-key".to_string(), "my-value".to_string()); let user_metadata = extract_user_defined_metadata(&metadata); assert_eq!(user_metadata.len(), 1); assert_eq!(user_metadata.get("my-key"), Some(&"my-value".to_string())); assert!(!user_metadata.contains_key("content-type")); assert!(!user_metadata.contains_key("cache-control")); assert!(!user_metadata.contains_key("etag")); } #[test] fn test_extract_user_defined_metadata_strip_amz_meta_prefix() { let mut metadata = HashMap::new(); metadata.insert("x-amz-meta-user-id".to_string(), "12345".to_string()); metadata.insert("x-amz-meta-project".to_string(), "test-project".to_string()); metadata.insert("x-amz-storage-class".to_string(), "STANDARD".to_string()); metadata.insert("x-amz-tagging".to_string(), "key=value".to_string()); metadata.insert("x-amz-replication-status".to_string(), "COMPLETED".to_string()); let user_metadata = extract_user_defined_metadata(&metadata); assert_eq!(user_metadata.len(), 2); assert_eq!(user_metadata.get("user-id"), Some(&"12345".to_string())); assert_eq!(user_metadata.get("project"), Some(&"test-project".to_string())); assert!(!user_metadata.contains_key("x-amz-meta-user-id")); assert!(!user_metadata.contains_key("x-amz-storage-class")); assert!(!user_metadata.contains_key("x-amz-tagging")); } #[test] fn test_extract_user_defined_metadata_exclude_rustfs_internal() { let mut metadata: HashMap<String, String> = HashMap::new(); metadata.insert("x-rustfs-internal-healing".to_string(), "true".to_string()); metadata.insert("x-rustfs-internal-data-mov".to_string(), "value".to_string()); metadata.insert("X-RustFS-Internal-purgestatus".to_string(), "status".to_string()); metadata.insert("x-rustfs-meta-custom".to_string(), "custom-value".to_string()); metadata.insert("my-key".to_string(), "my-value".to_string()); let user_metadata = extract_user_defined_metadata(&metadata); assert_eq!(user_metadata.len(), 2); assert_eq!(user_metadata.get("custom"), Some(&"custom-value".to_string())); assert_eq!(user_metadata.get("my-key"), Some(&"my-value".to_string())); assert!(!user_metadata.contains_key("x-rustfs-internal-healing")); assert!(!user_metadata.contains_key("x-rustfs-internal-data-mov")); } #[test] fn test_extract_user_defined_metadata_exclude_minio_headers() { let mut metadata = HashMap::new(); metadata.insert("x-minio-custom".to_string(), "minio-value".to_string()); metadata.insert("x-minio-internal".to_string(), "internal".to_string()); metadata.insert("my-key".to_string(), "my-value".to_string()); let user_metadata = extract_user_defined_metadata(&metadata); assert_eq!(user_metadata.len(), 1); assert_eq!(user_metadata.get("my-key"), Some(&"my-value".to_string())); assert!(!user_metadata.contains_key("x-minio-custom")); } #[test] fn test_extract_user_defined_metadata_mixed() { let mut metadata = HashMap::new(); // System headers metadata.insert("content-type".to_string(), "application/json".to_string()); metadata.insert("cache-control".to_string(), "no-cache".to_string()); // AMZ headers metadata.insert("x-amz-meta-version".to_string(), "1.0".to_string()); metadata.insert("x-amz-storage-class".to_string(), "STANDARD".to_string()); // RustFS internal metadata.insert("x-rustfs-internal-healing".to_string(), "true".to_string()); metadata.insert("x-rustfs-meta-source".to_string(), "upload".to_string()); // User defined metadata.insert("my-custom-key".to_string(), "custom-value".to_string()); metadata.insert("another-key".to_string(), "another-value".to_string()); let user_metadata = extract_user_defined_metadata(&metadata); assert_eq!(user_metadata.len(), 4); assert_eq!(user_metadata.get("version"), Some(&"1.0".to_string())); assert_eq!(user_metadata.get("source"), Some(&"upload".to_string())); assert_eq!(user_metadata.get("my-custom-key"), Some(&"custom-value".to_string())); assert_eq!(user_metadata.get("another-key"), Some(&"another-value".to_string())); assert!(!user_metadata.contains_key("content-type")); assert!(!user_metadata.contains_key("x-amz-storage-class")); assert!(!user_metadata.contains_key("x-rustfs-internal-healing")); } #[test] fn test_extract_user_defined_metadata_empty() { let metadata = HashMap::new(); let user_metadata = extract_user_defined_metadata(&metadata); assert!(user_metadata.is_empty()); } #[test] fn test_extract_user_defined_metadata_case_insensitive() { let mut metadata = HashMap::new(); metadata.insert("Content-Type".to_string(), "application/json".to_string()); metadata.insert("CACHE-CONTROL".to_string(), "no-cache".to_string()); metadata.insert("X-Amz-Meta-UserId".to_string(), "12345".to_string()); metadata.insert("My-Custom-Key".to_string(), "value".to_string()); let user_metadata = extract_user_defined_metadata(&metadata); assert_eq!(user_metadata.len(), 2); assert_eq!(user_metadata.get("userid"), Some(&"12345".to_string())); assert_eq!(user_metadata.get("My-Custom-Key"), Some(&"value".to_string())); assert!(!user_metadata.contains_key("Content-Type")); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/os/unix.rs
crates/utils/src/os/unix.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use super::{DiskInfo, IOStats}; use nix::sys::statfs::Statfs; use nix::sys::{stat::stat, statfs::statfs}; use std::io::Error; use std::path::Path; // FreeBSD and OpenBSD return a signed integer for blocks_available. // Cast to an unsigned integer to use with DiskInfo. #[cfg(any(target_os = "freebsd", target_os = "openbsd"))] fn blocks_available(stat: &Statfs) -> u64 { match stat.blocks_available().try_into() { Ok(bavail) => bavail, Err(e) => { tracing::warn!("blocks_available returned a negative value: Using 0 as fallback. {}", e); 0 } } } // FreeBSD returns a signed integer for files_free. Cast to an unsigned integer // to use with DiskInfo #[cfg(target_os = "freebsd")] fn files_free(stat: &Statfs) -> u64 { match stat.files_free().try_into() { Ok(files_free) => files_free, Err(e) => { tracing::warn!("files_free returned a negative value: Using 0 as fallback. {}", e); 0 } } } #[cfg(not(target_os = "freebsd"))] fn files_free(stat: &Statfs) -> u64 { stat.files_free() } #[cfg(not(any(target_os = "freebsd", target_os = "openbsd")))] fn blocks_available(stat: &Statfs) -> u64 { stat.blocks_available() } /// Returns total and free bytes available in a directory, e.g. `/`. pub fn get_info(p: impl AsRef<Path>) -> std::io::Result<DiskInfo> { let path_display = p.as_ref().display(); let stat = statfs(p.as_ref())?; let bsize = stat.block_size() as u64; let bfree = stat.blocks_free(); let bavail = blocks_available(&stat); let blocks = stat.blocks(); let reserved = match bfree.checked_sub(bavail) { Some(reserved) => reserved, None => { return Err(Error::other(format!( "detected f_bavail space ({bavail}) > f_bfree space ({bfree}), fs corruption at ({path_display}). please run 'fsck'", ))); } }; let total = match blocks.checked_sub(reserved) { Some(total) => total * bsize, None => { return Err(Error::other(format!( "detected reserved space ({reserved}) > blocks space ({blocks}), fs corruption at ({path_display}). please run 'fsck'", ))); } }; let free = bavail * bsize; let used = match total.checked_sub(free) { Some(used) => used, None => { return Err(Error::other(format!( "detected free space ({free}) > total drive space ({total}), fs corruption at ({path_display}). please run 'fsck'" ))); } }; Ok(DiskInfo { total, free, used, files: stat.files(), ffree: files_free(&stat), fstype: stat.filesystem_type_name().to_string(), ..Default::default() }) } pub fn same_disk(disk1: &str, disk2: &str) -> std::io::Result<bool> { let stat1 = stat(disk1)?; let stat2 = stat(disk2)?; Ok(stat1.st_dev == stat2.st_dev) } pub fn get_drive_stats(_major: u32, _minor: u32) -> std::io::Result<IOStats> { Ok(IOStats::default()) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/os/linux.rs
crates/utils/src/os/linux.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use nix::sys::stat::{self, stat}; use nix::sys::statfs::{self, FsType, statfs}; use std::fs::File; use std::io::{self, BufRead, Error, ErrorKind}; use std::path::Path; use super::{DiskInfo, IOStats}; /// Returns total and free bytes available in a directory, e.g. `/`. pub fn get_info(p: impl AsRef<Path>) -> std::io::Result<DiskInfo> { let path_display = p.as_ref().display(); let stat_fs = statfs(p.as_ref())?; let bsize = stat_fs.block_size() as u64; let bfree = stat_fs.blocks_free() as u64; let bavail = stat_fs.blocks_available() as u64; let blocks = stat_fs.blocks() as u64; let reserved = match bfree.checked_sub(bavail) { Some(reserved) => reserved, None => { return Err(Error::other(format!( "detected f_bavail space ({bavail}) > f_bfree space ({bfree}), fs corruption at ({path_display}). please run 'fsck'" ))); } }; let total = match blocks.checked_sub(reserved) { Some(total) => total * bsize, None => { return Err(Error::other(format!( "detected reserved space ({reserved}) > blocks space ({blocks}), fs corruption at ({path_display}). please run 'fsck'" ))); } }; let free = bavail * bsize; let used = match total.checked_sub(free) { Some(used) => used, None => { return Err(Error::other(format!( "detected free space ({free}) > total drive space ({total}), fs corruption at ({path_display}). please run 'fsck'" ))); } }; let st = stat(p.as_ref())?; Ok(DiskInfo { total, free, used, files: stat_fs.files(), ffree: stat_fs.files_free(), fstype: get_fs_type(stat_fs.filesystem_type()).to_string(), major: stat::major(st.st_dev), minor: stat::minor(st.st_dev), ..Default::default() }) } /// Returns the filesystem type of the underlying mounted filesystem /// /// TODO The following mapping could not find the corresponding constant in `nix`: /// /// "137d" => "EXT", /// "4244" => "HFS", /// "5346544e" => "NTFS", /// "61756673" => "AUFS", /// "ef51" => "EXT2OLD", /// "2fc12fc1" => "zfs", /// "ff534d42" => "cifs", /// "53464846" => "wslfs", fn get_fs_type(fs_type: FsType) -> &'static str { match fs_type { statfs::TMPFS_MAGIC => "TMPFS", statfs::MSDOS_SUPER_MAGIC => "MSDOS", // statfs::XFS_SUPER_MAGIC => "XFS", statfs::NFS_SUPER_MAGIC => "NFS", statfs::EXT4_SUPER_MAGIC => "EXT4", statfs::ECRYPTFS_SUPER_MAGIC => "ecryptfs", statfs::OVERLAYFS_SUPER_MAGIC => "overlayfs", statfs::REISERFS_SUPER_MAGIC => "REISERFS", _ => "UNKNOWN", } } pub fn same_disk(disk1: &str, disk2: &str) -> std::io::Result<bool> { let stat1 = stat(disk1)?; let stat2 = stat(disk2)?; Ok(stat1.st_dev == stat2.st_dev) } pub fn get_drive_stats(major: u32, minor: u32) -> std::io::Result<IOStats> { read_drive_stats(&format!("/sys/dev/block/{major}:{minor}/stat")) } fn read_drive_stats(stats_file: &str) -> std::io::Result<IOStats> { let stats = read_stat(stats_file)?; if stats.len() < 11 { return Err(Error::new( ErrorKind::InvalidData, format!("found invalid format while reading {stats_file}"), )); } let mut io_stats = IOStats { read_ios: stats[0], read_merges: stats[1], read_sectors: stats[2], read_ticks: stats[3], write_ios: stats[4], write_merges: stats[5], write_sectors: stats[6], write_ticks: stats[7], current_ios: stats[8], total_ticks: stats[9], req_ticks: stats[10], ..Default::default() }; if stats.len() > 14 { io_stats.discard_ios = stats[11]; io_stats.discard_merges = stats[12]; io_stats.discard_sectors = stats[13]; io_stats.discard_ticks = stats[14]; } Ok(io_stats) } fn read_stat(file_name: &str) -> std::io::Result<Vec<u64>> { // Open file let path = Path::new(file_name); let file = File::open(path)?; // Create a BufReader let reader = io::BufReader::new(file); // Read first line let mut stats = Vec::new(); if let Some(line) = reader.lines().next() { let line = line?; // Split line and parse as u64 // https://rust-lang.github.io/rust-clippy/master/index.html#trim_split_whitespace for token in line.split_whitespace() { let ui64: u64 = token .parse() .map_err(|e| Error::new(ErrorKind::InvalidData, format!("failed to parse '{token}' as u64: {e}")))?; stats.push(ui64); } } Ok(stats) } #[cfg(test)] mod test { use super::get_drive_stats; use tracing::debug; #[ignore] // FIXME: failed in github actions #[test] fn test_stats() { let major = 7; let minor = 11; let s = get_drive_stats(major, minor).unwrap(); debug!("Drive stats for major: {}, minor: {} - {:?}", major, minor, s); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/os/windows.rs
crates/utils/src/os/windows.rs
#![allow(unsafe_code)] // TODO: audit unsafe code // Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use super::{DiskInfo, IOStats}; use std::io::Error; use std::mem; use std::os::windows::ffi::OsStrExt; use std::path::Path; use winapi::shared::minwindef::{DWORD, MAX_PATH}; use winapi::shared::ntdef::ULARGE_INTEGER; use winapi::um::fileapi::{GetDiskFreeSpaceExW, GetDiskFreeSpaceW, GetVolumeInformationW, GetVolumePathNameW}; use winapi::um::winnt::{LPCWSTR, WCHAR}; /// Returns total and free bytes available in a directory, e.g. `C:\`. pub fn get_info(p: impl AsRef<Path>) -> std::io::Result<DiskInfo> { let path_display = p.as_ref().display(); let path_wide: Vec<WCHAR> = p .as_ref() .to_path_buf() .into_os_string() .encode_wide() .chain(std::iter::once(0)) // Null-terminate the string .collect(); let mut lp_free_bytes_available: ULARGE_INTEGER = unsafe { mem::zeroed() }; let mut lp_total_number_of_bytes: ULARGE_INTEGER = unsafe { mem::zeroed() }; let mut lp_total_number_of_free_bytes: ULARGE_INTEGER = unsafe { mem::zeroed() }; let success = unsafe { GetDiskFreeSpaceExW( path_wide.as_ptr(), &mut lp_free_bytes_available, &mut lp_total_number_of_bytes, &mut lp_total_number_of_free_bytes, ) }; if success == 0 { return Err(Error::last_os_error()); } let total = unsafe { *lp_total_number_of_bytes.QuadPart() }; let free = unsafe { *lp_total_number_of_free_bytes.QuadPart() }; if free > total { return Err(Error::other(format!( "detected free space ({free}) > total drive space ({total}), fs corruption at ({path_display}). please run 'fsck'" ))); } let mut lp_sectors_per_cluster: DWORD = 0; let mut lp_bytes_per_sector: DWORD = 0; let mut lp_number_of_free_clusters: DWORD = 0; let mut lp_total_number_of_clusters: DWORD = 0; let success = unsafe { GetDiskFreeSpaceW( path_wide.as_ptr(), &mut lp_sectors_per_cluster, &mut lp_bytes_per_sector, &mut lp_number_of_free_clusters, &mut lp_total_number_of_clusters, ) }; if success == 0 { return Err(Error::last_os_error()); } Ok(DiskInfo { total, free, used: total - free, files: lp_total_number_of_clusters as u64, ffree: lp_number_of_free_clusters as u64, // TODO This field is currently unused, and since this logic causes a // NotFound error during startup on Windows systems, it has been commented out here // // The error occurs in GetVolumeInformationW where the path parameter // is of type [WCHAR; MAX_PATH]. For a drive letter, there are excessive // trailing zeros, which causes the failure here. // // fstype: get_fs_type(&path_wide)?, ..Default::default() }) } /// Returns leading volume name. #[allow(dead_code)] fn get_volume_name(v: &[WCHAR]) -> std::io::Result<LPCWSTR> { let volume_name_size: DWORD = MAX_PATH as _; let mut lp_volume_name_buffer: [WCHAR; MAX_PATH] = [0; MAX_PATH]; let success = unsafe { GetVolumePathNameW(v.as_ptr(), lp_volume_name_buffer.as_mut_ptr(), volume_name_size) }; if success == 0 { return Err(Error::last_os_error()); } Ok(lp_volume_name_buffer.as_ptr()) } #[allow(dead_code)] fn utf16_to_string(v: &[WCHAR]) -> String { let len = v.iter().position(|&x| x == 0).unwrap_or(v.len()); String::from_utf16_lossy(&v[..len]) } /// Returns the filesystem type of the underlying mounted filesystem #[allow(dead_code)] fn get_fs_type(p: &[WCHAR]) -> std::io::Result<String> { let path = get_volume_name(p)?; let volume_name_size: DWORD = MAX_PATH as _; let n_file_system_name_size: DWORD = MAX_PATH as _; let mut lp_volume_serial_number: DWORD = 0; let mut lp_maximum_component_length: DWORD = 0; let mut lp_file_system_flags: DWORD = 0; let mut lp_volume_name_buffer: [WCHAR; MAX_PATH] = [0; MAX_PATH]; let mut lp_file_system_name_buffer: [WCHAR; MAX_PATH] = [0; MAX_PATH]; let success = unsafe { GetVolumeInformationW( path, lp_volume_name_buffer.as_mut_ptr(), volume_name_size, &mut lp_volume_serial_number, &mut lp_maximum_component_length, &mut lp_file_system_flags, lp_file_system_name_buffer.as_mut_ptr(), n_file_system_name_size, ) }; if success == 0 { return Err(Error::last_os_error()); } Ok(utf16_to_string(&lp_file_system_name_buffer)) } pub fn same_disk(_disk1: &str, _disk2: &str) -> std::io::Result<bool> { Ok(false) } pub fn get_drive_stats(_major: u32, _minor: u32) -> std::io::Result<IOStats> { Ok(IOStats::default()) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/os/mod.rs
crates/utils/src/os/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #[cfg(target_os = "linux")] mod linux; #[cfg(all(unix, not(target_os = "linux")))] mod unix; #[cfg(target_os = "windows")] mod windows; #[cfg(target_os = "linux")] pub use linux::{get_drive_stats, get_info, same_disk}; // pub use linux::same_disk; #[cfg(all(unix, not(target_os = "linux")))] pub use unix::{get_drive_stats, get_info, same_disk}; #[cfg(target_os = "windows")] pub use windows::{get_drive_stats, get_info, same_disk}; #[derive(Debug, Default, PartialEq)] pub struct IOStats { pub read_ios: u64, pub read_merges: u64, pub read_sectors: u64, pub read_ticks: u64, pub write_ios: u64, pub write_merges: u64, pub write_sectors: u64, pub write_ticks: u64, pub current_ios: u64, pub total_ticks: u64, pub req_ticks: u64, pub discard_ios: u64, pub discard_merges: u64, pub discard_sectors: u64, pub discard_ticks: u64, pub flush_ios: u64, pub flush_ticks: u64, } #[derive(Debug, Default, PartialEq)] pub struct DiskInfo { pub total: u64, pub free: u64, pub used: u64, pub files: u64, pub ffree: u64, pub fstype: String, pub major: u64, pub minor: u64, pub name: String, pub rotational: bool, pub nrrequests: u64, } #[cfg(test)] mod tests { use super::*; use std::path::PathBuf; #[test] fn test_get_info_valid_path() { let temp_dir = tempfile::tempdir().unwrap(); let info = get_info(temp_dir.path()).unwrap(); // Verify disk info is valid assert!(info.total > 0); assert!(info.free > 0); assert!(info.used > 0); assert!(info.files > 0); assert!(info.ffree > 0); assert!(!info.fstype.is_empty()); } #[test] fn test_get_info_invalid_path() { let invalid_path = PathBuf::from("/invalid/path"); let result = get_info(&invalid_path); assert!(result.is_err()); } #[test] fn test_same_disk_same_path() { let temp_dir = tempfile::tempdir().unwrap(); let path = temp_dir.path().to_str().unwrap(); let result = same_disk(path, path).unwrap(); assert!(result); } #[test] fn test_same_disk_different_paths() { let temp_dir1 = tempfile::tempdir().unwrap(); let temp_dir2 = tempfile::tempdir().unwrap(); let path1 = temp_dir1.path().to_str().unwrap(); let path2 = temp_dir2.path().to_str().unwrap(); let _result = same_disk(path1, path2).unwrap(); // Since both temporary directories are created in the same file system, // they should be on the same disk in most cases // Test passes if the function doesn't panic - the actual result depends on test environment } #[ignore] // FIXME: failed in github actions #[test] fn test_get_drive_stats_default() { let stats = get_drive_stats(0, 0).unwrap(); assert_eq!(stats, IOStats::default()); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/notify/mod.rs
crates/utils/src/notify/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. mod net; use hashbrown::HashMap; use hyper::HeaderMap; use s3s::{S3Request, S3Response}; pub use net::*; /// Extract request parameters from S3Request, mainly header information. #[allow(dead_code)] pub fn extract_req_params<T>(req: &S3Request<T>) -> HashMap<String, String> { let mut params = HashMap::new(); for (key, value) in req.headers.iter() { if let Ok(val_str) = value.to_str() { params.insert(key.as_str().to_string(), val_str.to_string()); } } params } /// Extract request parameters from hyper::HeaderMap, mainly header information. /// This function is useful when you have a raw HTTP request and need to extract parameters. pub fn extract_req_params_header(head: &HeaderMap) -> HashMap<String, String> { let mut params = HashMap::new(); for (key, value) in head.iter() { if let Ok(val_str) = value.to_str() { params.insert(key.as_str().to_string(), val_str.to_string()); } } params } /// Extract response elements from S3Response, mainly header information. #[allow(dead_code)] pub fn extract_resp_elements<T>(resp: &S3Response<T>) -> HashMap<String, String> { let mut params = HashMap::new(); for (key, value) in resp.headers.iter() { if let Ok(val_str) = value.to_str() { params.insert(key.as_str().to_string(), val_str.to_string()); } } params } /// Get host from header information. #[allow(dead_code)] pub fn get_request_host(headers: &HeaderMap) -> String { headers .get("host") .and_then(|v| v.to_str().ok()) .unwrap_or_default() .to_string() } /// Get user-agent from header information. #[allow(dead_code)] pub fn get_request_user_agent(headers: &HeaderMap) -> String { headers .get("user-agent") .and_then(|v| v.to_str().ok()) .unwrap_or_default() .to_string() }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/notify/net.rs
crates/utils/src/notify/net.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use regex::Regex; use serde::{Deserialize, Serialize}; use std::net::IpAddr; use std::path::Path; use std::sync::LazyLock; use thiserror::Error; use url::Url; // Lazy static for the host label regex. static HOST_LABEL_REGEX: LazyLock<Regex> = LazyLock::new(|| Regex::new(r"^[a-zA-Z0-9]([a-zA-Z0-9-]*[a-zA-Z0-9])?$").unwrap()); /// NetError represents errors that can occur in network operations. #[derive(Error, Debug)] pub enum NetError { #[error("invalid argument")] InvalidArgument, #[error("invalid hostname")] InvalidHost, #[error("missing '[' in host")] MissingBracket, #[error("parse error: {0}")] ParseError(String), #[error("unexpected scheme: {0}")] UnexpectedScheme(String), #[error("scheme appears with empty host")] SchemeWithEmptyHost, } /// Host represents a network host with IP/name and port. /// Similar to Go's net.Host structure. #[derive(Debug, Clone, PartialEq, Serialize, Deserialize)] pub struct Host { pub name: String, pub port: Option<u16>, // Using Option<u16> to represent if port is set, similar to IsPortSet. } // Implementation of Host methods. impl Host { // is_empty returns true if the host name is empty. pub fn is_empty(&self) -> bool { self.name.is_empty() } // equal checks if two hosts are equal by comparing their string representations. pub fn equal(&self, other: &Host) -> bool { self.to_string() == other.to_string() } } impl std::fmt::Display for Host { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { match self.port { Some(p) => write!(f, "{}:{}", self.name, p), None => write!(f, "{}", self.name), } } } // parse_host parses a string into a Host, with validation similar to Go's ParseHost. pub fn parse_host(s: &str) -> Result<Host, NetError> { if s.is_empty() { return Err(NetError::InvalidArgument); } // is_valid_host validates the host string, checking for IP or hostname validity. let is_valid_host = |host: &str| -> bool { if host.is_empty() { return true; } if host.parse::<IpAddr>().is_ok() { return true; } if !(1..=253).contains(&host.len()) { return false; } for (i, label) in host.split('.').enumerate() { if i + 1 == host.split('.').count() && label.is_empty() { continue; } if !(1..=63).contains(&label.len()) || !HOST_LABEL_REGEX.is_match(label) { return false; } } true }; // Split host and port, similar to net.SplitHostPort. let (host_str, port_str) = s.rsplit_once(':').map_or((s, ""), |(h, p)| (h, p)); let port = if !port_str.is_empty() { Some(port_str.parse().map_err(|_| NetError::ParseError(port_str.to_string()))?) } else { None }; // Trim IPv6 brackets if present. let host = trim_ipv6(host_str)?; // Handle IPv6 zone identifier. let trimmed_host = host.split('%').next().unwrap_or(&host); if !is_valid_host(trimmed_host) { return Err(NetError::InvalidHost); } Ok(Host { name: host, port }) } // trim_ipv6 removes square brackets from IPv6 addresses, similar to Go's trimIPv6. fn trim_ipv6(host: &str) -> Result<String, NetError> { if host.ends_with(']') { if !host.starts_with('[') { return Err(NetError::MissingBracket); } Ok(host[1..host.len() - 1].to_string()) } else { Ok(host.to_string()) } } /// URL is a wrapper around url::Url for custom handling. /// Provides methods similar to Go's URL struct. #[derive(Debug, Clone)] pub struct ParsedURL(pub Url); impl ParsedURL { /// is_empty returns true if the URL is empty or "about:blank". /// /// # Arguments /// * `&self` - Reference to the ParsedURL instance. /// /// # Returns /// * `bool` - True if the URL is empty or "about:blank", false otherwise. /// pub fn is_empty(&self) -> bool { self.0.as_str() == "" || (self.0.scheme() == "about" && self.0.path() == "blank") } /// hostname returns the hostname of the URL. /// /// # Returns /// * `String` - The hostname of the URL, or an empty string if not set. /// pub fn hostname(&self) -> String { self.0.host_str().unwrap_or("").to_string() } /// port returns the port of the URL as a string, defaulting to "80" for http and "443" for https if not set. /// /// # Returns /// * `String` - The port of the URL as a string. /// pub fn port(&self) -> String { match self.0.port() { Some(p) => p.to_string(), None => match self.0.scheme() { "http" => "80".to_string(), "https" => "443".to_string(), _ => "".to_string(), }, } } /// scheme returns the scheme of the URL. /// /// # Returns /// * `&str` - The scheme of the URL. /// pub fn scheme(&self) -> &str { self.0.scheme() } /// url returns a reference to the underlying Url. /// /// # Returns /// * `&Url` - Reference to the underlying Url. /// pub fn url(&self) -> &Url { &self.0 } } impl std::fmt::Display for ParsedURL { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { let mut url = self.0.clone(); if let Some(host) = url.host_str().map(|h| h.to_string()) && let Some(port) = url.port() && ((url.scheme() == "http" && port == 80) || (url.scheme() == "https" && port == 443)) { url.set_host(Some(&host)).unwrap(); url.set_port(None).unwrap(); } let mut s = url.to_string(); // If the URL ends with a slash and the path is just "/", remove the trailing slash. if s.ends_with('/') && url.path() == "/" { s.pop(); } write!(f, "{s}") } } impl serde::Serialize for ParsedURL { fn serialize<S>(&self, serializer: S) -> Result<S::Ok, S::Error> where S: serde::Serializer, { serializer.serialize_str(&self.to_string()) } } impl<'de> serde::Deserialize<'de> for ParsedURL { fn deserialize<D>(deserializer: D) -> Result<Self, D::Error> where D: serde::Deserializer<'de>, { let s: String = serde::Deserialize::deserialize(deserializer)?; if s.is_empty() { Ok(ParsedURL(Url::parse("about:blank").unwrap())) } else { parse_url(&s).map_err(serde::de::Error::custom) } } } /// parse_url parses a string into a ParsedURL, with host validation and path cleaning. /// /// # Arguments /// * `s` - The URL string to parse. /// /// # Returns /// * `Ok(ParsedURL)` - If parsing is successful. /// * `Err(NetError)` - If parsing fails or host is invalid. /// /// # Errors /// Returns NetError if parsing fails or host is invalid. /// pub fn parse_url(s: &str) -> Result<ParsedURL, NetError> { if let Some(scheme_end) = s.find("://") && s[scheme_end + 3..].starts_with('/') { let scheme = &s[..scheme_end]; if !scheme.is_empty() { return Err(NetError::SchemeWithEmptyHost); } } let mut uu = Url::parse(s).map_err(|e| NetError::ParseError(e.to_string()))?; if uu.host_str().is_none_or(|h| h.is_empty()) { if uu.scheme() != "" { return Err(NetError::SchemeWithEmptyHost); } } else { let port_str = uu.port().map(|p| p.to_string()).unwrap_or_else(|| match uu.scheme() { "http" => "80".to_string(), "https" => "443".to_string(), _ => "".to_string(), }); if !port_str.is_empty() { let host_port = format!("{}:{}", uu.host_str().unwrap(), port_str); parse_host(&host_port)?; // Validate host. } } // Clean path: Use Url's path_segments to normalize. if !uu.path().is_empty() { // Url automatically cleans paths, but we ensure trailing slash if original had it. let mut cleaned_path = String::new(); for comp in Path::new(uu.path()).components() { use std::path::Component; match comp { Component::RootDir => cleaned_path.push('/'), Component::Normal(s) => { if !cleaned_path.ends_with('/') { cleaned_path.push('/'); } cleaned_path.push_str(&s.to_string_lossy()); } _ => {} } } if s.ends_with('/') && !cleaned_path.ends_with('/') { cleaned_path.push('/'); } if cleaned_path.is_empty() { cleaned_path.push('/'); } uu.set_path(&cleaned_path); } Ok(ParsedURL(uu)) } #[allow(dead_code)] /// parse_http_url parses a string into a ParsedURL, ensuring the scheme is http or https. /// /// # Arguments /// * `s` - The URL string to parse. /// /// # Returns /// * `Ok(ParsedURL)` - If parsing is successful and scheme is http/https. /// * `Err(NetError)` - If parsing fails or scheme is not http/https. /// pub fn parse_http_url(s: &str) -> Result<ParsedURL, NetError> { let u = parse_url(s)?; match u.0.scheme() { "http" | "https" => Ok(u), _ => Err(NetError::UnexpectedScheme(u.0.scheme().to_string())), } } #[allow(dead_code)] /// is_network_or_host_down checks if an error indicates network or host down, considering timeouts. /// /// # Arguments /// * `err` - The std::io::Error to check. /// * `expect_timeouts` - Whether timeouts are expected. /// /// # Returns /// * `bool` - True if the error indicates network or host down, false otherwise. /// pub fn is_network_or_host_down(err: &std::io::Error, expect_timeouts: bool) -> bool { if err.kind() == std::io::ErrorKind::TimedOut { return !expect_timeouts; } // Simplified checks based on Go logic; adapt for Rust as needed let err_str = err.to_string().to_lowercase(); err_str.contains("connection reset by peer") || err_str.contains("connection timed out") || err_str.contains("broken pipe") || err_str.contains("use of closed network connection") } #[allow(dead_code)] /// is_conn_reset_err checks if an error indicates a connection reset by peer. /// /// # Arguments /// * `err` - The std::io::Error to check. /// /// # Returns /// * `bool` - True if the error indicates connection reset, false otherwise. /// pub fn is_conn_reset_err(err: &std::io::Error) -> bool { err.to_string().contains("connection reset by peer") || matches!(err.raw_os_error(), Some(libc::ECONNRESET)) } #[allow(dead_code)] /// is_conn_refused_err checks if an error indicates a connection refused. /// /// # Arguments /// * `err` - The std::io::Error to check. /// /// # Returns /// * `bool` - True if the error indicates connection refused, false otherwise. /// pub fn is_conn_refused_err(err: &std::io::Error) -> bool { err.to_string().contains("connection refused") || matches!(err.raw_os_error(), Some(libc::ECONNREFUSED)) } #[cfg(test)] mod tests { use super::*; #[test] fn parse_host_with_empty_string_returns_error() { let result = parse_host(""); assert!(matches!(result, Err(NetError::InvalidArgument))); } #[test] fn parse_host_with_valid_ipv4() { let result = parse_host("192.168.1.1:8080"); assert!(result.is_ok()); let host = result.unwrap(); assert_eq!(host.name, "192.168.1.1"); assert_eq!(host.port, Some(8080)); } #[test] fn parse_host_with_valid_hostname() { let result = parse_host("example.com:443"); assert!(result.is_ok()); let host = result.unwrap(); assert_eq!(host.name, "example.com"); assert_eq!(host.port, Some(443)); } #[test] fn parse_host_with_ipv6_brackets() { let result = parse_host("[::1]:8080"); assert!(result.is_ok()); let host = result.unwrap(); assert_eq!(host.name, "::1"); assert_eq!(host.port, Some(8080)); } #[test] fn parse_host_with_invalid_ipv6_missing_bracket() { let result = parse_host("::1]:8080"); assert!(matches!(result, Err(NetError::MissingBracket))); } #[test] fn parse_host_with_invalid_hostname() { let result = parse_host("invalid..host:80"); assert!(matches!(result, Err(NetError::InvalidHost))); } #[test] fn parse_host_without_port() { let result = parse_host("example.com"); assert!(result.is_ok()); let host = result.unwrap(); assert_eq!(host.name, "example.com"); assert_eq!(host.port, None); } #[test] fn host_is_empty_when_name_is_empty() { let host = Host { name: "".to_string(), port: None, }; assert!(host.is_empty()); } #[test] fn host_is_not_empty_when_name_present() { let host = Host { name: "example.com".to_string(), port: Some(80), }; assert!(!host.is_empty()); } #[test] fn host_to_string_with_port() { let host = Host { name: "example.com".to_string(), port: Some(80), }; assert_eq!(host.to_string(), "example.com:80"); } #[test] fn host_to_string_without_port() { let host = Host { name: "example.com".to_string(), port: None, }; assert_eq!(host.to_string(), "example.com"); } #[test] fn host_equal_when_same() { let host1 = Host { name: "example.com".to_string(), port: Some(80), }; let host2 = Host { name: "example.com".to_string(), port: Some(80), }; assert!(host1.equal(&host2)); } #[test] fn host_not_equal_when_different() { let host1 = Host { name: "example.com".to_string(), port: Some(80), }; let host2 = Host { name: "example.com".to_string(), port: Some(443), }; assert!(!host1.equal(&host2)); } #[test] fn parse_url_with_valid_http_url() { let result = parse_url("http://example.com/path"); assert!(result.is_ok()); let parsed = result.unwrap(); assert_eq!(parsed.hostname(), "example.com"); assert_eq!(parsed.port(), "80"); } #[test] fn parse_url_with_valid_https_url() { let result = parse_url("https://example.com:443/path"); assert!(result.is_ok()); let parsed = result.unwrap(); assert_eq!(parsed.hostname(), "example.com"); assert_eq!(parsed.port(), "443"); } #[test] fn parse_url_with_scheme_but_empty_host() { let result = parse_url("http:///path"); assert!(matches!(result, Err(NetError::SchemeWithEmptyHost))); } #[test] fn parse_url_with_invalid_host() { let result = parse_url("http://invalid..host/path"); assert!(matches!(result, Err(NetError::InvalidHost))); } #[test] fn parse_url_with_path_cleaning() { let result = parse_url("http://example.com//path/../path/"); assert!(result.is_ok()); let parsed = result.unwrap(); assert_eq!(parsed.0.path(), "/path/"); } #[test] fn parse_http_url_with_http_scheme() { let result = parse_http_url("http://example.com"); assert!(result.is_ok()); } #[test] fn parse_http_url_with_https_scheme() { let result = parse_http_url("https://example.com"); assert!(result.is_ok()); } #[test] fn parse_http_url_with_invalid_scheme() { let result = parse_http_url("ftp://example.com"); assert!(matches!(result, Err(NetError::UnexpectedScheme(_)))); } #[test] fn parsed_url_is_empty_when_url_is_empty() { let url = ParsedURL(Url::parse("about:blank").unwrap()); assert!(url.is_empty()); } #[test] fn parsed_url_hostname() { let url = ParsedURL(Url::parse("http://example.com:8080").unwrap()); assert_eq!(url.hostname(), "example.com"); } #[test] fn parsed_url_port() { let url = ParsedURL(Url::parse("http://example.com:8080").unwrap()); assert_eq!(url.port(), "8080"); } #[test] fn parsed_url_to_string_removes_default_ports() { let url = ParsedURL(Url::parse("http://example.com:80").unwrap()); assert_eq!(url.to_string(), "http://example.com"); } #[test] fn is_network_or_host_down_with_timeout() { let err = std::io::Error::new(std::io::ErrorKind::TimedOut, "timeout"); assert!(is_network_or_host_down(&err, false)); } #[test] fn is_network_or_host_down_with_expected_timeout() { let err = std::io::Error::new(std::io::ErrorKind::TimedOut, "timeout"); assert!(!is_network_or_host_down(&err, true)); } #[test] fn is_conn_reset_err_with_reset_message() { let err = std::io::Error::other("connection reset by peer"); assert!(is_conn_reset_err(&err)); } #[test] fn is_conn_refused_err_with_refused_message() { let err = std::io::Error::other("connection refused"); assert!(is_conn_refused_err(&err)); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/sys/user_agent.rs
crates/utils/src/sys/user_agent.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use rustfs_config::VERSION; use std::env; use std::fmt; use sysinfo::System; /// Business Type Enumeration #[derive(Debug, Clone, PartialEq)] pub enum ServiceType { Basis, Core, Event, Logger, Custom(String), } impl ServiceType { fn as_str(&self) -> &str { match self { ServiceType::Basis => "basis", ServiceType::Core => "core", ServiceType::Event => "event", ServiceType::Logger => "logger", ServiceType::Custom(s) => s.as_str(), } } } /// UserAgent structure to hold User-Agent information /// including OS platform, architecture, version, and service type. /// It provides methods to generate a formatted User-Agent string. /// # Examples /// ``` /// use rustfs_utils::{get_user_agent, ServiceType}; /// /// let ua = get_user_agent(ServiceType::Core); /// println!("User-Agent: {}", ua); /// ``` #[derive(Debug)] struct UserAgent { os_platform: String, arch: String, version: String, service: ServiceType, } impl UserAgent { /// Create a new UserAgent instance and accept business type parameters /// /// # Arguments /// * `service` - The type of service for which the User-Agent is being created. /// # Returns /// A new instance of `UserAgent` with the current OS platform, architecture, version, and service type. fn new(service: ServiceType) -> Self { let os_platform = Self::get_os_platform(); let arch = env::consts::ARCH.to_string(); let version = VERSION.to_string(); UserAgent { os_platform, arch, version, service, } } /// Obtain operating system platform information fn get_os_platform() -> String { if cfg!(target_os = "windows") { Self::get_windows_platform() } else if cfg!(target_os = "macos") { Self::get_macos_platform() } else if cfg!(target_os = "linux") { Self::get_linux_platform() } else { "Unknown".to_string() } } /// Get Windows platform information #[cfg(windows)] fn get_windows_platform() -> String { // Priority to using sysinfo to get versions let version = match System::os_version() { Some(version) => version, None => "Windows NT Unknown".to_string(), }; format!("Windows NT {version}") } #[cfg(not(windows))] fn get_windows_platform() -> String { "N/A".to_string() } /// Get macOS platform information #[cfg(target_os = "macos")] fn get_macos_platform() -> String { let binding = System::os_version().unwrap_or("14.5.0".to_string()); let version = binding.split('.').collect::<Vec<&str>>(); let major = version.first().unwrap_or(&"14").to_string(); let minor = version.get(1).unwrap_or(&"5").to_string(); let patch = version.get(2).unwrap_or(&"0").to_string(); let arch = env::consts::ARCH; let cpu_info = if arch == "aarch64" { "Apple" } else { "Intel" }; // Convert to User-Agent format format!("Macintosh; {cpu_info} Mac OS X {major}_{minor}_{patch}") } #[cfg(not(target_os = "macos"))] fn get_macos_platform() -> String { "N/A".to_string() } /// Get Linux platform information #[cfg(target_os = "linux")] fn get_linux_platform() -> String { format!("X11; {}", System::long_os_version().unwrap_or("Linux Unknown".to_string())) } #[cfg(not(target_os = "linux"))] fn get_linux_platform() -> String { "N/A".to_string() } } /// Implement Display trait to format User-Agent impl fmt::Display for UserAgent { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { if self.service == ServiceType::Basis { return write!(f, "Mozilla/5.0 ({}; {}) RustFS/{}", self.os_platform, self.arch, self.version); } write!( f, "Mozilla/5.0 ({}; {}) RustFS/{} ({})", self.os_platform, self.arch, self.version, self.service.as_str() ) } } /// Get the User-Agent string and accept business type parameters /// /// # Arguments /// * `service` - The type of service for which the User-Agent is being created. /// /// # Returns /// A formatted User-Agent string. /// pub fn get_user_agent(service: ServiceType) -> String { UserAgent::new(service).to_string() } #[cfg(test)] mod tests { use super::*; use rustfs_config::VERSION; use tracing::debug; #[test] fn test_user_agent_format_basis() { let ua = get_user_agent(ServiceType::Basis); assert!(ua.starts_with("Mozilla/5.0")); assert!(ua.contains(&format!("RustFS/{VERSION}").to_string())); debug!("Basic User-Agent: {}", ua); } #[test] fn test_user_agent_format_core() { let ua = get_user_agent(ServiceType::Core); assert!(ua.starts_with("Mozilla/5.0")); assert!(ua.contains(&format!("RustFS/{VERSION} (core)").to_string())); debug!("Core User-Agent: {}", ua); } #[test] fn test_user_agent_format_event() { let ua = get_user_agent(ServiceType::Event); assert!(ua.starts_with("Mozilla/5.0")); assert!(ua.contains(&format!("RustFS/{VERSION} (event)").to_string())); debug!("Event User-Agent: {}", ua); } #[test] fn test_user_agent_format_logger() { let ua = get_user_agent(ServiceType::Logger); assert!(ua.starts_with("Mozilla/5.0")); assert!(ua.contains(&format!("RustFS/{VERSION} (logger)").to_string())); debug!("Logger User-Agent: {}", ua); } #[test] fn test_user_agent_format_custom() { let ua = get_user_agent(ServiceType::Custom("monitor".to_string())); assert!(ua.starts_with("Mozilla/5.0")); assert!(ua.contains(&format!("RustFS/{VERSION} (monitor)").to_string())); debug!("Monitor User-Agent: {}", ua); } #[test] fn test_all_service_type() { // Example: Generate User-Agents of Different Business Types let ua_core = get_user_agent(ServiceType::Core); let ua_event = get_user_agent(ServiceType::Event); let ua_logger = get_user_agent(ServiceType::Logger); let ua_custom = get_user_agent(ServiceType::Custom("monitor".to_string())); debug!("Core User-Agent: {}", ua_core); debug!("Event User-Agent: {}", ua_event); debug!("Logger User-Agent: {}", ua_logger); debug!("Custom User-Agent: {}", ua_custom); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/sys/mod.rs
crates/utils/src/sys/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. pub(crate) mod user_agent;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/http/headers.rs
crates/utils/src/http/headers.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use convert_case::{Case, Casing}; use std::collections::HashMap; use std::sync::LazyLock; pub const LAST_MODIFIED: &str = "Last-Modified"; pub const DATE: &str = "Date"; pub const ETAG: &str = "ETag"; pub const CONTENT_TYPE: &str = "Content-Type"; pub const CONTENT_MD5: &str = "Content-Md5"; pub const CONTENT_ENCODING: &str = "Content-Encoding"; pub const EXPIRES: &str = "Expires"; pub const CONTENT_LENGTH: &str = "Content-Length"; pub const CONTENT_LANGUAGE: &str = "Content-Language"; pub const CONTENT_RANGE: &str = "Content-Range"; pub const CONNECTION: &str = "Connection"; pub const ACCEPT_RANGES: &str = "Accept-Ranges"; pub const AMZ_BUCKET_REGION: &str = "X-Amz-Bucket-Region"; pub const SERVER_INFO: &str = "Server"; pub const RETRY_AFTER: &str = "Retry-After"; pub const LOCATION: &str = "Location"; pub const CACHE_CONTROL: &str = "Cache-Control"; pub const CONTENT_DISPOSITION: &str = "Content-Disposition"; pub const AUTHORIZATION: &str = "Authorization"; pub const ACTION: &str = "Action"; pub const RANGE: &str = "Range"; // S3 storage class pub const AMZ_STORAGE_CLASS: &str = "x-amz-storage-class"; // S3 object version ID pub const AMZ_VERSION_ID: &str = "x-amz-version-id"; pub const AMZ_DELETE_MARKER: &str = "x-amz-delete-marker"; // S3 object tagging pub const AMZ_OBJECT_TAGGING: &str = "X-Amz-Tagging"; pub const AMZ_TAG_COUNT: &str = "x-amz-tagging-count"; pub const AMZ_TAG_DIRECTIVE: &str = "X-Amz-Tagging-Directive"; // S3 transition restore pub const AMZ_RESTORE_EXPIRY_DAYS: &str = "X-Amz-Restore-Expiry-Days"; pub const AMZ_RESTORE_REQUEST_DATE: &str = "X-Amz-Restore-Request-Date"; // S3 extensions pub const AMZ_COPY_SOURCE_IF_MODIFIED_SINCE: &str = "x-amz-copy-source-if-modified-since"; pub const AMZ_COPY_SOURCE_IF_UNMODIFIED_SINCE: &str = "x-amz-copy-source-if-unmodified-since"; pub const AMZ_COPY_SOURCE_IF_NONE_MATCH: &str = "x-amz-copy-source-if-none-match"; pub const AMZ_COPY_SOURCE_IF_MATCH: &str = "x-amz-copy-source-if-match"; pub const AMZ_COPY_SOURCE: &str = "X-Amz-Copy-Source"; pub const AMZ_COPY_SOURCE_VERSION_ID: &str = "X-Amz-Copy-Source-Version-Id"; pub const AMZ_COPY_SOURCE_RANGE: &str = "X-Amz-Copy-Source-Range"; pub const AMZ_METADATA_DIRECTIVE: &str = "X-Amz-Metadata-Directive"; pub const AMZ_OBJECT_LOCK_MODE: &str = "X-Amz-Object-Lock-Mode"; pub const AMZ_OBJECT_LOCK_RETAIN_UNTIL_DATE: &str = "X-Amz-Object-Lock-Retain-Until-Date"; pub const AMZ_OBJECT_LOCK_LEGAL_HOLD: &str = "X-Amz-Object-Lock-Legal-Hold"; pub const AMZ_OBJECT_LOCK_BYPASS_GOVERNANCE: &str = "X-Amz-Bypass-Governance-Retention"; pub const AMZ_BUCKET_REPLICATION_STATUS: &str = "X-Amz-Replication-Status"; // AmzSnowballExtract will trigger unpacking of an archive content pub const AMZ_SNOWBALL_EXTRACT: &str = "X-Amz-Meta-Snowball-Auto-Extract"; // Object lock enabled pub const AMZ_OBJECT_LOCK_ENABLED: &str = "x-amz-bucket-object-lock-enabled"; // Multipart parts count pub const AMZ_MP_PARTS_COUNT: &str = "x-amz-mp-parts-count"; // Object date/time of expiration pub const AMZ_EXPIRATION: &str = "x-amz-expiration"; // Dummy putBucketACL pub const AMZ_ACL: &str = "x-amz-acl"; // Signature V4 related constants. pub const AMZ_CONTENT_SHA256: &str = "X-Amz-Content-Sha256"; pub const AMZ_DATE: &str = "X-Amz-Date"; pub const AMZ_ALGORITHM: &str = "X-Amz-Algorithm"; pub const AMZ_EXPIRES: &str = "X-Amz-Expires"; pub const AMZ_SIGNED_HEADERS: &str = "X-Amz-SignedHeaders"; pub const AMZ_SIGNATURE: &str = "X-Amz-Signature"; pub const AMZ_CREDENTIAL: &str = "X-Amz-Credential"; pub const AMZ_SECURITY_TOKEN: &str = "X-Amz-Security-Token"; pub const AMZ_DECODED_CONTENT_LENGTH: &str = "X-Amz-Decoded-Content-Length"; pub const AMZ_TRAILER: &str = "X-Amz-Trailer"; pub const AMZ_MAX_PARTS: &str = "X-Amz-Max-Parts"; pub const AMZ_PART_NUMBER_MARKER: &str = "X-Amz-Part-Number-Marker"; // Constants used for GetObjectAttributes and GetObjectVersionAttributes pub const AMZ_OBJECT_ATTRIBUTES: &str = "X-Amz-Object-Attributes"; // AWS server-side encryption headers for SSE-S3, SSE-KMS and SSE-C. pub const AMZ_SERVER_SIDE_ENCRYPTION: &str = "X-Amz-Server-Side-Encryption"; pub const AMZ_SERVER_SIDE_ENCRYPTION_KMS_ID: &str = "X-Amz-Server-Side-Encryption-Aws-Kms-Key-Id"; pub const AMZ_SERVER_SIDE_ENCRYPTION_KMS_CONTEXT: &str = "X-Amz-Server-Side-Encryption-Context"; pub const AMZ_SERVER_SIDE_ENCRYPTION_CUSTOMER_ALGORITHM: &str = "X-Amz-Server-Side-Encryption-Customer-Algorithm"; pub const AMZ_SERVER_SIDE_ENCRYPTION_CUSTOMER_KEY: &str = "X-Amz-Server-Side-Encryption-Customer-Key"; pub const AMZ_SERVER_SIDE_ENCRYPTION_CUSTOMER_KEY_MD5: &str = "X-Amz-Server-Side-Encryption-Customer-Key-Md5"; pub const AMZ_SERVER_SIDE_ENCRYPTION_COPY_CUSTOMER_ALGORITHM: &str = "X-Amz-Copy-Source-Server-Side-Encryption-Customer-Algorithm"; pub const AMZ_SERVER_SIDE_ENCRYPTION_COPY_CUSTOMER_KEY: &str = "X-Amz-Copy-Source-Server-Side-Encryption-Customer-Key"; pub const AMZ_SERVER_SIDE_ENCRYPTION_COPY_CUSTOMER_KEY_MD5: &str = "X-Amz-Copy-Source-Server-Side-Encryption-Customer-Key-Md5"; pub const AMZ_ENCRYPTION_AES: &str = "AES256"; pub const AMZ_ENCRYPTION_KMS: &str = "aws:kms"; // Signature v2 related constants pub const AMZ_SIGNATURE_V2: &str = "Signature"; pub const AMZ_ACCESS_KEY_ID: &str = "AWSAccessKeyId"; // Response request id. pub const AMZ_REQUEST_ID: &str = "x-amz-request-id"; pub const AMZ_REQUEST_HOST_ID: &str = "x-amz-id-2"; // Content Checksums pub const AMZ_CHECKSUM_ALGO: &str = "x-amz-checksum-algorithm"; pub const AMZ_CHECKSUM_CRC32: &str = "x-amz-checksum-crc32"; pub const AMZ_CHECKSUM_CRC32C: &str = "x-amz-checksum-crc32c"; pub const AMZ_CHECKSUM_SHA1: &str = "x-amz-checksum-sha1"; pub const AMZ_CHECKSUM_SHA256: &str = "x-amz-checksum-sha256"; pub const AMZ_CHECKSUM_CRC64NVME: &str = "x-amz-checksum-crc64nvme"; pub const AMZ_CHECKSUM_MODE: &str = "x-amz-checksum-mode"; pub const AMZ_CHECKSUM_TYPE: &str = "x-amz-checksum-type"; pub const AMZ_CHECKSUM_TYPE_FULL_OBJECT: &str = "FULL_OBJECT"; pub const AMZ_CHECKSUM_TYPE_COMPOSITE: &str = "COMPOSITE"; // Post Policy related pub const AMZ_META_UUID: &str = "X-Amz-Meta-Uuid"; pub const AMZ_META_NAME: &str = "X-Amz-Meta-Name"; pub const AMZ_META_UNENCRYPTED_CONTENT_LENGTH: &str = "X-Amz-Meta-X-Amz-Unencrypted-Content-Length"; pub const AMZ_META_UNENCRYPTED_CONTENT_MD5: &str = "X-Amz-Meta-X-Amz-Unencrypted-Content-Md5"; pub const RESERVED_METADATA_PREFIX: &str = "X-RustFS-Internal-"; pub const RESERVED_METADATA_PREFIX_LOWER: &str = "x-rustfs-internal-"; pub const RUSTFS_HEALING: &str = "X-Rustfs-Internal-healing"; // pub const RUSTFS_DATA_MOVE: &str = "X-Rustfs-Internal-data-mov"; // pub const X_RUSTFS_INLINE_DATA: &str = "x-rustfs-inline-data"; pub const VERSION_PURGE_STATUS_KEY: &str = "X-Rustfs-Internal-purgestatus"; pub const X_RUSTFS_HEALING: &str = "X-Rustfs-Internal-healing"; pub const X_RUSTFS_DATA_MOV: &str = "X-Rustfs-Internal-data-mov"; pub const AMZ_TAGGING_DIRECTIVE: &str = "X-Amz-Tagging-Directive"; pub const RUSTFS_DATA_MOVE: &str = "X-Rustfs-Internal-data-mov"; pub const RUSTFS_FORCE_DELETE: &str = "X-Rustfs-Force-Delete"; pub const RUSTFS_INCLUDE_DELETED: &str = "X-Rustfs-Include-Deleted"; pub const RUSTFS_REPLICATION_RESET_STATUS: &str = "X-Rustfs-Replication-Reset-Status"; pub const RUSTFS_REPLICATION_ACTUAL_OBJECT_SIZE: &str = "X-Rustfs-Replication-Actual-Object-Size"; pub const RUSTFS_BUCKET_SOURCE_VERSION_ID: &str = "X-Rustfs-Source-Version-Id"; pub const RUSTFS_BUCKET_SOURCE_MTIME: &str = "X-Rustfs-Source-Mtime"; pub const RUSTFS_BUCKET_SOURCE_ETAG: &str = "X-Rustfs-Source-Etag"; pub const RUSTFS_BUCKET_REPLICATION_DELETE_MARKER: &str = "X-Rustfs-Source-DeleteMarker"; pub const RUSTFS_BUCKET_REPLICATION_PROXY_REQUEST: &str = "X-Rustfs-Source-Proxy-Request"; pub const RUSTFS_BUCKET_REPLICATION_REQUEST: &str = "X-Rustfs-Source-Replication-Request"; pub const RUSTFS_BUCKET_REPLICATION_CHECK: &str = "X-Rustfs-Source-Replication-Check"; pub const RUSTFS_BUCKET_REPLICATION_SSEC_CHECKSUM: &str = "X-Rustfs-Source-Replication-Ssec-Crc"; // SSEC encryption header constants pub const SSEC_ALGORITHM_HEADER: &str = "x-amz-server-side-encryption-customer-algorithm"; pub const SSEC_KEY_HEADER: &str = "x-amz-server-side-encryption-customer-key"; pub const SSEC_KEY_MD5_HEADER: &str = "x-amz-server-side-encryption-customer-key-md5"; pub const AMZ_WEBSITE_REDIRECT_LOCATION: &str = "x-amz-website-redirect-location"; pub trait HeaderExt { fn lookup(&self, s: &str) -> Option<&str>; } impl HeaderExt for HashMap<String, String> { fn lookup(&self, s: &str) -> Option<&str> { let train = s.to_case(Case::Train); let lower = s.to_ascii_lowercase(); let keys = [s, lower.as_str(), train.as_str()]; for key in keys { if let Some(v) = self.get(key) { return Some(v); } } None } } static SUPPORTED_QUERY_VALUES: LazyLock<HashMap<String, bool>> = LazyLock::new(|| { let mut m = HashMap::new(); m.insert("attributes".to_string(), true); m.insert("partNumber".to_string(), true); m.insert("versionId".to_string(), true); m.insert("response-cache-control".to_string(), true); m.insert("response-content-disposition".to_string(), true); m.insert("response-content-encoding".to_string(), true); m.insert("response-content-language".to_string(), true); m.insert("response-content-type".to_string(), true); m.insert("response-expires".to_string(), true); m }); static SUPPORTED_HEADERS: LazyLock<HashMap<String, bool>> = LazyLock::new(|| { let mut m = HashMap::new(); m.insert("content-type".to_string(), true); m.insert("cache-control".to_string(), true); m.insert("content-encoding".to_string(), true); m.insert("content-disposition".to_string(), true); m.insert("content-language".to_string(), true); m.insert("x-amz-website-redirect-location".to_string(), true); m.insert("x-amz-object-lock-mode".to_string(), true); m.insert("x-amz-metadata-directive".to_string(), true); m.insert("x-amz-object-lock-retain-until-date".to_string(), true); m.insert("expires".to_string(), true); m.insert("x-amz-replication-status".to_string(), true); m }); static SSE_HEADERS: LazyLock<HashMap<String, bool>> = LazyLock::new(|| { let mut m = HashMap::new(); m.insert("x-amz-server-side-encryption".to_string(), true); m.insert("x-amz-server-side-encryption-aws-kms-key-id".to_string(), true); m.insert("x-amz-server-side-encryption-context".to_string(), true); m.insert("x-amz-server-side-encryption-customer-algorithm".to_string(), true); m.insert("x-amz-server-side-encryption-customer-key".to_string(), true); m.insert("x-amz-server-side-encryption-customer-key-md5".to_string(), true); m }); pub fn is_standard_query_value(qs_key: &str) -> bool { *SUPPORTED_QUERY_VALUES.get(qs_key).unwrap_or(&false) } pub fn is_storageclass_header(header_key: &str) -> bool { header_key.to_lowercase() == AMZ_STORAGE_CLASS.to_lowercase() } pub fn is_standard_header(header_key: &str) -> bool { *SUPPORTED_HEADERS.get(&header_key.to_lowercase()).unwrap_or(&false) } pub fn is_sse_header(header_key: &str) -> bool { *SSE_HEADERS.get(&header_key.to_lowercase()).unwrap_or(&false) } pub fn is_amz_header(header_key: &str) -> bool { let key = header_key.to_lowercase(); key.starts_with("x-amz-meta-") || key.starts_with("x-amz-grant-") || key == "x-amz-acl" || is_sse_header(header_key) || key.starts_with("x-amz-checksum-") } pub fn is_rustfs_header(header_key: &str) -> bool { header_key.to_lowercase().starts_with("x-rustfs-") } pub fn is_minio_header(header_key: &str) -> bool { header_key.to_lowercase().starts_with("x-minio-") }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/http/mod.rs
crates/utils/src/http/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. pub mod headers; pub mod ip; pub use headers::*; pub use ip::*;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/utils/src/http/ip.rs
crates/utils/src/http/ip.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use http::HeaderMap; use regex::Regex; use std::env; use std::net::SocketAddr; use std::str::FromStr; use std::sync::LazyLock; /// De-facto standard header keys. const X_FORWARDED_FOR: &str = "x-forwarded-for"; const X_FORWARDED_PROTO: &str = "x-forwarded-proto"; const X_FORWARDED_SCHEME: &str = "x-forwarded-scheme"; const X_REAL_IP: &str = "x-real-ip"; /// RFC7239 defines a new "Forwarded: " header designed to replace the /// existing use of X-Forwarded-* headers. /// e.g. Forwarded: for=192.0.2.60;proto=https;by=203.0.113.43 const FORWARDED: &str = "forwarded"; static FOR_REGEX: LazyLock<Regex> = LazyLock::new(|| Regex::new(r"(?i)(?:for=)([^(;|,| )]+)(.*)").unwrap()); static PROTO_REGEX: LazyLock<Regex> = LazyLock::new(|| Regex::new(r"(?i)^(;|,| )+(?:proto=)(https|http)").unwrap()); /// Used to disable all processing of the X-Forwarded-For header in source IP discovery. /// /// # Returns /// A `bool` indicating whether the X-Forwarded-For header is enabled /// fn is_xff_header_enabled() -> bool { env::var("_RUSTFS_API_XFF_HEADER") .unwrap_or_else(|_| "on".to_string()) .to_lowercase() == "on" } /// GetSourceScheme retrieves the scheme from the X-Forwarded-Proto and RFC7239 /// Forwarded headers (in that order). /// /// # Arguments /// * `headers` - HTTP headers from the request /// /// # Returns /// An `Option<String>` containing the source scheme if found /// pub fn get_source_scheme(headers: &HeaderMap) -> Option<String> { // Retrieve the scheme from X-Forwarded-Proto. if let Some(proto) = headers.get(X_FORWARDED_PROTO) && let Ok(proto_str) = proto.to_str() { return Some(proto_str.to_lowercase()); } if let Some(proto) = headers.get(X_FORWARDED_SCHEME) && let Ok(proto_str) = proto.to_str() { return Some(proto_str.to_lowercase()); } if let Some(forwarded) = headers.get(FORWARDED) && let Ok(forwarded_str) = forwarded.to_str() { // match should contain at least two elements if the protocol was // specified in the Forwarded header. The first element will always be // the 'for=', which we ignore, subsequently we proceed to look for // 'proto=' which should precede right after `for=` if not // we simply ignore the values and return empty. This is in line // with the approach we took for returning first ip from multiple // params. if let Some(for_match) = FOR_REGEX.captures(forwarded_str) && for_match.len() > 1 { let remaining = &for_match[2]; if let Some(proto_match) = PROTO_REGEX.captures(remaining) && proto_match.len() > 1 { return Some(proto_match[2].to_lowercase()); } } } None } /// GetSourceIPFromHeaders retrieves the IP from the X-Forwarded-For, X-Real-IP /// and RFC7239 Forwarded headers (in that order) /// /// # Arguments /// * `headers` - HTTP headers from the request /// /// # Returns /// An `Option<String>` containing the source IP address if found /// pub fn get_source_ip_from_headers(headers: &HeaderMap) -> Option<String> { let mut addr = None; if is_xff_header_enabled() && let Some(forwarded_for) = headers.get(X_FORWARDED_FOR) && let Ok(forwarded_str) = forwarded_for.to_str() { // Only grab the first (client) address. Note that '192.168.0.1, // 10.1.1.1' is a valid key for X-Forwarded-For where addresses after // the first may represent forwarding proxies earlier in the chain. let first_comma = forwarded_str.find(", "); let end = first_comma.unwrap_or(forwarded_str.len()); addr = Some(forwarded_str[..end].to_string()); } if addr.is_none() { if let Some(real_ip) = headers.get(X_REAL_IP) { if let Ok(real_ip_str) = real_ip.to_str() { // X-Real-IP should only contain one IP address (the client making the // request). addr = Some(real_ip_str.to_string()); } } else if let Some(forwarded) = headers.get(FORWARDED) && let Ok(forwarded_str) = forwarded.to_str() { // match should contain at least two elements if the protocol was // specified in the Forwarded header. The first element will always be // the 'for=' capture, which we ignore. In the case of multiple IP // addresses (for=8.8.8.8, 8.8.4.4, 172.16.1.20 is valid) we only // extract the first, which should be the client IP. if let Some(for_match) = FOR_REGEX.captures(forwarded_str) && for_match.len() > 1 { // IPv6 addresses in Forwarded headers are quoted-strings. We strip // these quotes. let ip = for_match[1].trim_matches('"'); addr = Some(ip.to_string()); } } } addr } /// GetSourceIPRaw retrieves the IP from the request headers /// and falls back to remote_addr when necessary. /// however returns without bracketing. /// /// # Arguments /// * `headers` - HTTP headers from the request /// * `remote_addr` - Remote address as a string /// /// # Returns /// A `String` containing the source IP address /// pub fn get_source_ip_raw(headers: &HeaderMap, remote_addr: &str) -> String { let addr = get_source_ip_from_headers(headers).unwrap_or_else(|| remote_addr.to_string()); // Default to remote address if headers not set. if let Ok(socket_addr) = SocketAddr::from_str(&addr) { socket_addr.ip().to_string() } else { addr } } /// GetSourceIP retrieves the IP from the request headers /// and falls back to remote_addr when necessary. /// It brackets IPv6 addresses. /// /// # Arguments /// * `headers` - HTTP headers from the request /// * `remote_addr` - Remote address as a string /// /// # Returns /// A `String` containing the source IP address, with IPv6 addresses bracketed /// pub fn get_source_ip(headers: &HeaderMap, remote_addr: &str) -> String { let addr = get_source_ip_raw(headers, remote_addr); if addr.contains(':') { format!("[{addr}]") } else { addr } } #[cfg(test)] mod tests { use super::*; use http::HeaderValue; fn create_test_headers() -> HeaderMap { let mut headers = HeaderMap::new(); headers.insert("x-forwarded-for", HeaderValue::from_static("192.168.1.1")); headers.insert("x-forwarded-proto", HeaderValue::from_static("https")); headers } #[test] fn test_get_source_scheme() { let headers = create_test_headers(); assert_eq!(get_source_scheme(&headers), Some("https".to_string())); } #[test] fn test_get_source_ip_from_headers() { let headers = create_test_headers(); assert_eq!(get_source_ip_from_headers(&headers), Some("192.168.1.1".to_string())); } #[test] fn test_get_source_ip_raw() { let headers = create_test_headers(); let remote_addr = "127.0.0.1:8080"; let result = get_source_ip_raw(&headers, remote_addr); assert_eq!(result, "192.168.1.1"); } #[test] fn test_get_source_ip() { let headers = create_test_headers(); let remote_addr = "127.0.0.1:8080"; let result = get_source_ip(&headers, remote_addr); assert_eq!(result, "192.168.1.1"); } #[test] fn test_get_source_ip_ipv6() { let mut headers = HeaderMap::new(); headers.insert("x-forwarded-for", HeaderValue::from_static("2001:db8::1")); let remote_addr = "127.0.0.1:8080"; let result = get_source_ip(&headers, remote_addr); assert_eq!(result, "[2001:db8::1]"); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/lib.rs
crates/s3select-query/src/lib.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. pub mod data_source; pub mod dispatcher; pub mod execution; pub mod function; pub mod instance; pub mod metadata; pub mod sql; #[cfg(test)] mod test; use rustfs_s3select_api::{QueryResult, server::dbms::DatabaseManagerSystem}; use s3s::dto::SelectObjectContentInput; use std::sync::{Arc, LazyLock}; use crate::{ execution::{factory::SqlQueryExecutionFactory, scheduler::local::LocalScheduler}, function::simple_func_manager::SimpleFunctionMetadataManager, metadata::base_table::BaseTableProvider, sql::{optimizer::CascadeOptimizerBuilder, parser::DefaultParser}, }; // Global cached components that can be reused across database instances struct GlobalComponents { func_manager: Arc<SimpleFunctionMetadataManager>, parser: Arc<DefaultParser>, query_execution_factory: Arc<SqlQueryExecutionFactory>, default_table_provider: Arc<BaseTableProvider>, } static GLOBAL_COMPONENTS: LazyLock<GlobalComponents> = LazyLock::new(|| { let func_manager = Arc::new(SimpleFunctionMetadataManager::default()); let parser = Arc::new(DefaultParser::default()); let optimizer = Arc::new(CascadeOptimizerBuilder::default().build()); let scheduler = Arc::new(LocalScheduler {}); let query_execution_factory = Arc::new(SqlQueryExecutionFactory::new(optimizer, scheduler)); let default_table_provider = Arc::new(BaseTableProvider::default()); GlobalComponents { func_manager, parser, query_execution_factory, default_table_provider, } }); /// Get or create database instance with cached components pub async fn get_global_db( input: SelectObjectContentInput, enable_debug: bool, ) -> QueryResult<Arc<dyn DatabaseManagerSystem + Send + Sync>> { let components = &*GLOBAL_COMPONENTS; let db = crate::instance::make_rustfsms_with_components( Arc::new(input), enable_debug, components.func_manager.clone(), components.parser.clone(), components.query_execution_factory.clone(), components.default_table_provider.clone(), ) .await?; Ok(Arc::new(db) as Arc<dyn DatabaseManagerSystem + Send + Sync>) } /// Create a fresh database instance without using cached components (for testing) pub async fn create_fresh_db() -> QueryResult<Arc<dyn DatabaseManagerSystem + Send + Sync>> { // Create a default test input for fresh database creation let default_input = SelectObjectContentInput { bucket: "test-bucket".to_string(), expected_bucket_owner: None, key: "test.csv".to_string(), sse_customer_algorithm: None, sse_customer_key: None, sse_customer_key_md5: None, request: s3s::dto::SelectObjectContentRequest { expression: "SELECT * FROM S3Object".to_string(), expression_type: s3s::dto::ExpressionType::from_static("SQL"), input_serialization: s3s::dto::InputSerialization::default(), output_serialization: s3s::dto::OutputSerialization::default(), request_progress: None, scan_range: None, }, }; let db = crate::instance::make_rustfsms(Arc::new(default_input), true).await?; Ok(Arc::new(db) as Arc<dyn DatabaseManagerSystem + Send + Sync>) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/instance.rs
crates/s3select-query/src/instance.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::sync::Arc; use async_trait::async_trait; use derive_builder::Builder; use rustfs_s3select_api::{ QueryResult, query::{ Query, dispatcher::QueryDispatcher, execution::QueryStateMachineRef, logical_planner::Plan, session::SessionCtxFactory, }, server::dbms::{DatabaseManagerSystem, QueryHandle}, }; use s3s::dto::SelectObjectContentInput; use crate::{ dispatcher::manager::SimpleQueryDispatcherBuilder, execution::{factory::SqlQueryExecutionFactory, scheduler::local::LocalScheduler}, function::simple_func_manager::SimpleFunctionMetadataManager, metadata::base_table::BaseTableProvider, sql::{optimizer::CascadeOptimizerBuilder, parser::DefaultParser}, }; #[derive(Builder)] pub struct RustFSms<D: QueryDispatcher> { // query dispatcher & query execution query_dispatcher: Arc<D>, } #[async_trait] impl<D> DatabaseManagerSystem for RustFSms<D> where D: QueryDispatcher, { async fn execute(&self, query: &Query) -> QueryResult<QueryHandle> { let result = self.query_dispatcher.execute_query(query).await?; Ok(QueryHandle::new(query.clone(), result)) } async fn build_query_state_machine(&self, query: Query) -> QueryResult<QueryStateMachineRef> { let query_state_machine = self.query_dispatcher.build_query_state_machine(query).await?; Ok(query_state_machine) } async fn build_logical_plan(&self, query_state_machine: QueryStateMachineRef) -> QueryResult<Option<Plan>> { let logical_plan = self.query_dispatcher.build_logical_plan(query_state_machine).await?; Ok(logical_plan) } async fn execute_logical_plan( &self, logical_plan: Plan, query_state_machine: QueryStateMachineRef, ) -> QueryResult<QueryHandle> { let query = query_state_machine.query.clone(); let result = self .query_dispatcher .execute_logical_plan(logical_plan, query_state_machine) .await?; Ok(QueryHandle::new(query.clone(), result)) } } pub async fn make_rustfsms(input: Arc<SelectObjectContentInput>, is_test: bool) -> QueryResult<impl DatabaseManagerSystem> { // init Function Manager, we can define some UDF if need let func_manager = SimpleFunctionMetadataManager::default(); // TODO session config need load global system config let session_factory = Arc::new(SessionCtxFactory { is_test }); let parser = Arc::new(DefaultParser::default()); let optimizer = Arc::new(CascadeOptimizerBuilder::default().build()); // TODO wrap, and num_threads configurable let scheduler = Arc::new(LocalScheduler {}); let query_execution_factory = Arc::new(SqlQueryExecutionFactory::new(optimizer, scheduler)); let default_table_provider = Arc::new(BaseTableProvider::default()); let query_dispatcher = SimpleQueryDispatcherBuilder::default() .with_input(input) .with_func_manager(Arc::new(func_manager)) .with_default_table_provider(default_table_provider) .with_session_factory(session_factory) .with_parser(parser) .with_query_execution_factory(query_execution_factory) .build()?; let mut builder = RustFSmsBuilder::default(); let db_server = builder.query_dispatcher(query_dispatcher).build().expect("build db server"); Ok(db_server) } pub async fn make_rustfsms_with_components( input: Arc<SelectObjectContentInput>, is_test: bool, func_manager: Arc<SimpleFunctionMetadataManager>, parser: Arc<DefaultParser>, query_execution_factory: Arc<SqlQueryExecutionFactory>, default_table_provider: Arc<BaseTableProvider>, ) -> QueryResult<impl DatabaseManagerSystem> { // TODO session config need load global system config let session_factory = Arc::new(SessionCtxFactory { is_test }); let query_dispatcher = SimpleQueryDispatcherBuilder::default() .with_input(input) .with_func_manager(func_manager) .with_default_table_provider(default_table_provider) .with_session_factory(session_factory) .with_parser(parser) .with_query_execution_factory(query_execution_factory) .build()?; let mut builder = RustFSmsBuilder::default(); let db_server = builder.query_dispatcher(query_dispatcher).build().expect("build db server"); Ok(db_server) } #[cfg(test)] mod tests { use std::sync::Arc; use datafusion::{arrow::util::pretty, assert_batches_eq}; use rustfs_s3select_api::query::{Context, Query}; use s3s::dto::{ CSVInput, CSVOutput, ExpressionType, FieldDelimiter, FileHeaderInfo, InputSerialization, OutputSerialization, RecordDelimiter, SelectObjectContentInput, SelectObjectContentRequest, }; use crate::get_global_db; #[tokio::test] #[ignore] async fn test_simple_sql() { let sql = "select * from S3Object"; let input = SelectObjectContentInput { bucket: "dandan".to_string(), expected_bucket_owner: None, key: "test.csv".to_string(), sse_customer_algorithm: None, sse_customer_key: None, sse_customer_key_md5: None, request: SelectObjectContentRequest { expression: sql.to_string(), expression_type: ExpressionType::from_static("SQL"), input_serialization: InputSerialization { csv: Some(CSVInput { file_header_info: Some(FileHeaderInfo::from_static(FileHeaderInfo::USE)), ..Default::default() }), ..Default::default() }, output_serialization: OutputSerialization { csv: Some(CSVOutput::default()), ..Default::default() }, request_progress: None, scan_range: None, }, }; let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, sql.to_string()); let result = db.execute(&query).await.unwrap(); let results = result.result().chunk_result().await.unwrap().to_vec(); let expected = [ "+----------------+---------+-----+------------+--------+", "| id | name | age | department | salary |", "+----------------+---------+-----+------------+--------+", "| 1 | Alice | 25 | HR | 5000 |", "| 2 | Bob | 30 | IT | 6000 |", "| 3 | Charlie | 35 | Finance | 7000 |", "| 4 | Diana | 22 | Marketing | 4500 |", "| 5 | Eve | 28 | IT | 5500 |", "| 6 | Frank | 40 | Finance | 8000 |", "| 7 | Grace | 26 | HR | 5200 |", "| 8 | Henry | 32 | IT | 6200 |", "| 9 | Ivy | 24 | Marketing | 4800 |", "| 10 | Jack | 38 | Finance | 7500 |", "+----------------+---------+-----+------------+--------+", ]; assert_batches_eq!(expected, &results); pretty::print_batches(&results).unwrap(); } #[tokio::test] #[ignore] async fn test_func_sql() { let sql = "SELECT * FROM S3Object s"; let input = SelectObjectContentInput { bucket: "dandan".to_string(), expected_bucket_owner: None, key: "test.csv".to_string(), sse_customer_algorithm: None, sse_customer_key: None, sse_customer_key_md5: None, request: SelectObjectContentRequest { expression: sql.to_string(), expression_type: ExpressionType::from_static("SQL"), input_serialization: InputSerialization { csv: Some(CSVInput { file_header_info: Some(FileHeaderInfo::from_static(FileHeaderInfo::IGNORE)), field_delimiter: Some(FieldDelimiter::from("╦")), record_delimiter: Some(RecordDelimiter::from("\n")), ..Default::default() }), ..Default::default() }, output_serialization: OutputSerialization { csv: Some(CSVOutput::default()), ..Default::default() }, request_progress: None, scan_range: None, }, }; let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, sql.to_string()); let result = db.execute(&query).await.unwrap(); let results = result.result().chunk_result().await.unwrap().to_vec(); pretty::print_batches(&results).unwrap(); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/metadata/base_table.rs
crates/s3select-query/src/metadata/base_table.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::sync::Arc; use datafusion::common::Result as DFResult; use datafusion::datasource::listing::ListingTable; use crate::data_source::table_source::TableHandle; use super::TableHandleProvider; #[derive(Default)] pub struct BaseTableProvider {} impl TableHandleProvider for BaseTableProvider { fn build_table_handle(&self, provider: Arc<ListingTable>) -> DFResult<TableHandle> { Ok(TableHandle::External(provider)) } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/metadata/mod.rs
crates/s3select-query/src/metadata/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::sync::Arc; use async_trait::async_trait; use datafusion::arrow::datatypes::DataType; use datafusion::common::Result as DFResult; use datafusion::datasource::listing::ListingTable; use datafusion::logical_expr::var_provider::is_system_variables; use datafusion::logical_expr::{AggregateUDF, ScalarUDF, TableSource, WindowUDF}; use datafusion::variable::VarType; use datafusion::{ config::ConfigOptions, sql::{TableReference, planner::ContextProvider}, }; use rustfs_s3select_api::query::{function::FuncMetaManagerRef, session::SessionCtx}; use crate::data_source::table_source::{TableHandle, TableSourceAdapter}; pub mod base_table; #[async_trait] pub trait ContextProviderExtension: ContextProvider { fn get_table_source_(&self, name: TableReference) -> datafusion::common::Result<Arc<TableSourceAdapter>>; } pub type TableHandleProviderRef = Arc<dyn TableHandleProvider + Send + Sync>; pub trait TableHandleProvider { fn build_table_handle(&self, provider: Arc<ListingTable>) -> DFResult<TableHandle>; } pub struct MetadataProvider { provider: Arc<ListingTable>, session: SessionCtx, config_options: ConfigOptions, func_manager: FuncMetaManagerRef, current_session_table_provider: TableHandleProviderRef, } impl MetadataProvider { #[allow(clippy::too_many_arguments)] pub fn new( provider: Arc<ListingTable>, current_session_table_provider: TableHandleProviderRef, func_manager: FuncMetaManagerRef, session: SessionCtx, ) -> Self { Self { provider, current_session_table_provider, config_options: session.inner().config_options().as_ref().clone(), session, func_manager, } } fn build_table_handle(&self) -> datafusion::common::Result<TableHandle> { self.current_session_table_provider.build_table_handle(self.provider.clone()) } } impl ContextProviderExtension for MetadataProvider { fn get_table_source_(&self, table_ref: TableReference) -> datafusion::common::Result<Arc<TableSourceAdapter>> { let name = table_ref.clone().resolve("", ""); let table_name = &*name.table; let table_handle = self.build_table_handle()?; Ok(Arc::new(TableSourceAdapter::try_new(table_ref.clone(), table_name, table_handle)?)) } } impl ContextProvider for MetadataProvider { fn get_function_meta(&self, name: &str) -> Option<Arc<ScalarUDF>> { self.func_manager .udf(name) .ok() .or(self.session.inner().scalar_functions().get(name).cloned()) } fn get_aggregate_meta(&self, name: &str) -> Option<Arc<AggregateUDF>> { self.func_manager.udaf(name).ok() } fn get_variable_type(&self, variable_names: &[String]) -> Option<DataType> { if variable_names.is_empty() { return None; } let var_type = if is_system_variables(variable_names) { VarType::System } else { VarType::UserDefined }; self.session .inner() .execution_props() .get_var_provider(var_type) .and_then(|p| p.get_type(variable_names)) } fn options(&self) -> &ConfigOptions { // TODO refactor &self.config_options } fn get_window_meta(&self, name: &str) -> Option<Arc<WindowUDF>> { self.func_manager.udwf(name).ok() } fn get_table_source(&self, name: TableReference) -> DFResult<Arc<dyn TableSource>> { Ok(self.get_table_source_(name)?) } fn udf_names(&self) -> Vec<String> { self.func_manager.udfs() } fn udaf_names(&self) -> Vec<String> { self.func_manager.udafs() } fn udwf_names(&self) -> Vec<String> { self.func_manager.udwfs() } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/dispatcher/manager.rs
crates/s3select-query/src/dispatcher/manager.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::{ ops::Deref, pin::Pin, sync::Arc, task::{Context, Poll}, }; use async_trait::async_trait; use datafusion::{ arrow::{ datatypes::{Schema, SchemaRef}, record_batch::RecordBatch, }, datasource::{ file_format::{csv::CsvFormat, json::JsonFormat, parquet::ParquetFormat}, listing::{ListingOptions, ListingTable, ListingTableConfig, ListingTableUrl}, }, error::Result as DFResult, execution::{RecordBatchStream, SendableRecordBatchStream}, }; use futures::{Stream, StreamExt}; use rustfs_s3select_api::{ QueryError, QueryResult, query::{ Query, ast::ExtStatement, dispatcher::QueryDispatcher, execution::{Output, QueryStateMachine}, function::FuncMetaManagerRef, logical_planner::{LogicalPlanner, Plan}, parser::Parser, session::{SessionCtx, SessionCtxFactory}, }, }; use s3s::dto::{FileHeaderInfo, SelectObjectContentInput}; use std::sync::LazyLock; use crate::{ execution::factory::QueryExecutionFactoryRef, metadata::{ContextProviderExtension, MetadataProvider, TableHandleProviderRef, base_table::BaseTableProvider}, sql::logical::planner::DefaultLogicalPlanner, }; static IGNORE: LazyLock<FileHeaderInfo> = LazyLock::new(|| FileHeaderInfo::from_static(FileHeaderInfo::IGNORE)); static NONE: LazyLock<FileHeaderInfo> = LazyLock::new(|| FileHeaderInfo::from_static(FileHeaderInfo::NONE)); static USE: LazyLock<FileHeaderInfo> = LazyLock::new(|| FileHeaderInfo::from_static(FileHeaderInfo::USE)); #[derive(Clone)] pub struct SimpleQueryDispatcher { input: Arc<SelectObjectContentInput>, // client for default tenant _default_table_provider: TableHandleProviderRef, session_factory: Arc<SessionCtxFactory>, // parser parser: Arc<dyn Parser + Send + Sync>, // get query execution factory query_execution_factory: QueryExecutionFactoryRef, func_manager: FuncMetaManagerRef, } #[async_trait] impl QueryDispatcher for SimpleQueryDispatcher { async fn execute_query(&self, query: &Query) -> QueryResult<Output> { let query_state_machine = { self.build_query_state_machine(query.clone()).await? }; let logical_plan = self.build_logical_plan(query_state_machine.clone()).await?; let logical_plan = match logical_plan { Some(plan) => plan, None => return Ok(Output::Nil(())), }; let result = self.execute_logical_plan(logical_plan, query_state_machine).await?; Ok(result) } async fn build_logical_plan(&self, query_state_machine: Arc<QueryStateMachine>) -> QueryResult<Option<Plan>> { let session = &query_state_machine.session; let query = &query_state_machine.query; let scheme_provider = self.build_scheme_provider(session).await?; let logical_planner = DefaultLogicalPlanner::new(&scheme_provider); let statements = self.parser.parse(query.content())?; // not allow multi statement if statements.len() > 1 { return Err(QueryError::MultiStatement { num: statements.len(), sql: query_state_machine.query.content().to_string(), }); } let stmt = match statements.front() { Some(stmt) => stmt.clone(), None => return Ok(None), }; let logical_plan = self .statement_to_logical_plan(stmt, &logical_planner, query_state_machine) .await?; Ok(Some(logical_plan)) } async fn execute_logical_plan(&self, logical_plan: Plan, query_state_machine: Arc<QueryStateMachine>) -> QueryResult<Output> { self.execute_logical_plan(logical_plan, query_state_machine).await } async fn build_query_state_machine(&self, query: Query) -> QueryResult<Arc<QueryStateMachine>> { let session = self.session_factory.create_session_ctx(query.context()).await?; let query_state_machine = Arc::new(QueryStateMachine::begin(query, session)); Ok(query_state_machine) } } impl SimpleQueryDispatcher { async fn statement_to_logical_plan<S: ContextProviderExtension + Send + Sync>( &self, stmt: ExtStatement, logical_planner: &DefaultLogicalPlanner<'_, S>, query_state_machine: Arc<QueryStateMachine>, ) -> QueryResult<Plan> { // begin analyze query_state_machine.begin_analyze(); let logical_plan = logical_planner .create_logical_plan(stmt, &query_state_machine.session) .await?; query_state_machine.end_analyze(); Ok(logical_plan) } async fn execute_logical_plan(&self, logical_plan: Plan, query_state_machine: Arc<QueryStateMachine>) -> QueryResult<Output> { let execution = self .query_execution_factory .create_query_execution(logical_plan, query_state_machine.clone()) .await?; match execution.start().await { Ok(Output::StreamData(stream)) => Ok(Output::StreamData(Box::pin(TrackedRecordBatchStream { inner: stream }))), Ok(nil @ Output::Nil(_)) => Ok(nil), Err(err) => Err(err), } } async fn build_scheme_provider(&self, session: &SessionCtx) -> QueryResult<MetadataProvider> { let path = format!("s3://{}/{}", self.input.bucket, self.input.key); let table_path = ListingTableUrl::parse(path)?; let (listing_options, need_rename_volume_name, need_ignore_volume_name) = if let Some(csv) = self.input.request.input_serialization.csv.as_ref() { let mut need_rename_volume_name = false; let mut need_ignore_volume_name = false; let mut file_format = CsvFormat::default() .with_comment( csv.comments .clone() .map(|c| c.as_bytes().first().copied().unwrap_or_default()), ) .with_escape( csv.quote_escape_character .clone() .map(|e| e.as_bytes().first().copied().unwrap_or_default()), ); if let Some(delimiter) = csv.field_delimiter.as_ref() && delimiter.len() == 1 { file_format = file_format.with_delimiter(delimiter.as_bytes()[0]); } // TODO waiting for processing @junxiang Mu // if csv.file_header_info.is_some() {} match csv.file_header_info.as_ref() { Some(info) => { if *info == *NONE { file_format = file_format.with_has_header(false); need_rename_volume_name = true; } else if *info == *IGNORE { file_format = file_format.with_has_header(true); need_rename_volume_name = true; need_ignore_volume_name = true; } else if *info == *USE { file_format = file_format.with_has_header(true); } else { return Err(QueryError::NotImplemented { err: "unsupported FileHeaderInfo".to_string(), }); } } _ => { return Err(QueryError::NotImplemented { err: "unsupported FileHeaderInfo".to_string(), }); } } if let Some(quote) = csv.quote_character.as_ref() { file_format = file_format.with_quote(quote.as_bytes().first().copied().unwrap_or_default()); } ( ListingOptions::new(Arc::new(file_format)).with_file_extension(".csv"), need_rename_volume_name, need_ignore_volume_name, ) } else if self.input.request.input_serialization.parquet.is_some() { let file_format = ParquetFormat::new(); (ListingOptions::new(Arc::new(file_format)).with_file_extension(".parquet"), false, false) } else if self.input.request.input_serialization.json.is_some() { let file_format = JsonFormat::default(); (ListingOptions::new(Arc::new(file_format)).with_file_extension(".json"), false, false) } else { return Err(QueryError::NotImplemented { err: "not support this file type".to_string(), }); }; let resolve_schema = listing_options.infer_schema(session.inner(), &table_path).await?; let config = if need_rename_volume_name { let mut new_fields = Vec::new(); for (i, field) in resolve_schema.fields().iter().enumerate() { let f_name = field.name(); let mut_field = field.deref().clone(); if f_name.starts_with("column_") { let re_name = f_name.replace("column_", "_"); new_fields.push(mut_field.with_name(re_name)); } else if need_ignore_volume_name { let re_name = format!("_{}", i + 1); new_fields.push(mut_field.with_name(re_name)); } else { new_fields.push(mut_field); } } let new_schema = Arc::new(Schema::new(new_fields).with_metadata(resolve_schema.metadata().clone())); ListingTableConfig::new(table_path) .with_listing_options(listing_options) .with_schema(new_schema) } else { ListingTableConfig::new(table_path) .with_listing_options(listing_options) .with_schema(resolve_schema) }; // rename default let provider = Arc::new(ListingTable::try_new(config)?); let current_session_table_provider = self.build_table_handle_provider()?; let metadata_provider = MetadataProvider::new(provider, current_session_table_provider, self.func_manager.clone(), session.clone()); Ok(metadata_provider) } fn build_table_handle_provider(&self) -> QueryResult<TableHandleProviderRef> { let current_session_table_provider: Arc<BaseTableProvider> = Arc::new(BaseTableProvider::default()); Ok(current_session_table_provider) } } pub struct TrackedRecordBatchStream { inner: SendableRecordBatchStream, } impl RecordBatchStream for TrackedRecordBatchStream { fn schema(&self) -> SchemaRef { self.inner.schema() } } impl Stream for TrackedRecordBatchStream { type Item = DFResult<RecordBatch>; fn poll_next(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Option<Self::Item>> { self.inner.poll_next_unpin(cx) } } #[derive(Default, Clone)] pub struct SimpleQueryDispatcherBuilder { input: Option<Arc<SelectObjectContentInput>>, default_table_provider: Option<TableHandleProviderRef>, session_factory: Option<Arc<SessionCtxFactory>>, parser: Option<Arc<dyn Parser + Send + Sync>>, query_execution_factory: Option<QueryExecutionFactoryRef>, func_manager: Option<FuncMetaManagerRef>, } impl SimpleQueryDispatcherBuilder { pub fn with_input(mut self, input: Arc<SelectObjectContentInput>) -> Self { self.input = Some(input); self } pub fn with_default_table_provider(mut self, default_table_provider: TableHandleProviderRef) -> Self { self.default_table_provider = Some(default_table_provider); self } pub fn with_session_factory(mut self, session_factory: Arc<SessionCtxFactory>) -> Self { self.session_factory = Some(session_factory); self } pub fn with_parser(mut self, parser: Arc<dyn Parser + Send + Sync>) -> Self { self.parser = Some(parser); self } pub fn with_query_execution_factory(mut self, query_execution_factory: QueryExecutionFactoryRef) -> Self { self.query_execution_factory = Some(query_execution_factory); self } pub fn with_func_manager(mut self, func_manager: FuncMetaManagerRef) -> Self { self.func_manager = Some(func_manager); self } pub fn build(self) -> QueryResult<Arc<SimpleQueryDispatcher>> { let input = self.input.ok_or_else(|| QueryError::BuildQueryDispatcher { err: "lost of input".to_string(), })?; let session_factory = self.session_factory.ok_or_else(|| QueryError::BuildQueryDispatcher { err: "lost of session_factory".to_string(), })?; let parser = self.parser.ok_or_else(|| QueryError::BuildQueryDispatcher { err: "lost of parser".to_string(), })?; let query_execution_factory = self.query_execution_factory.ok_or_else(|| QueryError::BuildQueryDispatcher { err: "lost of query_execution_factory".to_string(), })?; let func_manager = self.func_manager.ok_or_else(|| QueryError::BuildQueryDispatcher { err: "lost of func_manager".to_string(), })?; let default_table_provider = self.default_table_provider.ok_or_else(|| QueryError::BuildQueryDispatcher { err: "lost of default_table_provider".to_string(), })?; let dispatcher = Arc::new(SimpleQueryDispatcher { input, _default_table_provider: default_table_provider, session_factory, parser, query_execution_factory, func_manager, }); Ok(dispatcher) } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/dispatcher/mod.rs
crates/s3select-query/src/dispatcher/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. pub mod manager;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/test/integration_test.rs
crates/s3select-query/src/test/integration_test.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #[cfg(test)] mod integration_tests { use crate::{create_fresh_db, get_global_db, instance::make_rustfsms}; use rustfs_s3select_api::{ QueryError, query::{Context, Query}, }; use s3s::dto::{ CSVInput, CSVOutput, ExpressionType, FileHeaderInfo, InputSerialization, OutputSerialization, SelectObjectContentInput, SelectObjectContentRequest, }; use std::sync::Arc; fn create_test_input(sql: &str) -> SelectObjectContentInput { SelectObjectContentInput { bucket: "test-bucket".to_string(), expected_bucket_owner: None, key: "test.csv".to_string(), sse_customer_algorithm: None, sse_customer_key: None, sse_customer_key_md5: None, request: SelectObjectContentRequest { expression: sql.to_string(), expression_type: ExpressionType::from_static("SQL"), input_serialization: InputSerialization { csv: Some(CSVInput { file_header_info: Some(FileHeaderInfo::from_static(FileHeaderInfo::USE)), ..Default::default() }), ..Default::default() }, output_serialization: OutputSerialization { csv: Some(CSVOutput::default()), ..Default::default() }, request_progress: None, scan_range: None, }, } } #[tokio::test] async fn test_database_creation() { let input = create_test_input("SELECT * FROM S3Object"); let result = make_rustfsms(Arc::new(input), true).await; assert!(result.is_ok()); } #[tokio::test] async fn test_global_db_creation() { let input = create_test_input("SELECT * FROM S3Object"); let result = get_global_db(input.clone(), true).await; assert!(result.is_ok()); } #[tokio::test] async fn test_fresh_db_creation() { let result = create_fresh_db().await; assert!(result.is_ok()); } #[tokio::test] async fn test_simple_select_query() { let sql = "SELECT * FROM S3Object"; let input = create_test_input(sql); let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, sql.to_string()); let result = db.execute(&query).await; assert!(result.is_ok()); let query_handle = result.unwrap(); let output = query_handle.result().chunk_result().await; assert!(output.is_ok()); } #[tokio::test] async fn test_select_with_where_clause() { let sql = "SELECT name, age FROM S3Object WHERE age > 30"; let input = create_test_input(sql); let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, sql.to_string()); let result = db.execute(&query).await; assert!(result.is_ok()); } #[tokio::test] async fn test_select_with_aggregation() { let sql = "SELECT department, COUNT(*) as count FROM S3Object GROUP BY department"; let input = create_test_input(sql); let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, sql.to_string()); let result = db.execute(&query).await; // Aggregation queries might fail due to lack of actual data, which is acceptable match result { Ok(_) => { // If successful, that's great } Err(_) => { // Expected to fail due to no actual data source } } } #[tokio::test] async fn test_invalid_sql_syntax() { let sql = "INVALID SQL SYNTAX"; let input = create_test_input(sql); let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, sql.to_string()); let result = db.execute(&query).await; assert!(result.is_err()); } #[tokio::test] async fn test_multi_statement_error() { let sql = "SELECT * FROM S3Object; SELECT 1;"; let input = create_test_input(sql); let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, sql.to_string()); let result = db.execute(&query).await; assert!(result.is_err()); if let Err(QueryError::MultiStatement { num, .. }) = result { assert_eq!(num, 2); } else { panic!("Expected MultiStatement error"); } } #[tokio::test] async fn test_query_state_machine_workflow() { let sql = "SELECT * FROM S3Object"; let input = create_test_input(sql); let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, sql.to_string()); // Test state machine creation let state_machine = db.build_query_state_machine(query.clone()).await; assert!(state_machine.is_ok()); let state_machine = state_machine.unwrap(); // Test logical plan building let logical_plan = db.build_logical_plan(state_machine.clone()).await; assert!(logical_plan.is_ok()); // Test execution if plan exists if let Ok(Some(plan)) = logical_plan { let execution_result = db.execute_logical_plan(plan, state_machine).await; assert!(execution_result.is_ok()); } } #[tokio::test] async fn test_query_with_limit() { let sql = "SELECT * FROM S3Object LIMIT 5"; let input = create_test_input(sql); let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, sql.to_string()); let result = db.execute(&query).await; assert!(result.is_ok()); let query_handle = result.unwrap(); let output = query_handle.result().chunk_result().await.unwrap(); // Verify that we get results (exact count depends on test data) let total_rows: usize = output.iter().map(|batch| batch.num_rows()).sum(); assert!(total_rows <= 5); } #[tokio::test] async fn test_query_with_order_by() { let sql = "SELECT name, age FROM S3Object ORDER BY age DESC"; let input = create_test_input(sql); let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, sql.to_string()); let result = db.execute(&query).await; assert!(result.is_ok()); } #[tokio::test] async fn test_concurrent_queries() { let sql = "SELECT * FROM S3Object"; let input = create_test_input(sql); let db = get_global_db(input.clone(), true).await.unwrap(); // Execute multiple queries concurrently let mut handles = vec![]; for i in 0..3 { let query = Query::new( Context { input: Arc::new(input.clone()), }, format!("SELECT * FROM S3Object LIMIT {}", i + 1), ); let db_clone = db.clone(); let handle = tokio::spawn(async move { db_clone.execute(&query).await }); handles.push(handle); } // Wait for all queries to complete for handle in handles { let result = handle.await.unwrap(); assert!(result.is_ok()); } } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/test/error_handling_test.rs
crates/s3select-query/src/test/error_handling_test.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #[cfg(test)] mod error_handling_tests { use crate::get_global_db; use rustfs_s3select_api::{ QueryError, query::{Context, Query}, }; use s3s::dto::{ CSVInput, ExpressionType, FileHeaderInfo, InputSerialization, SelectObjectContentInput, SelectObjectContentRequest, }; use std::sync::Arc; fn create_test_input_with_sql(sql: &str) -> SelectObjectContentInput { SelectObjectContentInput { bucket: "test-bucket".to_string(), expected_bucket_owner: None, key: "test.csv".to_string(), sse_customer_algorithm: None, sse_customer_key: None, sse_customer_key_md5: None, request: SelectObjectContentRequest { expression: sql.to_string(), expression_type: ExpressionType::from_static("SQL"), input_serialization: InputSerialization { csv: Some(CSVInput { file_header_info: Some(FileHeaderInfo::from_static(FileHeaderInfo::USE)), ..Default::default() }), ..Default::default() }, output_serialization: s3s::dto::OutputSerialization::default(), request_progress: None, scan_range: None, }, } } #[tokio::test] async fn test_syntax_error_handling() { let invalid_sqls = vec![ "INVALID SQL", "SELECT FROM", "SELECT * FORM S3Object", // typo in FROM "SELECT * FROM", "SELECT * FROM S3Object WHERE", "SELECT COUNT( FROM S3Object", // missing closing parenthesis ]; for sql in invalid_sqls { let input = create_test_input_with_sql(sql); let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, sql.to_string()); let result = db.execute(&query).await; assert!(result.is_err(), "Expected error for SQL: {sql}"); } } #[tokio::test] async fn test_multi_statement_error() { let multi_statement_sqls = vec![ "SELECT * FROM S3Object; SELECT 1;", "SELECT 1; SELECT 2; SELECT 3;", "SELECT * FROM S3Object; DROP TABLE test;", ]; for sql in multi_statement_sqls { let input = create_test_input_with_sql(sql); let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, sql.to_string()); let result = db.execute(&query).await; assert!(result.is_err(), "Expected multi-statement error for SQL: {sql}"); if let Err(QueryError::MultiStatement { num, .. }) = result { assert!(num >= 2, "Expected at least 2 statements, got: {num}"); } } } #[tokio::test] async fn test_unsupported_operations() { let unsupported_sqls = vec![ "INSERT INTO S3Object VALUES (1, 'test')", "UPDATE S3Object SET name = 'test'", "DELETE FROM S3Object", "CREATE TABLE test (id INT)", "DROP TABLE S3Object", ]; for sql in unsupported_sqls { let input = create_test_input_with_sql(sql); let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, sql.to_string()); let result = db.execute(&query).await; // These should either fail with syntax error or not implemented error assert!(result.is_err(), "Expected error for unsupported SQL: {sql}"); } } #[tokio::test] async fn test_invalid_column_references() { let invalid_column_sqls = vec![ "SELECT nonexistent_column FROM S3Object", "SELECT * FROM S3Object WHERE nonexistent_column = 1", "SELECT * FROM S3Object ORDER BY nonexistent_column", "SELECT * FROM S3Object GROUP BY nonexistent_column", ]; for sql in invalid_column_sqls { let input = create_test_input_with_sql(sql); let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, sql.to_string()); let result = db.execute(&query).await; // These might succeed or fail depending on schema inference // The test verifies that the system handles them gracefully match result { Ok(_) => { // If it succeeds, verify we can get results let handle = result.unwrap(); let output = handle.result().chunk_result().await; // Should either succeed with empty results or fail gracefully let _ = output; } Err(_) => { // Expected to fail - this is acceptable } } } } #[tokio::test] async fn test_complex_query_error_recovery() { let complex_invalid_sql = r#" SELECT name, age, INVALID_FUNCTION(salary) as invalid_calc, department FROM S3Object WHERE age > 'invalid_number' GROUP BY department, nonexistent_column HAVING COUNT(*) > INVALID_FUNCTION() ORDER BY invalid_column "#; let input = create_test_input_with_sql(complex_invalid_sql); let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, complex_invalid_sql.to_string()); let result = db.execute(&query).await; assert!(result.is_err(), "Expected error for complex invalid SQL"); } #[tokio::test] async fn test_empty_query() { let empty_sqls = vec!["", " ", "\n\t \n"]; for sql in empty_sqls { let input = create_test_input_with_sql(sql); let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, sql.to_string()); let result = db.execute(&query).await; // Empty queries might be handled differently by the parser match result { Ok(_) => { // Some parsers might accept empty queries } Err(_) => { // Expected to fail for empty SQL } } } } #[tokio::test] async fn test_very_long_query() { // Create a very long but valid query let mut long_sql = "SELECT ".to_string(); for i in 0..1000 { if i > 0 { long_sql.push_str(", "); } long_sql.push_str(&format!("'column_{i}' as col_{i}")); } long_sql.push_str(" FROM S3Object LIMIT 1"); let input = create_test_input_with_sql(&long_sql); let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, long_sql); let result = db.execute(&query).await; // This should either succeed or fail gracefully match result { Ok(handle) => { let output = handle.result().chunk_result().await; assert!(output.is_ok(), "Query execution should complete successfully"); } Err(_) => { // Acceptable to fail due to resource constraints } } } #[tokio::test] async fn test_sql_injection_patterns() { let injection_patterns = vec![ "SELECT * FROM S3Object WHERE name = 'test'; DROP TABLE users; --", "SELECT * FROM S3Object UNION SELECT * FROM information_schema.tables", "SELECT * FROM S3Object WHERE 1=1 OR 1=1", ]; for sql in injection_patterns { let input = create_test_input_with_sql(sql); let db = get_global_db(input.clone(), true).await.unwrap(); let query = Query::new(Context { input: Arc::new(input) }, sql.to_string()); let result = db.execute(&query).await; // These should be handled safely - either succeed with limited scope or fail match result { Ok(_) => { // If successful, it should only access S3Object data } Err(_) => { // Expected to fail for security reasons } } } } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/test/mod.rs
crates/s3select-query/src/test/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Test modules for s3select-query pub mod error_handling_test; pub mod integration_test;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/execution/mod.rs
crates/s3select-query/src/execution/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. pub mod factory; pub mod query; pub mod scheduler;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/execution/query.rs
crates/s3select-query/src/execution/query.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::sync::Arc; use async_trait::async_trait; use futures::stream::AbortHandle; use parking_lot::Mutex; use rustfs_s3select_api::query::execution::{Output, QueryExecution, QueryStateMachineRef}; use rustfs_s3select_api::query::logical_planner::QueryPlan; use rustfs_s3select_api::query::optimizer::Optimizer; use rustfs_s3select_api::query::scheduler::SchedulerRef; use rustfs_s3select_api::{QueryError, QueryResult}; use tracing::debug; pub struct SqlQueryExecution { query_state_machine: QueryStateMachineRef, plan: QueryPlan, optimizer: Arc<dyn Optimizer + Send + Sync>, scheduler: SchedulerRef, abort_handle: Mutex<Option<AbortHandle>>, } impl SqlQueryExecution { pub fn new( query_state_machine: QueryStateMachineRef, plan: QueryPlan, optimizer: Arc<dyn Optimizer + Send + Sync>, scheduler: SchedulerRef, ) -> Self { Self { query_state_machine, plan, optimizer, scheduler, abort_handle: Mutex::new(None), } } async fn start(&self) -> QueryResult<Output> { // begin optimize self.query_state_machine.begin_optimize(); let physical_plan = self.optimizer.optimize(&self.plan, &self.query_state_machine.session).await?; self.query_state_machine.end_optimize(); // begin schedule self.query_state_machine.begin_schedule(); let stream = self .scheduler .schedule(physical_plan.clone(), self.query_state_machine.session.inner().task_ctx()) .await? .stream(); debug!("Success build result stream."); self.query_state_machine.end_schedule(); Ok(Output::StreamData(stream)) } } #[async_trait] impl QueryExecution for SqlQueryExecution { async fn start(&self) -> QueryResult<Output> { let (task, abort_handle) = futures::future::abortable(self.start()); { *self.abort_handle.lock() = Some(abort_handle); } task.await.map_err(|_| QueryError::Cancel)? } fn cancel(&self) -> QueryResult<()> { debug!( "cancel sql query execution: sql: {}, state: {:?}", self.query_state_machine.query.content(), self.query_state_machine.state() ); // change state self.query_state_machine.cancel(); // stop future task if let Some(e) = self.abort_handle.lock().as_ref() { e.abort() }; debug!( "canceled sql query execution: sql: {}, state: {:?}", self.query_state_machine.query.content(), self.query_state_machine.state() ); Ok(()) } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/execution/factory.rs
crates/s3select-query/src/execution/factory.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::sync::Arc; use async_trait::async_trait; use rustfs_s3select_api::{ QueryError, query::{ execution::{QueryExecutionFactory, QueryExecutionRef, QueryStateMachineRef}, logical_planner::Plan, optimizer::Optimizer, scheduler::SchedulerRef, }, }; use super::query::SqlQueryExecution; pub type QueryExecutionFactoryRef = Arc<dyn QueryExecutionFactory + Send + Sync>; pub struct SqlQueryExecutionFactory { optimizer: Arc<dyn Optimizer + Send + Sync>, scheduler: SchedulerRef, } impl SqlQueryExecutionFactory { #[inline(always)] pub fn new(optimizer: Arc<dyn Optimizer + Send + Sync>, scheduler: SchedulerRef) -> Self { Self { optimizer, scheduler } } } #[async_trait] impl QueryExecutionFactory for SqlQueryExecutionFactory { async fn create_query_execution( &self, plan: Plan, state_machine: QueryStateMachineRef, ) -> Result<QueryExecutionRef, QueryError> { match plan { Plan::Query(query_plan) => Ok(Arc::new(SqlQueryExecution::new( state_machine, query_plan, self.optimizer.clone(), self.scheduler.clone(), ))), } } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/execution/scheduler/local.rs
crates/s3select-query/src/execution/scheduler/local.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::sync::Arc; use async_trait::async_trait; use datafusion::error::DataFusionError; use datafusion::execution::context::TaskContext; use datafusion::physical_plan::{ExecutionPlan, execute_stream}; use rustfs_s3select_api::query::scheduler::{ExecutionResults, Scheduler}; pub struct LocalScheduler {} #[async_trait] impl Scheduler for LocalScheduler { async fn schedule( &self, plan: Arc<dyn ExecutionPlan>, context: Arc<TaskContext>, ) -> Result<ExecutionResults, DataFusionError> { let stream = execute_stream(plan, context)?; Ok(ExecutionResults::new(stream)) } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/execution/scheduler/mod.rs
crates/s3select-query/src/execution/scheduler/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. pub mod local;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/data_source/table_source.rs
crates/s3select-query/src/data_source/table_source.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::any::Any; use std::borrow::Cow; use std::fmt::Display; use std::sync::Arc; use std::write; use async_trait::async_trait; use datafusion::arrow::datatypes::SchemaRef; use datafusion::common::Result as DFResult; use datafusion::datasource::listing::ListingTable; use datafusion::datasource::{TableProvider, provider_as_source}; use datafusion::error::DataFusionError; use datafusion::logical_expr::{LogicalPlan, LogicalPlanBuilder, TableProviderFilterPushDown, TableSource}; use datafusion::prelude::Expr; use datafusion::sql::TableReference; use tracing::debug; pub const TEMP_LOCATION_TABLE_NAME: &str = "external_location_table"; pub struct TableSourceAdapter { database_name: String, table_name: String, table_handle: TableHandle, plan: LogicalPlan, } impl TableSourceAdapter { pub fn try_new( table_ref: impl Into<TableReference>, table_name: impl Into<String>, table_handle: impl Into<TableHandle>, ) -> Result<Self, DataFusionError> { let table_name: String = table_name.into(); let table_handle = table_handle.into(); let plan = match &table_handle { // TableScan TableHandle::External(t) => { let table_source = provider_as_source(t.clone()); LogicalPlanBuilder::scan(table_ref, table_source, None)?.build()? } // TableScan TableHandle::TableProvider(t) => { let table_source = provider_as_source(t.clone()); if let Some(plan) = table_source.get_logical_plan() { LogicalPlanBuilder::from(plan.into_owned()).build()? } else { LogicalPlanBuilder::scan(table_ref, table_source, None)?.build()? } } }; debug!("Table source logical plan node of {}:\n{}", table_name, plan.display_indent_schema()); Ok(Self { database_name: "default_db".to_string(), table_name, table_handle, plan, }) } pub fn database_name(&self) -> &str { &self.database_name } pub fn table_name(&self) -> &str { &self.table_name } pub fn table_handle(&self) -> &TableHandle { &self.table_handle } } #[async_trait] impl TableSource for TableSourceAdapter { fn as_any(&self) -> &dyn Any { self } fn schema(&self) -> SchemaRef { self.table_handle.schema() } fn supports_filters_pushdown(&self, filter: &[&Expr]) -> DFResult<Vec<TableProviderFilterPushDown>> { self.table_handle.supports_filters_pushdown(filter) } /// Called by [`InlineTableScan`] fn get_logical_plan(&self) -> Option<Cow<'_, LogicalPlan>> { Some(Cow::Owned(self.plan.clone())) } } #[derive(Clone)] pub enum TableHandle { TableProvider(Arc<dyn TableProvider>), External(Arc<ListingTable>), } impl TableHandle { pub fn schema(&self) -> SchemaRef { match self { Self::External(t) => t.schema(), Self::TableProvider(t) => t.schema(), } } pub fn supports_filters_pushdown(&self, filter: &[&Expr]) -> DFResult<Vec<TableProviderFilterPushDown>> { match self { Self::External(t) => t.supports_filters_pushdown(filter), Self::TableProvider(t) => t.supports_filters_pushdown(filter), } } } impl From<Arc<dyn TableProvider>> for TableHandle { fn from(value: Arc<dyn TableProvider>) -> Self { TableHandle::TableProvider(value) } } impl From<Arc<ListingTable>> for TableHandle { fn from(value: Arc<ListingTable>) -> Self { TableHandle::External(value) } } impl Display for TableHandle { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { match self { Self::External(e) => write!(f, "External({:?})", e.table_paths()), Self::TableProvider(_) => write!(f, "TableProvider"), } } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/data_source/mod.rs
crates/s3select-query/src/data_source/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. pub mod table_source;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/function/simple_func_manager.rs
crates/s3select-query/src/function/simple_func_manager.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::collections::HashMap; use std::sync::Arc; use datafusion::execution::SessionStateDefaults; use datafusion::logical_expr::{AggregateUDF, ScalarUDF, WindowUDF}; use rustfs_s3select_api::query::function::FunctionMetadataManager; use rustfs_s3select_api::{QueryError, QueryResult}; use tracing::debug; pub type SimpleFunctionMetadataManagerRef = Arc<SimpleFunctionMetadataManager>; #[derive(Debug)] pub struct SimpleFunctionMetadataManager { /// Scalar functions that are registered with the context pub scalar_functions: HashMap<String, Arc<ScalarUDF>>, /// Aggregate functions registered in the context pub aggregate_functions: HashMap<String, Arc<AggregateUDF>>, /// Window functions registered in the context pub window_functions: HashMap<String, Arc<WindowUDF>>, } impl Default for SimpleFunctionMetadataManager { fn default() -> Self { let mut func_meta_manager = Self { scalar_functions: Default::default(), aggregate_functions: Default::default(), window_functions: Default::default(), }; SessionStateDefaults::default_scalar_functions().into_iter().for_each(|udf| { let existing_udf = func_meta_manager.register_udf(udf.clone()); if let Ok(()) = existing_udf { debug!("Overwrote an existing UDF: {}", udf.name()); } }); SessionStateDefaults::default_aggregate_functions() .into_iter() .for_each(|udaf| { let existing_udaf = func_meta_manager.register_udaf(udaf.clone()); if let Ok(()) = existing_udaf { debug!("Overwrote an existing UDAF: {}", udaf.name()); } }); SessionStateDefaults::default_window_functions().into_iter().for_each(|udwf| { let existing_udwf = func_meta_manager.register_udwf(udwf.clone()); if let Ok(()) = existing_udwf { debug!("Overwrote an existing UDWF: {}", udwf.name()); } }); func_meta_manager } } impl FunctionMetadataManager for SimpleFunctionMetadataManager { fn register_udf(&mut self, f: Arc<ScalarUDF>) -> QueryResult<()> { self.scalar_functions.insert(f.inner().name().to_uppercase(), f); Ok(()) } fn register_udaf(&mut self, f: Arc<AggregateUDF>) -> QueryResult<()> { self.aggregate_functions.insert(f.inner().name().to_uppercase(), f); Ok(()) } fn register_udwf(&mut self, f: Arc<WindowUDF>) -> QueryResult<()> { self.window_functions.insert(f.inner().name().to_uppercase(), f); Ok(()) } fn udf(&self, name: &str) -> QueryResult<Arc<ScalarUDF>> { let result = self.scalar_functions.get(&name.to_uppercase()); result .cloned() .ok_or_else(|| QueryError::FunctionExists { name: name.to_string() }) } fn udaf(&self, name: &str) -> QueryResult<Arc<AggregateUDF>> { let result = self.aggregate_functions.get(&name.to_uppercase()); result .cloned() .ok_or_else(|| QueryError::FunctionNotExists { name: name.to_string() }) } fn udwf(&self, name: &str) -> QueryResult<Arc<WindowUDF>> { let result = self.window_functions.get(&name.to_uppercase()); result .cloned() .ok_or_else(|| QueryError::FunctionNotExists { name: name.to_string() }) } fn udfs(&self) -> Vec<String> { self.scalar_functions.keys().cloned().collect() } fn udafs(&self) -> Vec<String> { self.aggregate_functions.keys().cloned().collect() } fn udwfs(&self) -> Vec<String> { self.window_functions.keys().cloned().collect() } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/function/mod.rs
crates/s3select-query/src/function/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. pub mod simple_func_manager;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/sql/dialect.rs
crates/s3select-query/src/sql/dialect.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use datafusion::sql::sqlparser::dialect::Dialect; #[derive(Debug, Default)] pub struct RustFsDialect; impl Dialect for RustFsDialect { fn is_identifier_start(&self, ch: char) -> bool { ch.is_alphabetic() || ch == '_' || ch == '#' || ch == '@' } fn is_identifier_part(&self, ch: char) -> bool { ch.is_alphabetic() || ch.is_ascii_digit() || ch == '@' || ch == '$' || ch == '#' || ch == '_' } fn supports_group_by_expr(&self) -> bool { true } } #[cfg(test)] mod tests { use super::*; #[test] fn test_rustfs_dialect_creation() { let _dialect = RustFsDialect; // Test that dialect can be created successfully assert!(std::mem::size_of::<RustFsDialect>() == 0, "Dialect should be zero-sized"); } #[test] fn test_rustfs_dialect_debug() { let dialect = RustFsDialect; let debug_str = format!("{dialect:?}"); assert!(!debug_str.is_empty(), "Debug output should not be empty"); assert!(debug_str.contains("RustFsDialect"), "Debug output should contain dialect name"); } #[test] fn test_is_identifier_start_alphabetic() { let dialect = RustFsDialect; // Test alphabetic characters assert!(dialect.is_identifier_start('a'), "Lowercase letter should be valid identifier start"); assert!(dialect.is_identifier_start('A'), "Uppercase letter should be valid identifier start"); assert!(dialect.is_identifier_start('z'), "Last lowercase letter should be valid identifier start"); assert!(dialect.is_identifier_start('Z'), "Last uppercase letter should be valid identifier start"); // Test Unicode alphabetic characters assert!(dialect.is_identifier_start('Ξ±'), "Greek letter should be valid identifier start"); assert!(dialect.is_identifier_start('ΓΌ'), "Unicode character should be valid identifier start"); assert!(dialect.is_identifier_start('Γ±'), "Accented letter should be valid identifier start"); } #[test] fn test_is_identifier_start_special_chars() { let dialect = RustFsDialect; // Test special characters that are allowed assert!(dialect.is_identifier_start('_'), "Underscore should be valid identifier start"); assert!(dialect.is_identifier_start('#'), "Hash should be valid identifier start"); assert!(dialect.is_identifier_start('@'), "At symbol should be valid identifier start"); } #[test] fn test_is_identifier_start_invalid_chars() { let dialect = RustFsDialect; // Test characters that should not be valid identifier starts assert!(!dialect.is_identifier_start('0'), "Digit should not be valid identifier start"); assert!(!dialect.is_identifier_start('9'), "Digit should not be valid identifier start"); assert!(!dialect.is_identifier_start('$'), "Dollar sign should not be valid identifier start"); assert!(!dialect.is_identifier_start(' '), "Space should not be valid identifier start"); assert!(!dialect.is_identifier_start('\t'), "Tab should not be valid identifier start"); assert!(!dialect.is_identifier_start('\n'), "Newline should not be valid identifier start"); assert!(!dialect.is_identifier_start('.'), "Dot should not be valid identifier start"); assert!(!dialect.is_identifier_start(','), "Comma should not be valid identifier start"); assert!(!dialect.is_identifier_start(';'), "Semicolon should not be valid identifier start"); assert!(!dialect.is_identifier_start('('), "Left paren should not be valid identifier start"); assert!(!dialect.is_identifier_start(')'), "Right paren should not be valid identifier start"); assert!(!dialect.is_identifier_start('['), "Left bracket should not be valid identifier start"); assert!(!dialect.is_identifier_start(']'), "Right bracket should not be valid identifier start"); assert!(!dialect.is_identifier_start('{'), "Left brace should not be valid identifier start"); assert!(!dialect.is_identifier_start('}'), "Right brace should not be valid identifier start"); assert!(!dialect.is_identifier_start('='), "Equals should not be valid identifier start"); assert!(!dialect.is_identifier_start('+'), "Plus should not be valid identifier start"); assert!(!dialect.is_identifier_start('-'), "Minus should not be valid identifier start"); assert!(!dialect.is_identifier_start('*'), "Asterisk should not be valid identifier start"); assert!(!dialect.is_identifier_start('/'), "Slash should not be valid identifier start"); assert!(!dialect.is_identifier_start('%'), "Percent should not be valid identifier start"); assert!(!dialect.is_identifier_start('<'), "Less than should not be valid identifier start"); assert!(!dialect.is_identifier_start('>'), "Greater than should not be valid identifier start"); assert!(!dialect.is_identifier_start('!'), "Exclamation should not be valid identifier start"); assert!(!dialect.is_identifier_start('?'), "Question mark should not be valid identifier start"); assert!(!dialect.is_identifier_start('&'), "Ampersand should not be valid identifier start"); assert!(!dialect.is_identifier_start('|'), "Pipe should not be valid identifier start"); assert!(!dialect.is_identifier_start('^'), "Caret should not be valid identifier start"); assert!(!dialect.is_identifier_start('~'), "Tilde should not be valid identifier start"); assert!(!dialect.is_identifier_start('`'), "Backtick should not be valid identifier start"); assert!(!dialect.is_identifier_start('"'), "Double quote should not be valid identifier start"); assert!(!dialect.is_identifier_start('\''), "Single quote should not be valid identifier start"); } #[test] fn test_is_identifier_part_alphabetic() { let dialect = RustFsDialect; // Test alphabetic characters assert!(dialect.is_identifier_part('a'), "Lowercase letter should be valid identifier part"); assert!(dialect.is_identifier_part('A'), "Uppercase letter should be valid identifier part"); assert!(dialect.is_identifier_part('z'), "Last lowercase letter should be valid identifier part"); assert!(dialect.is_identifier_part('Z'), "Last uppercase letter should be valid identifier part"); // Test Unicode alphabetic characters assert!(dialect.is_identifier_part('Ξ±'), "Greek letter should be valid identifier part"); assert!(dialect.is_identifier_part('ΓΌ'), "Unicode character should be valid identifier part"); assert!(dialect.is_identifier_part('Γ±'), "Accented letter should be valid identifier part"); } #[test] fn test_is_identifier_part_digits() { let dialect = RustFsDialect; // Test ASCII digits assert!(dialect.is_identifier_part('0'), "Digit 0 should be valid identifier part"); assert!(dialect.is_identifier_part('1'), "Digit 1 should be valid identifier part"); assert!(dialect.is_identifier_part('5'), "Digit 5 should be valid identifier part"); assert!(dialect.is_identifier_part('9'), "Digit 9 should be valid identifier part"); } #[test] fn test_is_identifier_part_special_chars() { let dialect = RustFsDialect; // Test special characters that are allowed assert!(dialect.is_identifier_part('_'), "Underscore should be valid identifier part"); assert!(dialect.is_identifier_part('#'), "Hash should be valid identifier part"); assert!(dialect.is_identifier_part('@'), "At symbol should be valid identifier part"); assert!(dialect.is_identifier_part('$'), "Dollar sign should be valid identifier part"); } #[test] fn test_is_identifier_part_invalid_chars() { let dialect = RustFsDialect; // Test characters that should not be valid identifier parts assert!(!dialect.is_identifier_part(' '), "Space should not be valid identifier part"); assert!(!dialect.is_identifier_part('\t'), "Tab should not be valid identifier part"); assert!(!dialect.is_identifier_part('\n'), "Newline should not be valid identifier part"); assert!(!dialect.is_identifier_part('.'), "Dot should not be valid identifier part"); assert!(!dialect.is_identifier_part(','), "Comma should not be valid identifier part"); assert!(!dialect.is_identifier_part(';'), "Semicolon should not be valid identifier part"); assert!(!dialect.is_identifier_part('('), "Left paren should not be valid identifier part"); assert!(!dialect.is_identifier_part(')'), "Right paren should not be valid identifier part"); assert!(!dialect.is_identifier_part('['), "Left bracket should not be valid identifier part"); assert!(!dialect.is_identifier_part(']'), "Right bracket should not be valid identifier part"); assert!(!dialect.is_identifier_part('{'), "Left brace should not be valid identifier part"); assert!(!dialect.is_identifier_part('}'), "Right brace should not be valid identifier part"); assert!(!dialect.is_identifier_part('='), "Equals should not be valid identifier part"); assert!(!dialect.is_identifier_part('+'), "Plus should not be valid identifier part"); assert!(!dialect.is_identifier_part('-'), "Minus should not be valid identifier part"); assert!(!dialect.is_identifier_part('*'), "Asterisk should not be valid identifier part"); assert!(!dialect.is_identifier_part('/'), "Slash should not be valid identifier part"); assert!(!dialect.is_identifier_part('%'), "Percent should not be valid identifier part"); assert!(!dialect.is_identifier_part('<'), "Less than should not be valid identifier part"); assert!(!dialect.is_identifier_part('>'), "Greater than should not be valid identifier part"); assert!(!dialect.is_identifier_part('!'), "Exclamation should not be valid identifier part"); assert!(!dialect.is_identifier_part('?'), "Question mark should not be valid identifier part"); assert!(!dialect.is_identifier_part('&'), "Ampersand should not be valid identifier part"); assert!(!dialect.is_identifier_part('|'), "Pipe should not be valid identifier part"); assert!(!dialect.is_identifier_part('^'), "Caret should not be valid identifier part"); assert!(!dialect.is_identifier_part('~'), "Tilde should not be valid identifier part"); assert!(!dialect.is_identifier_part('`'), "Backtick should not be valid identifier part"); assert!(!dialect.is_identifier_part('"'), "Double quote should not be valid identifier part"); assert!(!dialect.is_identifier_part('\''), "Single quote should not be valid identifier part"); } #[test] fn test_supports_group_by_expr() { let dialect = RustFsDialect; assert!(dialect.supports_group_by_expr(), "RustFsDialect should support GROUP BY expressions"); } #[test] fn test_identifier_validation_comprehensive() { let dialect = RustFsDialect; // Test valid identifier patterns let valid_starts = ['a', 'A', 'z', 'Z', '_', '#', '@', 'Ξ±', 'ΓΌ']; let valid_parts = ['a', 'A', '0', '9', '_', '#', '@', '$', 'Ξ±', 'ΓΌ']; for start_char in valid_starts { assert!( dialect.is_identifier_start(start_char), "Character '{start_char}' should be valid identifier start" ); for part_char in valid_parts { assert!( dialect.is_identifier_part(part_char), "Character '{part_char}' should be valid identifier part" ); } } } #[test] fn test_identifier_edge_cases() { let dialect = RustFsDialect; // Test edge cases with control characters assert!(!dialect.is_identifier_start('\0'), "Null character should not be valid identifier start"); assert!(!dialect.is_identifier_part('\0'), "Null character should not be valid identifier part"); assert!( !dialect.is_identifier_start('\x01'), "Control character should not be valid identifier start" ); assert!( !dialect.is_identifier_part('\x01'), "Control character should not be valid identifier part" ); assert!(!dialect.is_identifier_start('\x7F'), "DEL character should not be valid identifier start"); assert!(!dialect.is_identifier_part('\x7F'), "DEL character should not be valid identifier part"); } #[test] fn test_identifier_unicode_support() { let dialect = RustFsDialect; // Test various Unicode categories let unicode_letters = ['Ξ±', 'Ξ²', 'Ξ³', 'Ξ‘', 'Ξ’', 'Ξ“', 'Γ±', 'ΓΌ', 'Γ§', 'ΓΈ', 'Γ¦', 'ß']; for ch in unicode_letters { assert!(dialect.is_identifier_start(ch), "Unicode letter '{ch}' should be valid identifier start"); assert!(dialect.is_identifier_part(ch), "Unicode letter '{ch}' should be valid identifier part"); } } #[test] fn test_identifier_ascii_digits() { let dialect = RustFsDialect; // Test all ASCII digits for digit in '0'..='9' { assert!( !dialect.is_identifier_start(digit), "ASCII digit '{digit}' should not be valid identifier start" ); assert!(dialect.is_identifier_part(digit), "ASCII digit '{digit}' should be valid identifier part"); } } #[test] fn test_dialect_consistency() { let dialect = RustFsDialect; // Test that all valid identifier starts are also valid identifier parts let test_chars = [ 'a', 'A', 'z', 'Z', '_', '#', '@', 'Ξ±', 'ΓΌ', 'Γ±', '0', '9', '$', ' ', '.', ',', ';', '(', ')', '=', '+', '-', ]; for ch in test_chars { if dialect.is_identifier_start(ch) { assert!( dialect.is_identifier_part(ch), "Character '{ch}' that is valid identifier start should also be valid identifier part" ); } } } #[test] fn test_dialect_memory_efficiency() { let dialect = RustFsDialect; // Test that dialect doesn't use excessive memory let dialect_size = std::mem::size_of_val(&dialect); assert!(dialect_size < 100, "Dialect should not use excessive memory"); } #[test] fn test_dialect_trait_implementation() { let dialect = RustFsDialect; // Test that dialect properly implements the Dialect trait let dialect_ref: &dyn Dialect = &dialect; // Test basic functionality through trait assert!(dialect_ref.is_identifier_start('a'), "Trait method should work for valid start"); assert!(!dialect_ref.is_identifier_start('0'), "Trait method should work for invalid start"); assert!(dialect_ref.is_identifier_part('a'), "Trait method should work for valid part"); assert!(dialect_ref.is_identifier_part('0'), "Trait method should work for digit part"); assert!( dialect_ref.supports_group_by_expr(), "Trait method should return true for GROUP BY support" ); } #[test] fn test_dialect_clone_and_default() { let dialect1 = RustFsDialect; let dialect2 = RustFsDialect; // Test that multiple instances behave the same let test_chars = ['a', 'A', '0', '_', '#', '@', '$', ' ', '.']; for ch in test_chars { assert_eq!( dialect1.is_identifier_start(ch), dialect2.is_identifier_start(ch), "Different instances should behave the same for is_identifier_start" ); assert_eq!( dialect1.is_identifier_part(ch), dialect2.is_identifier_part(ch), "Different instances should behave the same for is_identifier_part" ); } assert_eq!( dialect1.supports_group_by_expr(), dialect2.supports_group_by_expr(), "Different instances should behave the same for supports_group_by_expr" ); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/sql/analyzer.rs
crates/s3select-query/src/sql/analyzer.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use datafusion::logical_expr::LogicalPlan; use datafusion::optimizer::analyzer::Analyzer as DFAnalyzer; use rustfs_s3select_api::QueryResult; use rustfs_s3select_api::query::analyzer::Analyzer; use rustfs_s3select_api::query::session::SessionCtx; pub struct DefaultAnalyzer { inner: DFAnalyzer, } impl DefaultAnalyzer { pub fn new() -> Self { let analyzer = DFAnalyzer::default(); // we can add analyzer rule at here Self { inner: analyzer } } } impl Default for DefaultAnalyzer { fn default() -> Self { Self::new() } } impl Analyzer for DefaultAnalyzer { fn analyze(&self, plan: &LogicalPlan, session: &SessionCtx) -> QueryResult<LogicalPlan> { let plan = self .inner .execute_and_check(plan.to_owned(), session.inner().config_options(), |_, _| {})?; Ok(plan) } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/sql/parser.rs
crates/s3select-query/src/sql/parser.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::{collections::VecDeque, fmt::Display}; use datafusion::sql::sqlparser::{ dialect::Dialect, parser::{Parser, ParserError}, tokenizer::{Token, Tokenizer}, }; use rustfs_s3select_api::{ ParserSnafu, query::{ast::ExtStatement, parser::Parser as RustFsParser}, }; use snafu::ResultExt; use super::dialect::RustFsDialect; pub type Result<T, E = ParserError> = std::result::Result<T, E>; // Use `Parser::expected` instead, if possible macro_rules! parser_err { ($MSG:expr) => { Err(ParserError::ParserError($MSG.to_string())) }; } #[derive(Default)] pub struct DefaultParser {} impl RustFsParser for DefaultParser { fn parse(&self, sql: &str) -> rustfs_s3select_api::QueryResult<VecDeque<ExtStatement>> { ExtParser::parse_sql(sql).context(ParserSnafu) } } /// SQL Parser pub struct ExtParser<'a> { parser: Parser<'a>, } impl<'a> ExtParser<'a> { /// Parse the specified tokens with dialect fn new_with_dialect(sql: &str, dialect: &'a dyn Dialect) -> Result<Self> { let mut tokenizer = Tokenizer::new(dialect, sql); let tokens = tokenizer.tokenize()?; Ok(ExtParser { parser: Parser::new(dialect).with_tokens(tokens), }) } /// Parse a SQL statement and produce a set of statements pub fn parse_sql(sql: &str) -> Result<VecDeque<ExtStatement>> { let dialect = &RustFsDialect {}; ExtParser::parse_sql_with_dialect(sql, dialect) } /// Parse a SQL statement and produce a set of statements pub fn parse_sql_with_dialect(sql: &str, dialect: &dyn Dialect) -> Result<VecDeque<ExtStatement>> { let mut parser = ExtParser::new_with_dialect(sql, dialect)?; let mut stmts = VecDeque::new(); let mut expecting_statement_delimiter = false; loop { // ignore empty statements (between successive statement delimiters) while parser.parser.consume_token(&Token::SemiColon) { expecting_statement_delimiter = false; } if parser.parser.peek_token() == Token::EOF { break; } if expecting_statement_delimiter { return parser.expected("end of statement", parser.parser.peek_token()); } let statement = parser.parse_statement()?; stmts.push_back(statement); expecting_statement_delimiter = true; } // debug!("Parser sql: {}, stmts: {:#?}", sql, stmts); Ok(stmts) } /// Parse a new expression fn parse_statement(&mut self) -> Result<ExtStatement> { Ok(ExtStatement::SqlStatement(Box::new(self.parser.parse_statement()?))) } // Report unexpected token fn expected<T>(&self, expected: &str, found: impl Display) -> Result<T> { parser_err!(format!("Expected {}, found: {}", expected, found)) } } #[cfg(test)] mod tests { use super::*; use rustfs_s3select_api::query::ast::ExtStatement; #[test] fn test_default_parser_creation() { let _parser = DefaultParser::default(); // Test that parser can be created successfully assert!(std::mem::size_of::<DefaultParser>() == 0, "Parser should be zero-sized"); } #[test] fn test_default_parser_simple_select() { let parser = DefaultParser::default(); let sql = "SELECT * FROM S3Object"; let result = parser.parse(sql); assert!(result.is_ok(), "Simple SELECT should parse successfully"); let statements = result.unwrap(); assert_eq!(statements.len(), 1, "Should have exactly one statement"); // Just verify we get a SQL statement without diving into AST details match &statements[0] { ExtStatement::SqlStatement(_) => { // Successfully parsed as SQL statement } } } #[test] fn test_default_parser_select_with_columns() { let parser = DefaultParser::default(); let sql = "SELECT id, name, age FROM S3Object"; let result = parser.parse(sql); assert!(result.is_ok(), "SELECT with columns should parse successfully"); let statements = result.unwrap(); assert_eq!(statements.len(), 1, "Should have exactly one statement"); match &statements[0] { ExtStatement::SqlStatement(_) => { // Successfully parsed as SQL statement } } } #[test] fn test_default_parser_select_with_where() { let parser = DefaultParser::default(); let sql = "SELECT * FROM S3Object WHERE age > 25"; let result = parser.parse(sql); assert!(result.is_ok(), "SELECT with WHERE should parse successfully"); let statements = result.unwrap(); assert_eq!(statements.len(), 1, "Should have exactly one statement"); match &statements[0] { ExtStatement::SqlStatement(_) => { // Successfully parsed as SQL statement } } } #[test] fn test_default_parser_multiple_statements() { let parser = DefaultParser::default(); let sql = "SELECT * FROM S3Object; SELECT id FROM S3Object;"; let result = parser.parse(sql); assert!(result.is_ok(), "Multiple statements should parse successfully"); let statements = result.unwrap(); assert_eq!(statements.len(), 2, "Should have exactly two statements"); } #[test] fn test_default_parser_empty_statements() { let parser = DefaultParser::default(); let sql = ";;; SELECT * FROM S3Object; ;;;"; let result = parser.parse(sql); assert!(result.is_ok(), "Empty statements should be ignored"); let statements = result.unwrap(); assert_eq!(statements.len(), 1, "Should have exactly one non-empty statement"); } #[test] fn test_default_parser_invalid_sql() { let parser = DefaultParser::default(); let sql = "INVALID SQL SYNTAX"; let result = parser.parse(sql); assert!(result.is_err(), "Invalid SQL should return error"); } #[test] fn test_default_parser_empty_sql() { let parser = DefaultParser::default(); let sql = ""; let result = parser.parse(sql); assert!(result.is_ok(), "Empty SQL should parse successfully"); let statements = result.unwrap(); assert!(statements.is_empty(), "Should have no statements"); } #[test] fn test_default_parser_whitespace_only() { let parser = DefaultParser::default(); let sql = " \n\t "; let result = parser.parse(sql); assert!(result.is_ok(), "Whitespace-only SQL should parse successfully"); let statements = result.unwrap(); assert!(statements.is_empty(), "Should have no statements"); } #[test] fn test_ext_parser_parse_sql() { let sql = "SELECT * FROM S3Object"; let result = ExtParser::parse_sql(sql); assert!(result.is_ok(), "ExtParser::parse_sql should work"); let statements = result.unwrap(); assert_eq!(statements.len(), 1, "Should have exactly one statement"); } #[test] fn test_ext_parser_parse_sql_with_dialect() { let sql = "SELECT * FROM S3Object"; let dialect = &RustFsDialect; let result = ExtParser::parse_sql_with_dialect(sql, dialect); assert!(result.is_ok(), "ExtParser::parse_sql_with_dialect should work"); let statements = result.unwrap(); assert_eq!(statements.len(), 1, "Should have exactly one statement"); } #[test] fn test_ext_parser_new_with_dialect() { let sql = "SELECT * FROM S3Object"; let dialect = &RustFsDialect; let result = ExtParser::new_with_dialect(sql, dialect); assert!(result.is_ok(), "ExtParser::new_with_dialect should work"); } #[test] fn test_ext_parser_complex_query() { let sql = "SELECT id, name, age FROM S3Object WHERE age > 25 AND department = 'IT' ORDER BY age DESC LIMIT 10"; let result = ExtParser::parse_sql(sql); assert!(result.is_ok(), "Complex query should parse successfully"); let statements = result.unwrap(); assert_eq!(statements.len(), 1, "Should have exactly one statement"); match &statements[0] { ExtStatement::SqlStatement(_) => { // Successfully parsed as SQL statement } } } #[test] fn test_ext_parser_aggregate_functions() { let sql = "SELECT COUNT(*), AVG(age), MAX(salary) FROM S3Object GROUP BY department"; let result = ExtParser::parse_sql(sql); assert!(result.is_ok(), "Aggregate functions should parse successfully"); let statements = result.unwrap(); assert_eq!(statements.len(), 1, "Should have exactly one statement"); match &statements[0] { ExtStatement::SqlStatement(_) => { // Successfully parsed as SQL statement } } } #[test] fn test_ext_parser_join_query() { let sql = "SELECT s1.id, s2.name FROM S3Object s1 JOIN S3Object s2 ON s1.id = s2.id"; let result = ExtParser::parse_sql(sql); assert!(result.is_ok(), "JOIN query should parse successfully"); let statements = result.unwrap(); assert_eq!(statements.len(), 1, "Should have exactly one statement"); } #[test] fn test_ext_parser_subquery() { let sql = "SELECT * FROM S3Object WHERE id IN (SELECT id FROM S3Object WHERE age > 30)"; let result = ExtParser::parse_sql(sql); assert!(result.is_ok(), "Subquery should parse successfully"); let statements = result.unwrap(); assert_eq!(statements.len(), 1, "Should have exactly one statement"); } #[test] fn test_ext_parser_case_insensitive() { let sql = "select * from s3object where age > 25"; let result = ExtParser::parse_sql(sql); assert!(result.is_ok(), "Case insensitive SQL should parse successfully"); let statements = result.unwrap(); assert_eq!(statements.len(), 1, "Should have exactly one statement"); } #[test] fn test_ext_parser_quoted_identifiers() { let sql = r#"SELECT "id", "name" FROM "S3Object" WHERE "age" > 25"#; let result = ExtParser::parse_sql(sql); assert!(result.is_ok(), "Quoted identifiers should parse successfully"); let statements = result.unwrap(); assert_eq!(statements.len(), 1, "Should have exactly one statement"); } #[test] fn test_ext_parser_string_literals() { let sql = "SELECT * FROM S3Object WHERE name = 'John Doe' AND department = 'IT'"; let result = ExtParser::parse_sql(sql); assert!(result.is_ok(), "String literals should parse successfully"); let statements = result.unwrap(); assert_eq!(statements.len(), 1, "Should have exactly one statement"); } #[test] fn test_ext_parser_numeric_literals() { let sql = "SELECT * FROM S3Object WHERE age = 25 AND salary = 50000.50"; let result = ExtParser::parse_sql(sql); assert!(result.is_ok(), "Numeric literals should parse successfully"); let statements = result.unwrap(); assert_eq!(statements.len(), 1, "Should have exactly one statement"); } #[test] fn test_ext_parser_error_handling() { let invalid_sqls = vec![ "SELECT FROM", // Missing column list "SELECT * FROM", // Missing table name "SELECT * FROM S3Object WHERE", // Incomplete WHERE clause "SELECT * FROM S3Object GROUP", // Incomplete GROUP BY "SELECT * FROM S3Object ORDER", // Incomplete ORDER BY ]; for sql in invalid_sqls { let result = ExtParser::parse_sql(sql); assert!(result.is_err(), "Invalid SQL '{sql}' should return error"); } } #[test] fn test_ext_parser_memory_efficiency() { let sql = "SELECT * FROM S3Object"; // Test that parser doesn't use excessive memory let result = ExtParser::parse_sql(sql); assert!(result.is_ok(), "Parser should work efficiently"); let statements = result.unwrap(); let memory_size = std::mem::size_of_val(&statements); assert!(memory_size < 10000, "Parsed statements should not use excessive memory"); } #[test] fn test_ext_parser_large_query() { // Test with a reasonably large query let mut sql = String::from("SELECT "); for i in 0..100 { if i > 0 { sql.push_str(", "); } sql.push_str(&format!("col{i}")); } sql.push_str(" FROM S3Object WHERE "); for i in 0..50 { if i > 0 { sql.push_str(" AND "); } sql.push_str(&format!("col{i} > {i}")); } let result = ExtParser::parse_sql(&sql); assert!(result.is_ok(), "Large query should parse successfully"); let statements = result.unwrap(); assert_eq!(statements.len(), 1, "Should have exactly one statement"); } #[test] fn test_parser_err_macro() { let error: Result<()> = parser_err!("Test error message"); assert!(error.is_err(), "parser_err! macro should create error"); match error { Err(ParserError::ParserError(msg)) => { assert_eq!(msg, "Test error message", "Error message should match"); } _ => panic!("Expected ParserError::ParserError"), } } #[test] fn test_ext_parser_expected_method() { let sql = "SELECT * FROM S3Object"; let dialect = &RustFsDialect; let parser = ExtParser::new_with_dialect(sql, dialect).unwrap(); let result: Result<()> = parser.expected("test token", "found token"); assert!(result.is_err(), "expected method should return error"); match result { Err(ParserError::ParserError(msg)) => { assert!(msg.contains("Expected test token"), "Error should contain expected message"); assert!(msg.contains("found: found token"), "Error should contain found message"); } _ => panic!("Expected ParserError::ParserError"), } } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/sql/optimizer.rs
crates/s3select-query/src/sql/optimizer.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::sync::Arc; use async_trait::async_trait; use datafusion::physical_plan::{ExecutionPlan, displayable}; use rustfs_s3select_api::{ QueryResult, query::{logical_planner::QueryPlan, optimizer::Optimizer, physical_planner::PhysicalPlanner, session::SessionCtx}, }; use tracing::debug; use super::{ logical::optimizer::{DefaultLogicalOptimizer, LogicalOptimizer}, physical::{optimizer::PhysicalOptimizer, planner::DefaultPhysicalPlanner}, }; pub struct CascadeOptimizer { logical_optimizer: Arc<dyn LogicalOptimizer + Send + Sync>, physical_planner: Arc<dyn PhysicalPlanner + Send + Sync>, physical_optimizer: Arc<dyn PhysicalOptimizer + Send + Sync>, } #[async_trait] impl Optimizer for CascadeOptimizer { async fn optimize(&self, plan: &QueryPlan, session: &SessionCtx) -> QueryResult<Arc<dyn ExecutionPlan>> { debug!("Original logical plan:\n{}\n", plan.df_plan.display_indent_schema(),); let optimized_logical_plan = self.logical_optimizer.optimize(plan, session)?; debug!("Final logical plan:\n{}\n", optimized_logical_plan.display_indent_schema(),); let physical_plan = { self.physical_planner .create_physical_plan(&optimized_logical_plan, session) .await? }; debug!("Original physical plan:\n{}\n", displayable(physical_plan.as_ref()).indent(false)); let optimized_physical_plan = { self.physical_optimizer.optimize(physical_plan, session)? }; Ok(optimized_physical_plan) } } #[derive(Default)] pub struct CascadeOptimizerBuilder { logical_optimizer: Option<Arc<dyn LogicalOptimizer + Send + Sync>>, physical_planner: Option<Arc<dyn PhysicalPlanner + Send + Sync>>, physical_optimizer: Option<Arc<dyn PhysicalOptimizer + Send + Sync>>, } impl CascadeOptimizerBuilder { pub fn with_logical_optimizer(mut self, logical_optimizer: Arc<dyn LogicalOptimizer + Send + Sync>) -> Self { self.logical_optimizer = Some(logical_optimizer); self } pub fn with_physical_planner(mut self, physical_planner: Arc<dyn PhysicalPlanner + Send + Sync>) -> Self { self.physical_planner = Some(physical_planner); self } pub fn with_physical_optimizer(mut self, physical_optimizer: Arc<dyn PhysicalOptimizer + Send + Sync>) -> Self { self.physical_optimizer = Some(physical_optimizer); self } pub fn build(self) -> CascadeOptimizer { let default_logical_optimizer = Arc::new(DefaultLogicalOptimizer::default()); let default_physical_planner = Arc::new(DefaultPhysicalPlanner::default()); let logical_optimizer = self.logical_optimizer.unwrap_or(default_logical_optimizer); let physical_planner = self.physical_planner.unwrap_or_else(|| default_physical_planner.clone()); let physical_optimizer = self.physical_optimizer.unwrap_or(default_physical_planner); CascadeOptimizer { logical_optimizer, physical_planner, physical_optimizer, } } } #[cfg(test)] mod tests { use super::*; #[test] fn test_cascade_optimizer_builder_default() { let _builder = CascadeOptimizerBuilder::default(); // Test that builder can be created successfully assert!( std::mem::size_of::<CascadeOptimizerBuilder>() > 0, "Builder should be created successfully" ); } #[test] fn test_cascade_optimizer_builder_build_with_defaults() { let _builder = CascadeOptimizerBuilder::default(); let optimizer = _builder.build(); // Test that optimizer can be built with default components assert!(std::mem::size_of_val(&optimizer) > 0, "Optimizer should be built successfully"); } #[test] fn test_cascade_optimizer_builder_basic_functionality() { // Test that builder methods can be called and return self let _builder = CascadeOptimizerBuilder::default(); // Test that we can call builder methods (even if we don't have mock implementations) // This tests the builder pattern itself assert!( std::mem::size_of::<CascadeOptimizerBuilder>() > 0, "Builder should be created successfully" ); } #[test] fn test_cascade_optimizer_builder_memory_efficiency() { let _builder = CascadeOptimizerBuilder::default(); // Test that builder doesn't use excessive memory let builder_size = std::mem::size_of_val(&_builder); assert!(builder_size < 1000, "Builder should not use excessive memory"); let optimizer = _builder.build(); let optimizer_size = std::mem::size_of_val(&optimizer); assert!(optimizer_size < 1000, "Optimizer should not use excessive memory"); } #[test] fn test_cascade_optimizer_builder_multiple_builds() { let _builder = CascadeOptimizerBuilder::default(); // Test that we can build multiple optimizers from the same configuration let optimizer1 = _builder.build(); assert!(std::mem::size_of_val(&optimizer1) > 0, "First optimizer should be built successfully"); // Note: builder is consumed by build(), so we can't build again from the same instance // This is the expected behavior } #[test] fn test_cascade_optimizer_builder_default_fallbacks() { let _builder = CascadeOptimizerBuilder::default(); let optimizer = _builder.build(); // Test that default components are used when none are specified // We can't directly access the internal components, but we can verify the optimizer was built assert!(std::mem::size_of_val(&optimizer) > 0, "Optimizer should use default components"); } #[test] fn test_cascade_optimizer_component_types() { let optimizer = CascadeOptimizerBuilder::default().build(); // Test that optimizer contains the expected component types // We can't directly access the components, but we can verify the optimizer structure assert!(std::mem::size_of_val(&optimizer) > 0, "Optimizer should contain components"); // The optimizer should have three Arc fields for the components // This is a basic structural test } #[test] fn test_cascade_optimizer_builder_consistency() { // Test that multiple builders with the same configuration produce equivalent optimizers let optimizer1 = CascadeOptimizerBuilder::default().build(); let optimizer2 = CascadeOptimizerBuilder::default().build(); // Both optimizers should be built successfully assert!(std::mem::size_of_val(&optimizer1) > 0, "First optimizer should be built"); assert!(std::mem::size_of_val(&optimizer2) > 0, "Second optimizer should be built"); // They should have the same memory footprint (same structure) assert_eq!( std::mem::size_of_val(&optimizer1), std::mem::size_of_val(&optimizer2), "Optimizers with same configuration should have same size" ); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/sql/mod.rs
crates/s3select-query/src/sql/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. pub mod analyzer; pub mod dialect; pub mod logical; pub mod optimizer; pub mod parser; pub mod physical; pub mod planner;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/sql/planner.rs
crates/s3select-query/src/sql/planner.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use async_recursion::async_recursion; use async_trait::async_trait; use datafusion::sql::{planner::SqlToRel, sqlparser::ast::Statement}; use rustfs_s3select_api::{ QueryError, QueryResult, query::{ ast::ExtStatement, logical_planner::{LogicalPlanner, Plan, QueryPlan}, session::SessionCtx, }, }; use crate::metadata::ContextProviderExtension; pub struct SqlPlanner<'a, S: ContextProviderExtension> { _schema_provider: &'a S, df_planner: SqlToRel<'a, S>, } #[async_trait] impl<S: ContextProviderExtension + Send + Sync> LogicalPlanner for SqlPlanner<'_, S> { async fn create_logical_plan(&self, statement: ExtStatement, session: &SessionCtx) -> QueryResult<Plan> { let plan = { self.statement_to_plan(statement, session).await? }; Ok(plan) } } impl<'a, S: ContextProviderExtension + Send + Sync + 'a> SqlPlanner<'a, S> { /// Create a new query planner pub fn new(schema_provider: &'a S) -> Self { SqlPlanner { _schema_provider: schema_provider, df_planner: SqlToRel::new(schema_provider), } } /// Generate a logical plan from an Extent SQL statement #[async_recursion] pub(crate) async fn statement_to_plan(&self, statement: ExtStatement, session: &SessionCtx) -> QueryResult<Plan> { match statement { ExtStatement::SqlStatement(stmt) => self.df_sql_to_plan(*stmt, session).await, } } async fn df_sql_to_plan(&self, stmt: Statement, _session: &SessionCtx) -> QueryResult<Plan> { match stmt { Statement::Query(_) => { let df_plan = self.df_planner.sql_statement_to_plan(stmt)?; let plan = Plan::Query(QueryPlan { df_plan, is_tag_scan: false, }); Ok(plan) } _ => Err(QueryError::NotImplemented { err: stmt.to_string() }), } } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/sql/physical/optimizer.rs
crates/s3select-query/src/sql/physical/optimizer.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::sync::Arc; use datafusion::physical_optimizer::PhysicalOptimizerRule; use datafusion::physical_plan::ExecutionPlan; use rustfs_s3select_api::QueryResult; use rustfs_s3select_api::query::session::SessionCtx; pub trait PhysicalOptimizer { fn optimize(&self, plan: Arc<dyn ExecutionPlan>, session: &SessionCtx) -> QueryResult<Arc<dyn ExecutionPlan>>; fn inject_optimizer_rule(&mut self, optimizer_rule: Arc<dyn PhysicalOptimizerRule + Send + Sync>); }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/sql/physical/mod.rs
crates/s3select-query/src/sql/physical/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. pub mod optimizer; pub mod planner;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/sql/physical/planner.rs
crates/s3select-query/src/sql/physical/planner.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::sync::Arc; use async_trait::async_trait; use datafusion::execution::SessionStateBuilder; use datafusion::logical_expr::LogicalPlan; use datafusion::physical_optimizer::PhysicalOptimizerRule; use datafusion::physical_optimizer::aggregate_statistics::AggregateStatistics; use datafusion::physical_optimizer::coalesce_batches::CoalesceBatches; use datafusion::physical_optimizer::join_selection::JoinSelection; use datafusion::physical_plan::ExecutionPlan; use datafusion::physical_planner::{ DefaultPhysicalPlanner as DFDefaultPhysicalPlanner, ExtensionPlanner, PhysicalPlanner as DFPhysicalPlanner, }; use rustfs_s3select_api::QueryResult; use rustfs_s3select_api::query::physical_planner::PhysicalPlanner; use rustfs_s3select_api::query::session::SessionCtx; use super::optimizer::PhysicalOptimizer; pub struct DefaultPhysicalPlanner { ext_physical_transform_rules: Vec<Arc<dyn ExtensionPlanner + Send + Sync>>, /// Responsible for optimizing a physical execution plan ext_physical_optimizer_rules: Vec<Arc<dyn PhysicalOptimizerRule + Send + Sync>>, } impl DefaultPhysicalPlanner { #[allow(dead_code)] fn with_physical_transform_rules(mut self, rules: Vec<Arc<dyn ExtensionPlanner + Send + Sync>>) -> Self { self.ext_physical_transform_rules = rules; self } } impl DefaultPhysicalPlanner { #[allow(dead_code)] fn with_optimizer_rules(mut self, rules: Vec<Arc<dyn PhysicalOptimizerRule + Send + Sync>>) -> Self { self.ext_physical_optimizer_rules = rules; self } } impl Default for DefaultPhysicalPlanner { fn default() -> Self { let ext_physical_transform_rules: Vec<Arc<dyn ExtensionPlanner + Send + Sync>> = vec![ // can add rules at here ]; // We need to take care of the rule ordering. They may influence each other. let ext_physical_optimizer_rules: Vec<Arc<dyn PhysicalOptimizerRule + Sync + Send>> = vec![ Arc::new(AggregateStatistics::new()), // Statistics-based join selection will change the Auto mode to a real join implementation, // like collect left, or hash join, or future sort merge join, which will influence the // EnforceDistribution and EnforceSorting rules as they decide whether to add additional // repartitioning and local sorting steps to meet distribution and ordering requirements. // Therefore, it should run before EnforceDistribution and EnforceSorting. Arc::new(JoinSelection::new()), // The CoalesceBatches rule will not influence the distribution and ordering of the // whole plan tree. Therefore, to avoid influencing other rules, it should run last. Arc::new(CoalesceBatches::new()), ]; Self { ext_physical_transform_rules, ext_physical_optimizer_rules, } } } #[async_trait] impl PhysicalPlanner for DefaultPhysicalPlanner { async fn create_physical_plan( &self, logical_plan: &LogicalPlan, session: &SessionCtx, ) -> QueryResult<Arc<dyn ExecutionPlan>> { // Inject extended physical plan optimization rules into df's session state let new_state = SessionStateBuilder::new_from_existing(session.inner().clone()) .with_physical_optimizer_rules(self.ext_physical_optimizer_rules.clone()) .build(); // Construct df's Physical Planner with extended physical plan transformation rules let planner = DFDefaultPhysicalPlanner::with_extension_planners(self.ext_physical_transform_rules.clone()); // Execute df's physical plan planning and optimization planner .create_physical_plan(logical_plan, &new_state) .await .map_err(|e| e.into()) } fn inject_physical_transform_rule(&mut self, rule: Arc<dyn ExtensionPlanner + Send + Sync>) { self.ext_physical_transform_rules.push(rule) } } impl PhysicalOptimizer for DefaultPhysicalPlanner { fn optimize(&self, plan: Arc<dyn ExecutionPlan>, _session: &SessionCtx) -> QueryResult<Arc<dyn ExecutionPlan>> { Ok(plan) } fn inject_optimizer_rule(&mut self, optimizer_rule: Arc<dyn PhysicalOptimizerRule + Send + Sync>) { self.ext_physical_optimizer_rules.push(optimizer_rule); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/sql/logical/optimizer.rs
crates/s3select-query/src/sql/logical/optimizer.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use std::sync::Arc; use datafusion::{ execution::SessionStateBuilder, logical_expr::LogicalPlan, optimizer::{ OptimizerRule, common_subexpr_eliminate::CommonSubexprEliminate, decorrelate_predicate_subquery::DecorrelatePredicateSubquery, eliminate_cross_join::EliminateCrossJoin, eliminate_duplicated_expr::EliminateDuplicatedExpr, eliminate_filter::EliminateFilter, eliminate_join::EliminateJoin, eliminate_limit::EliminateLimit, eliminate_outer_join::EliminateOuterJoin, extract_equijoin_predicate::ExtractEquijoinPredicate, filter_null_join_keys::FilterNullJoinKeys, propagate_empty_relation::PropagateEmptyRelation, push_down_filter::PushDownFilter, push_down_limit::PushDownLimit, replace_distinct_aggregate::ReplaceDistinctWithAggregate, scalar_subquery_to_join::ScalarSubqueryToJoin, simplify_expressions::SimplifyExpressions, single_distinct_to_groupby::SingleDistinctToGroupBy, }, }; use rustfs_s3select_api::{ QueryResult, query::{analyzer::AnalyzerRef, logical_planner::QueryPlan, session::SessionCtx}, }; use tracing::debug; use crate::sql::analyzer::DefaultAnalyzer; pub trait LogicalOptimizer: Send + Sync { fn optimize(&self, plan: &QueryPlan, session: &SessionCtx) -> QueryResult<LogicalPlan>; fn inject_optimizer_rule(&mut self, optimizer_rule: Arc<dyn OptimizerRule + Send + Sync>); } pub struct DefaultLogicalOptimizer { // fit datafusion // TODO refactor analyzer: AnalyzerRef, rules: Vec<Arc<dyn OptimizerRule + Send + Sync>>, } impl DefaultLogicalOptimizer { #[allow(dead_code)] fn with_optimizer_rules(mut self, rules: Vec<Arc<dyn OptimizerRule + Send + Sync>>) -> Self { self.rules = rules; self } } impl Default for DefaultLogicalOptimizer { fn default() -> Self { let analyzer = Arc::new(DefaultAnalyzer::default()); // additional optimizer rule let rules: Vec<Arc<dyn OptimizerRule + Send + Sync>> = vec![ // df default rules start Arc::new(SimplifyExpressions::new()), Arc::new(ReplaceDistinctWithAggregate::new()), Arc::new(EliminateJoin::new()), Arc::new(DecorrelatePredicateSubquery::new()), Arc::new(ScalarSubqueryToJoin::new()), Arc::new(ExtractEquijoinPredicate::new()), // simplify expressions does not simplify expressions in subqueries, so we // run it again after running the optimizations that potentially converted // subqueries to joins Arc::new(SimplifyExpressions::new()), Arc::new(EliminateDuplicatedExpr::new()), Arc::new(EliminateFilter::new()), Arc::new(EliminateCrossJoin::new()), Arc::new(CommonSubexprEliminate::new()), Arc::new(EliminateLimit::new()), Arc::new(PropagateEmptyRelation::new()), Arc::new(FilterNullJoinKeys::default()), Arc::new(EliminateOuterJoin::new()), // Filters can't be pushed down past Limits, we should do PushDownFilter after PushDownLimit Arc::new(PushDownLimit::new()), Arc::new(PushDownFilter::new()), Arc::new(SingleDistinctToGroupBy::new()), // The previous optimizations added expressions and projections, // that might benefit from the following rules Arc::new(SimplifyExpressions::new()), Arc::new(CommonSubexprEliminate::new()), // PushDownProjection can pushdown Projections through Limits, do PushDownLimit again. Arc::new(PushDownLimit::new()), // df default rules end // custom rules can add at here ]; Self { analyzer, rules } } } impl LogicalOptimizer for DefaultLogicalOptimizer { fn optimize(&self, plan: &QueryPlan, session: &SessionCtx) -> QueryResult<LogicalPlan> { let analyzed_plan = { self.analyzer.analyze(&plan.df_plan, session)? }; debug!("Analyzed logical plan:\n{}\n", plan.df_plan.display_indent_schema(),); let optimizeed_plan = { SessionStateBuilder::new_from_existing(session.inner().clone()) .with_optimizer_rules(self.rules.clone()) .build() .optimize(&analyzed_plan)? }; Ok(optimizeed_plan) } fn inject_optimizer_rule(&mut self, optimizer_rule: Arc<dyn OptimizerRule + Send + Sync>) { self.rules.push(optimizer_rule); } }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/sql/logical/mod.rs
crates/s3select-query/src/sql/logical/mod.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. pub mod optimizer; pub mod planner;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/s3select-query/src/sql/logical/planner.rs
crates/s3select-query/src/sql/logical/planner.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use crate::sql::planner::SqlPlanner; pub type DefaultLogicalPlanner<'a, S> = SqlPlanner<'a, S>;
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false
rustfs/rustfs
https://github.com/rustfs/rustfs/blob/666c0a9a38636eb6653dff7d9c98ff7122601ce2/crates/obs/src/config.rs
crates/obs/src/config.rs
// Copyright 2024 RustFS Team // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use rustfs_config::observability::{ DEFAULT_OBS_ENVIRONMENT_PRODUCTION, ENV_OBS_ENDPOINT, ENV_OBS_ENVIRONMENT, ENV_OBS_LOG_DIRECTORY, ENV_OBS_LOG_ENDPOINT, ENV_OBS_LOG_FILENAME, ENV_OBS_LOG_KEEP_FILES, ENV_OBS_LOG_ROTATION_SIZE_MB, ENV_OBS_LOG_ROTATION_TIME, ENV_OBS_LOG_STDOUT_ENABLED, ENV_OBS_LOGGER_LEVEL, ENV_OBS_METER_INTERVAL, ENV_OBS_METRIC_ENDPOINT, ENV_OBS_SAMPLE_RATIO, ENV_OBS_SERVICE_NAME, ENV_OBS_SERVICE_VERSION, ENV_OBS_TRACE_ENDPOINT, ENV_OBS_USE_STDOUT, }; use rustfs_config::{ APP_NAME, DEFAULT_LOG_KEEP_FILES, DEFAULT_LOG_LEVEL, DEFAULT_LOG_ROTATION_SIZE_MB, DEFAULT_LOG_ROTATION_TIME, DEFAULT_OBS_LOG_FILENAME, DEFAULT_OBS_LOG_STDOUT_ENABLED, ENVIRONMENT, METER_INTERVAL, SAMPLE_RATIO, SERVICE_VERSION, USE_STDOUT, }; use rustfs_utils::dirs::get_log_directory_to_string; use rustfs_utils::{get_env_bool, get_env_f64, get_env_opt_str, get_env_str, get_env_u64, get_env_usize}; use serde::{Deserialize, Serialize}; use std::env; /// Observability: OpenTelemetry configuration /// # Fields /// * `endpoint`: Endpoint for metric collection /// * `use_stdout`: Output to stdout /// * `sample_ratio`: Trace sampling ratio /// * `meter_interval`: Metric collection interval /// * `service_name`: Service name /// * `service_version`: Service version /// * `environment`: Environment /// * `logger_level`: Logger level /// * `local_logging_enabled`: Local logging enabled /// # Added flexi_logger related configurations /// * `log_directory`: Log file directory /// * `log_filename`: The name of the log file /// * `log_rotation_size_mb`: Log file size cut threshold (MB) /// * `log_rotation_time`: Logs are cut by time (Hour,Day,Minute,Second) /// * `log_keep_files`: Number of log files to be retained /// # Returns /// A new instance of OtelConfig /// /// # Example /// ```no_run /// use rustfs_obs::OtelConfig; /// /// let config = OtelConfig::new(); /// ``` #[derive(Debug, Deserialize, Serialize, Clone)] pub struct OtelConfig { pub endpoint: String, // Endpoint for metric collection pub trace_endpoint: Option<String>, // Endpoint for trace collection pub metric_endpoint: Option<String>, // Endpoint for metric collection pub log_endpoint: Option<String>, // Endpoint for log collection pub use_stdout: Option<bool>, // Output to stdout pub sample_ratio: Option<f64>, // Trace sampling ratio pub meter_interval: Option<u64>, // Metric collection interval pub service_name: Option<String>, // Service name pub service_version: Option<String>, // Service version pub environment: Option<String>, // Environment pub logger_level: Option<String>, // Logger level pub log_stdout_enabled: Option<bool>, // Stdout logging enabled // Added flexi_logger related configurations pub log_directory: Option<String>, // LOG FILE DIRECTORY pub log_filename: Option<String>, // The name of the log file pub log_rotation_size_mb: Option<u64>, // Log file size cut threshold (MB) pub log_rotation_time: Option<String>, // Logs are cut by time (Hour, Day,Minute, Second) pub log_keep_files: Option<usize>, // Number of log files to be retained } impl OtelConfig { /// Helper function: Extract observable configuration from environment variables pub fn extract_otel_config_from_env(endpoint: Option<String>) -> OtelConfig { let endpoint = if let Some(endpoint) = endpoint { if endpoint.is_empty() { env::var(ENV_OBS_ENDPOINT).unwrap_or_else(|_| "".to_string()) } else { endpoint } } else { env::var(ENV_OBS_ENDPOINT).unwrap_or_else(|_| "".to_string()) }; let mut use_stdout = get_env_bool(ENV_OBS_USE_STDOUT, USE_STDOUT); if endpoint.is_empty() { use_stdout = true; } OtelConfig { endpoint, trace_endpoint: get_env_opt_str(ENV_OBS_TRACE_ENDPOINT), metric_endpoint: get_env_opt_str(ENV_OBS_METRIC_ENDPOINT), log_endpoint: get_env_opt_str(ENV_OBS_LOG_ENDPOINT), use_stdout: Some(use_stdout), sample_ratio: Some(get_env_f64(ENV_OBS_SAMPLE_RATIO, SAMPLE_RATIO)), meter_interval: Some(get_env_u64(ENV_OBS_METER_INTERVAL, METER_INTERVAL)), service_name: Some(get_env_str(ENV_OBS_SERVICE_NAME, APP_NAME)), service_version: Some(get_env_str(ENV_OBS_SERVICE_VERSION, SERVICE_VERSION)), environment: Some(get_env_str(ENV_OBS_ENVIRONMENT, ENVIRONMENT)), logger_level: Some(get_env_str(ENV_OBS_LOGGER_LEVEL, DEFAULT_LOG_LEVEL)), log_stdout_enabled: Some(get_env_bool(ENV_OBS_LOG_STDOUT_ENABLED, DEFAULT_OBS_LOG_STDOUT_ENABLED)), log_directory: Some(get_log_directory_to_string(ENV_OBS_LOG_DIRECTORY)), log_filename: Some(get_env_str(ENV_OBS_LOG_FILENAME, DEFAULT_OBS_LOG_FILENAME)), log_rotation_size_mb: Some(get_env_u64(ENV_OBS_LOG_ROTATION_SIZE_MB, DEFAULT_LOG_ROTATION_SIZE_MB)), // Default to 100 MB log_rotation_time: Some(get_env_str(ENV_OBS_LOG_ROTATION_TIME, DEFAULT_LOG_ROTATION_TIME)), // Default to "Hour" log_keep_files: Some(get_env_usize(ENV_OBS_LOG_KEEP_FILES, DEFAULT_LOG_KEEP_FILES)), // Default to keeping 30 log files } } /// Create a new instance of OtelConfig with default values /// /// # Returns /// A new instance of OtelConfig /// /// # Example /// ```no_run /// use rustfs_obs::OtelConfig; /// /// let config = OtelConfig::new(); /// ``` pub fn new() -> Self { Self::extract_otel_config_from_env(None) } } /// Implement Default trait for OtelConfig /// This allows creating a default instance of OtelConfig using OtelConfig::default() /// which internally calls OtelConfig::new() /// /// # Example /// ```no_run /// use rustfs_obs::OtelConfig; /// /// let config = OtelConfig::default(); /// ``` impl Default for OtelConfig { fn default() -> Self { Self::new() } } /// Overall application configuration /// Add observability configuration /// /// Observability: OpenTelemetry configuration /// /// # Example /// ``` /// use rustfs_obs::AppConfig; /// /// let config = AppConfig::new_with_endpoint(None); /// ``` #[derive(Debug, Deserialize, Clone)] pub struct AppConfig { pub observability: OtelConfig, } impl AppConfig { /// Create a new instance of AppConfig with default values /// /// # Returns /// A new instance of AppConfig pub fn new() -> Self { Self { observability: OtelConfig::default(), } } /// Create a new instance of AppConfig with specified endpoint /// /// # Arguments /// * `endpoint` - An optional string representing the endpoint for metric collection /// /// # Returns /// A new instance of AppConfig /// /// # Example /// ```no_run /// use rustfs_obs::AppConfig; /// /// let config = AppConfig::new_with_endpoint(Some("http://localhost:4317".to_string())); /// ``` pub fn new_with_endpoint(endpoint: Option<String>) -> Self { Self { observability: OtelConfig::extract_otel_config_from_env(endpoint), } } } /// Implement Default trait for AppConfig /// This allows creating a default instance of AppConfig using AppConfig::default() /// which internally calls AppConfig::new() /// /// # Example /// ```no_run /// use rustfs_obs::AppConfig; /// /// let config = AppConfig::default(); /// ``` impl Default for AppConfig { fn default() -> Self { Self::new() } } /// Check if the current environment is production /// /// # Returns /// true if production, false otherwise /// pub fn is_production_environment() -> bool { get_env_str(ENV_OBS_ENVIRONMENT, ENVIRONMENT).eq_ignore_ascii_case(DEFAULT_OBS_ENVIRONMENT_PRODUCTION) }
rust
Apache-2.0
666c0a9a38636eb6653dff7d9c98ff7122601ce2
2026-01-04T15:42:12.458416Z
false