id
stringlengths
20
153
type
stringclasses
1 value
granularity
stringclasses
14 values
content
stringlengths
16
84.3k
metadata
dict
connector-service_snippet_-8421620542795988629_75_15
clm
snippet
// connector-service/backend/connector-integration/src/utils/xml_utils.rs // Return JSON as bytes Ok(Bytes::from(json_string.into_bytes())) } /// Flattens a nested JSON structure, extracting values from "$text" fields pub fn flatten_json_structure(json_value: Value) -> Value { let mut flattened = Map::ne...
{ "chunk": null, "crate": "connector-integration", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "sta...
connector-service_snippet_-8421620542795988629_75_30
clm
snippet
// connector-service/backend/connector-integration/src/utils/xml_utils.rs // Return JSON as bytes Ok(Bytes::from(json_string.into_bytes())) } /// Flattens a nested JSON structure, extracting values from "$text" fields pub fn flatten_json_structure(json_value: Value) -> Value { let mut flattened = Map::ne...
{ "chunk": null, "crate": "connector-integration", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "sta...
connector-service_snippet_-8421620542795988629_75_50
clm
snippet
// connector-service/backend/connector-integration/src/utils/xml_utils.rs // Return JSON as bytes Ok(Bytes::from(json_string.into_bytes())) } /// Flattens a nested JSON structure, extracting values from "$text" fields pub fn flatten_json_structure(json_value: Value) -> Value { let mut flattened = Map::ne...
{ "chunk": null, "crate": "connector-integration", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 44, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "sta...
connector-service_snippet_-8421620542795988629_100_15
clm
snippet
// connector-service/backend/connector-integration/src/utils/xml_utils.rs if let Some(text_value) = value_obj.get("$text") { // Extract the value from "$text" field flattened.insert(key.clone(), text_value.clone()); } else if value_obj.is_empty() ...
{ "chunk": null, "crate": "connector-integration", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "sta...
connector-service_snippet_-8421620542795988629_100_30
clm
snippet
// connector-service/backend/connector-integration/src/utils/xml_utils.rs if let Some(text_value) = value_obj.get("$text") { // Extract the value from "$text" field flattened.insert(key.clone(), text_value.clone()); } else if value_obj.is_empty() ...
{ "chunk": null, "crate": "connector-integration", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 19, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "sta...
connector-service_snippet_-8421620542795988629_100_50
clm
snippet
// connector-service/backend/connector-integration/src/utils/xml_utils.rs if let Some(text_value) = value_obj.get("$text") { // Extract the value from "$text" field flattened.insert(key.clone(), text_value.clone()); } else if value_obj.is_empty() ...
{ "chunk": null, "crate": "connector-integration", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 19, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "sta...
connector-service_snippet_5501410502799828961_0_15
clm
snippet
// connector-service/backend/tracing-kafka/src/lib.rs //! A Kafka tracing layer that integrates with the tracing ecosystem. //! //! This crate provides a simple way to send tracing logs to Kafka while maintaining //! consistent JSON formatting through the log_utils infrastructure. //! //! # Examples //! ```no_run //! ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_5501410502799828961_0_30
clm
snippet
// connector-service/backend/tracing-kafka/src/lib.rs //! A Kafka tracing layer that integrates with the tracing ecosystem. //! //! This crate provides a simple way to send tracing logs to Kafka while maintaining //! consistent JSON formatting through the log_utils infrastructure. //! //! # Examples //! ```no_run //! ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_5501410502799828961_0_50
clm
snippet
// connector-service/backend/tracing-kafka/src/lib.rs //! A Kafka tracing layer that integrates with the tracing ecosystem. //! //! This crate provides a simple way to send tracing logs to Kafka while maintaining //! consistent JSON formatting through the log_utils infrastructure. //! //! # Examples //! ```no_run //! ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_5501410502799828961_25_15
clm
snippet
// connector-service/backend/tracing-kafka/src/lib.rs //! //! ```no_run //! use tracing_kafka::KafkaWriter; //! use rdkafka::message::OwnedHeaders; //! //! let writer = KafkaWriter::new( //! vec!["localhost:9092".to_string()], //! "default-topic".to_string(), //! None, None, None, None, None, None //! ).ex...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_5501410502799828961_25_30
clm
snippet
// connector-service/backend/tracing-kafka/src/lib.rs //! //! ```no_run //! use tracing_kafka::KafkaWriter; //! use rdkafka::message::OwnedHeaders; //! //! let writer = KafkaWriter::new( //! vec!["localhost:9092".to_string()], //! "default-topic".to_string(), //! None, None, None, None, None, None //! ).ex...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_5501410502799828961_25_50
clm
snippet
// connector-service/backend/tracing-kafka/src/lib.rs //! //! ```no_run //! use tracing_kafka::KafkaWriter; //! use rdkafka::message::OwnedHeaders; //! //! let writer = KafkaWriter::new( //! vec!["localhost:9092".to_string()], //! "default-topic".to_string(), //! None, None, None, None, None, None //! ).ex...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 46, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_5501410502799828961_50_15
clm
snippet
// connector-service/backend/tracing-kafka/src/lib.rs pub mod builder; mod layer; mod writer; pub use layer::{KafkaLayer, KafkaLayerError}; pub use writer::{KafkaWriter, KafkaWriterError}; #[cfg(feature = "kafka-metrics")] mod metrics; /// Initializes the metrics for the tracing kafka. /// This function should be c...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_5501410502799828961_50_30
clm
snippet
// connector-service/backend/tracing-kafka/src/lib.rs pub mod builder; mod layer; mod writer; pub use layer::{KafkaLayer, KafkaLayerError}; pub use writer::{KafkaWriter, KafkaWriterError}; #[cfg(feature = "kafka-metrics")] mod metrics; /// Initializes the metrics for the tracing kafka. /// This function should be c...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 21, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_5501410502799828961_50_50
clm
snippet
// connector-service/backend/tracing-kafka/src/lib.rs pub mod builder; mod layer; mod writer; pub use layer::{KafkaLayer, KafkaLayerError}; pub use writer::{KafkaWriter, KafkaWriterError}; #[cfg(feature = "kafka-metrics")] mod metrics; /// Initializes the metrics for the tracing kafka. /// This function should be c...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 21, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_0_15
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs //! Prometheus metrics for Kafka writer use std::sync::LazyLock; use prometheus::{register_int_counter, register_int_gauge, IntCounter, IntGauge}; /// Total number of logs successfully sent to Kafka #[allow(clippy::expect_used)] pub static KAFKA_LOGS_SENT: L...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_0_30
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs //! Prometheus metrics for Kafka writer use std::sync::LazyLock; use prometheus::{register_int_counter, register_int_gauge, IntCounter, IntGauge}; /// Total number of logs successfully sent to Kafka #[allow(clippy::expect_used)] pub static KAFKA_LOGS_SENT: L...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_0_50
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs //! Prometheus metrics for Kafka writer use std::sync::LazyLock; use prometheus::{register_int_counter, register_int_gauge, IntCounter, IntGauge}; /// Total number of logs successfully sent to Kafka #[allow(clippy::expect_used)] pub static KAFKA_LOGS_SENT: L...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_25_15
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs /// Current size of Kafka producer queue #[allow(clippy::expect_used)] pub static KAFKA_QUEUE_SIZE: LazyLock<IntGauge> = LazyLock::new(|| { register_int_gauge!( "kafka_producer_queue_size", "Current size of Kafka producer queue" ) ....
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_25_30
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs /// Current size of Kafka producer queue #[allow(clippy::expect_used)] pub static KAFKA_QUEUE_SIZE: LazyLock<IntGauge> = LazyLock::new(|| { register_int_gauge!( "kafka_producer_queue_size", "Current size of Kafka producer queue" ) ....
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_25_50
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs /// Current size of Kafka producer queue #[allow(clippy::expect_used)] pub static KAFKA_QUEUE_SIZE: LazyLock<IntGauge> = LazyLock::new(|| { register_int_gauge!( "kafka_producer_queue_size", "Current size of Kafka producer queue" ) ....
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_50_15
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs "kafka_drops_msg_too_large_total", "Total number of logs dropped due to message size exceeding limit" ) .expect("Failed to register kafka_drops_msg_too_large_total metric") }); /// Logs dropped due to timeout #[allow(clippy::expect_used...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_50_30
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs "kafka_drops_msg_too_large_total", "Total number of logs dropped due to message size exceeding limit" ) .expect("Failed to register kafka_drops_msg_too_large_total metric") }); /// Logs dropped due to timeout #[allow(clippy::expect_used...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_50_50
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs "kafka_drops_msg_too_large_total", "Total number of logs dropped due to message size exceeding limit" ) .expect("Failed to register kafka_drops_msg_too_large_total metric") }); /// Logs dropped due to timeout #[allow(clippy::expect_used...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_75_15
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs /// Total number of audit events successfully sent to Kafka #[allow(clippy::expect_used)] pub static KAFKA_AUDIT_EVENTS_SENT: LazyLock<IntCounter> = LazyLock::new(|| { register_int_counter!( "kafka_audit_events_sent_total", "Total number of...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_75_30
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs /// Total number of audit events successfully sent to Kafka #[allow(clippy::expect_used)] pub static KAFKA_AUDIT_EVENTS_SENT: LazyLock<IntCounter> = LazyLock::new(|| { register_int_counter!( "kafka_audit_events_sent_total", "Total number of...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_75_50
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs /// Total number of audit events successfully sent to Kafka #[allow(clippy::expect_used)] pub static KAFKA_AUDIT_EVENTS_SENT: LazyLock<IntCounter> = LazyLock::new(|| { register_int_counter!( "kafka_audit_events_sent_total", "Total number of...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_100_15
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs "kafka_audit_event_queue_size", "Current size of Kafka audit event producer queue" ) .expect("Failed to register kafka_audit_event_queue_size metric") }); /// Audit events dropped due to queue full #[allow(clippy::expect_used)] pub stat...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_100_30
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs "kafka_audit_event_queue_size", "Current size of Kafka audit event producer queue" ) .expect("Failed to register kafka_audit_event_queue_size metric") }); /// Audit events dropped due to queue full #[allow(clippy::expect_used)] pub stat...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_100_50
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs "kafka_audit_event_queue_size", "Current size of Kafka audit event producer queue" ) .expect("Failed to register kafka_audit_event_queue_size metric") }); /// Audit events dropped due to queue full #[allow(clippy::expect_used)] pub stat...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_125_15
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs /// Audit events dropped due to timeout #[allow(clippy::expect_used)] pub static KAFKA_AUDIT_DROPS_TIMEOUT: LazyLock<IntCounter> = LazyLock::new(|| { register_int_counter!( "kafka_audit_drops_timeout_total", "Total number of audit events dr...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_125_30
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs /// Audit events dropped due to timeout #[allow(clippy::expect_used)] pub static KAFKA_AUDIT_DROPS_TIMEOUT: LazyLock<IntCounter> = LazyLock::new(|| { register_int_counter!( "kafka_audit_drops_timeout_total", "Total number of audit events dr...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_125_50
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs /// Audit events dropped due to timeout #[allow(clippy::expect_used)] pub static KAFKA_AUDIT_DROPS_TIMEOUT: LazyLock<IntCounter> = LazyLock::new(|| { register_int_counter!( "kafka_audit_drops_timeout_total", "Total number of audit events dr...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 44, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_150_15
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs /// metric name), the application will panic immediately. #[cfg(feature = "kafka-metrics")] pub fn initialize_all_metrics() { // Force evaluation of all lazy metrics to fail fast if registration fails. let _ = &*KAFKA_LOGS_SENT; let _ = &*KAFKA_LOGS...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_150_30
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs /// metric name), the application will panic immediately. #[cfg(feature = "kafka-metrics")] pub fn initialize_all_metrics() { // Force evaluation of all lazy metrics to fail fast if registration fails. let _ = &*KAFKA_LOGS_SENT; let _ = &*KAFKA_LOGS...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 19, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_863436229918585844_150_50
clm
snippet
// connector-service/backend/tracing-kafka/src/metrics.rs /// metric name), the application will panic immediately. #[cfg(feature = "kafka-metrics")] pub fn initialize_all_metrics() { // Force evaluation of all lazy metrics to fail fast if registration fails. let _ = &*KAFKA_LOGS_SENT; let _ = &*KAFKA_LOGS...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 19, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_0_15
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs //! Kafka writer implementation for sending formatted log messages to Kafka. use std::{ io::{self, Write}, sync::Arc, time::Duration, }; use rdkafka::{ config::ClientConfig, error::{KafkaError, RDKafkaErrorCode}, message::OwnedHeaders, ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_0_30
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs //! Kafka writer implementation for sending formatted log messages to Kafka. use std::{ io::{self, Write}, sync::Arc, time::Duration, }; use rdkafka::{ config::ClientConfig, error::{KafkaError, RDKafkaErrorCode}, message::OwnedHeaders, ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_0_50
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs //! Kafka writer implementation for sending formatted log messages to Kafka. use std::{ io::{self, Write}, sync::Arc, time::Duration, }; use rdkafka::{ config::ClientConfig, error::{KafkaError, RDKafkaErrorCode}, message::OwnedHeaders, ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_25_15
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs /// A `ProducerContext` that handles delivery callbacks to increment metrics. #[derive(Clone)] struct MetricsProducerContext; impl ClientContext for MetricsProducerContext {} impl ProducerContext for MetricsProducerContext { type DeliveryOpaque = Box<Kafka...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_25_30
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs /// A `ProducerContext` that handles delivery callbacks to increment metrics. #[derive(Clone)] struct MetricsProducerContext; impl ClientContext for MetricsProducerContext {} impl ProducerContext for MetricsProducerContext { type DeliveryOpaque = Box<Kafka...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_25_50
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs /// A `ProducerContext` that handles delivery callbacks to increment metrics. #[derive(Clone)] struct MetricsProducerContext; impl ClientContext for MetricsProducerContext {} impl ProducerContext for MetricsProducerContext { type DeliveryOpaque = Box<Kafka...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_50_15
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs match (message_type, &kafka_error) { ( KafkaMessageType::Event, KafkaError::MessageProduction(RDKafkaErrorCode::QueueFull), ) => { KAFKA_AUDIT_DROPS_QUEUE_FUL...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_50_30
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs match (message_type, &kafka_error) { ( KafkaMessageType::Event, KafkaError::MessageProduction(RDKafkaErrorCode::QueueFull), ) => { KAFKA_AUDIT_DROPS_QUEUE_FUL...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_50_50
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs match (message_type, &kafka_error) { ( KafkaMessageType::Event, KafkaError::MessageProduction(RDKafkaErrorCode::QueueFull), ) => { KAFKA_AUDIT_DROPS_QUEUE_FUL...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_75_15
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs ) => { KAFKA_DROPS_QUEUE_FULL.inc(); } ( KafkaMessageType::Log, KafkaError::MessageProduction(RDKafkaErrorCode::MessageSizeTooLarge), ) =>...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_75_30
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs ) => { KAFKA_DROPS_QUEUE_FULL.inc(); } ( KafkaMessageType::Log, KafkaError::MessageProduction(RDKafkaErrorCode::MessageSizeTooLarge), ) =>...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_75_50
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs ) => { KAFKA_DROPS_QUEUE_FULL.inc(); } ( KafkaMessageType::Log, KafkaError::MessageProduction(RDKafkaErrorCode::MessageSizeTooLarge), ) =>...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_100_15
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs enum KafkaMessageType { Event, Log, } /// Kafka writer that implements std::io::Write for seamless integration with tracing #[derive(Clone)] pub struct KafkaWriter { producer: Arc<ThreadedProducer<MetricsProducerContext>>, topic: String, } impl...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_100_30
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs enum KafkaMessageType { Event, Log, } /// Kafka writer that implements std::io::Write for seamless integration with tracing #[derive(Clone)] pub struct KafkaWriter { producer: Arc<ThreadedProducer<MetricsProducerContext>>, topic: String, } impl...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_100_50
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs enum KafkaMessageType { Event, Log, } /// Kafka writer that implements std::io::Write for seamless integration with tracing #[derive(Clone)] pub struct KafkaWriter { producer: Arc<ThreadedProducer<MetricsProducerContext>>, topic: String, } impl...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_125_15
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs topic: String, batch_size: Option<usize>, linger_ms: Option<u64>, queue_buffering_max_messages: Option<usize>, queue_buffering_max_kbytes: Option<usize>, reconnect_backoff_min_ms: Option<u64>, reconnect_bac...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_125_30
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs topic: String, batch_size: Option<usize>, linger_ms: Option<u64>, queue_buffering_max_messages: Option<usize>, queue_buffering_max_kbytes: Option<usize>, reconnect_backoff_min_ms: Option<u64>, reconnect_bac...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_125_50
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs topic: String, batch_size: Option<usize>, linger_ms: Option<u64>, queue_buffering_max_messages: Option<usize>, queue_buffering_max_kbytes: Option<usize>, reconnect_backoff_min_ms: Option<u64>, reconnect_bac...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_150_15
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs } if let Some(ms) = linger_ms { config.set("linger.ms", ms.to_string()); } let producer: ThreadedProducer<MetricsProducerContext> = config .create_with_context(MetricsProducerContext) .map_err(...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_150_30
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs } if let Some(ms) = linger_ms { config.set("linger.ms", ms.to_string()); } let producer: ThreadedProducer<MetricsProducerContext> = config .create_with_context(MetricsProducerContext) .map_err(...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_150_50
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs } if let Some(ms) = linger_ms { config.set("linger.ms", ms.to_string()); } let producer: ThreadedProducer<MetricsProducerContext> = config .create_with_context(MetricsProducerContext) .map_err(...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_175_15
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs key: Option<&str>, payload: &[u8], headers: Option<OwnedHeaders>, ) -> Result<(), KafkaError> { #[cfg(feature = "kafka-metrics")] { let queue_size = self.producer.in_flight_count(); KAFKA_AUDIT_...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_175_30
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs key: Option<&str>, payload: &[u8], headers: Option<OwnedHeaders>, ) -> Result<(), KafkaError> { #[cfg(feature = "kafka-metrics")] { let queue_size = self.producer.in_flight_count(); KAFKA_AUDIT_...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_175_50
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs key: Option<&str>, payload: &[u8], headers: Option<OwnedHeaders>, ) -> Result<(), KafkaError> { #[cfg(feature = "kafka-metrics")] { let queue_size = self.producer.in_flight_count(); KAFKA_AUDIT_...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_200_15
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs } match self.producer.send(record) { Ok(_) => Ok(()), Err((kafka_error, _)) => { #[cfg(feature = "kafka-metrics")] { KAFKA_AUDIT_EVENTS_DROPPED.inc(); ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_200_30
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs } match self.producer.send(record) { Ok(_) => Ok(()), Err((kafka_error, _)) => { #[cfg(feature = "kafka-metrics")] { KAFKA_AUDIT_EVENTS_DROPPED.inc(); ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_200_50
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs } match self.producer.send(record) { Ok(_) => Ok(()), Err((kafka_error, _)) => { #[cfg(feature = "kafka-metrics")] { KAFKA_AUDIT_EVENTS_DROPPED.inc(); ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_225_15
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs pub fn builder() -> crate::builder::KafkaWriterBuilder { crate::builder::KafkaWriterBuilder::new() } } impl Write for KafkaWriter { fn write(&mut self, buf: &[u8]) -> io::Result<usize> { #[cfg(feature = "kafka-metrics")] { ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_225_30
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs pub fn builder() -> crate::builder::KafkaWriterBuilder { crate::builder::KafkaWriterBuilder::new() } } impl Write for KafkaWriter { fn write(&mut self, buf: &[u8]) -> io::Result<usize> { #[cfg(feature = "kafka-metrics")] { ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_225_50
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs pub fn builder() -> crate::builder::KafkaWriterBuilder { crate::builder::KafkaWriterBuilder::new() } } impl Write for KafkaWriter { fn write(&mut self, buf: &[u8]) -> io::Result<usize> { #[cfg(feature = "kafka-metrics")] { ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_250_15
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs KAFKA_LOGS_DROPPED.inc(); match &kafka_error { KafkaError::MessageProduction(RDKafkaErrorCode::QueueFull) => { KAFKA_DROPS_QUEUE_FULL.inc(); } _ ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_250_30
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs KAFKA_LOGS_DROPPED.inc(); match &kafka_error { KafkaError::MessageProduction(RDKafkaErrorCode::QueueFull) => { KAFKA_DROPS_QUEUE_FULL.inc(); } _ ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_250_50
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs KAFKA_LOGS_DROPPED.inc(); match &kafka_error { KafkaError::MessageProduction(RDKafkaErrorCode::QueueFull) => { KAFKA_DROPS_QUEUE_FULL.inc(); } _ ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_275_15
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs /// Errors that can occur when creating or using a KafkaWriter. #[derive(Debug, thiserror::Error)] pub enum KafkaWriterError { #[error("Failed to create Kafka producer: {0}")] ProducerCreation(KafkaError), #[error("Failed to fetch Kafka metadata: {0}...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_275_30
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs /// Errors that can occur when creating or using a KafkaWriter. #[derive(Debug, thiserror::Error)] pub enum KafkaWriterError { #[error("Failed to create Kafka producer: {0}")] ProducerCreation(KafkaError), #[error("Failed to fetch Kafka metadata: {0}...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-8840925468162898005_275_50
clm
snippet
// connector-service/backend/tracing-kafka/src/writer.rs /// Errors that can occur when creating or using a KafkaWriter. #[derive(Debug, thiserror::Error)] pub enum KafkaWriterError { #[error("Failed to create Kafka producer: {0}")] ProducerCreation(KafkaError), #[error("Failed to fetch Kafka metadata: {0}...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_0_15
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs //! Kafka layer implementation that reuses log_utils formatting. use std::{ collections::{HashMap, HashSet}, time::Duration, }; use log_utils::{ AdditionalFieldsPlacement, JsonFormattingLayer, JsonFormattingLayerConfig, LoggerError, }; use tracing::...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_0_30
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs //! Kafka layer implementation that reuses log_utils formatting. use std::{ collections::{HashMap, HashSet}, time::Duration, }; use log_utils::{ AdditionalFieldsPlacement, JsonFormattingLayer, JsonFormattingLayerConfig, LoggerError, }; use tracing::...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_0_50
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs //! Kafka layer implementation that reuses log_utils formatting. use std::{ collections::{HashMap, HashSet}, time::Duration, }; use log_utils::{ AdditionalFieldsPlacement, JsonFormattingLayer, JsonFormattingLayerConfig, LoggerError, }; use tracing::...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_25_15
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs impl KafkaLayer { /// Creates a new builder for configuring a KafkaLayer. pub fn builder() -> KafkaLayerBuilder { KafkaLayerBuilder::new() } /// Creates a new KafkaLayer from a pre-configured KafkaWriter. /// This is primarily used in...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_25_30
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs impl KafkaLayer { /// Creates a new builder for configuring a KafkaLayer. pub fn builder() -> KafkaLayerBuilder { KafkaLayerBuilder::new() } /// Creates a new KafkaLayer from a pre-configured KafkaWriter. /// This is primarily used in...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_25_50
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs impl KafkaLayer { /// Creates a new builder for configuring a KafkaLayer. pub fn builder() -> KafkaLayerBuilder { KafkaLayerBuilder::new() } /// Creates a new KafkaLayer from a pre-configured KafkaWriter. /// This is primarily used in...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_50_15
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs impl<S> Layer<S> for KafkaLayer where S: Subscriber + for<'lookup> tracing_subscriber::registry::LookupSpan<'lookup>, { fn on_event(&self, event: &tracing::Event<'_>, ctx: tracing_subscriber::layer::Context<'_, S>) { self.inner.on_event(event, ct...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_50_30
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs impl<S> Layer<S> for KafkaLayer where S: Subscriber + for<'lookup> tracing_subscriber::registry::LookupSpan<'lookup>, { fn on_event(&self, event: &tracing::Event<'_>, ctx: tracing_subscriber::layer::Context<'_, S>) { self.inner.on_event(event, ct...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_50_50
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs impl<S> Layer<S> for KafkaLayer where S: Subscriber + for<'lookup> tracing_subscriber::registry::LookupSpan<'lookup>, { fn on_event(&self, event: &tracing::Event<'_>, ctx: tracing_subscriber::layer::Context<'_, S>) { self.inner.on_event(event, ct...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_75_15
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs fn on_close(&self, id: tracing::span::Id, ctx: tracing_subscriber::layer::Context<'_, S>) { self.inner.on_close(id, ctx); } } impl KafkaLayer { /// Boxes the layer, making it easier to compose with other layers. pub fn boxed<S>(self) -> ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_75_30
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs fn on_close(&self, id: tracing::span::Id, ctx: tracing_subscriber::layer::Context<'_, S>) { self.inner.on_close(id, ctx); } } impl KafkaLayer { /// Boxes the layer, making it easier to compose with other layers. pub fn boxed<S>(self) -> ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_75_50
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs fn on_close(&self, id: tracing::span::Id, ctx: tracing_subscriber::layer::Context<'_, S>) { self.inner.on_close(id, ctx); } } impl KafkaLayer { /// Boxes the layer, making it easier to compose with other layers. pub fn boxed<S>(self) -> ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_100_15
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs #[error("Missing brokers configuration")] MissingBrokers, #[error("Missing topic configuration")] MissingTopic, } /// Builder for creating a KafkaLayer with custom configuration. #[derive(Debug, Clone, Default)] pub struct KafkaLayerBuilder { ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_100_30
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs #[error("Missing brokers configuration")] MissingBrokers, #[error("Missing topic configuration")] MissingTopic, } /// Builder for creating a KafkaLayer with custom configuration. #[derive(Debug, Clone, Default)] pub struct KafkaLayerBuilder { ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_100_50
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs #[error("Missing brokers configuration")] MissingBrokers, #[error("Missing topic configuration")] MissingTopic, } /// Builder for creating a KafkaLayer with custom configuration. #[derive(Debug, Clone, Default)] pub struct KafkaLayerBuilder { ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_125_15
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs .brokers(brokers.iter().map(|s| s.to_string()).collect()); self } /// Sets the Kafka topic to send logs to. pub fn topic(mut self, topic: impl Into<String>) -> Self { self.writer_builder = self.writer_builder.topic(topic);...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_125_30
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs .brokers(brokers.iter().map(|s| s.to_string()).collect()); self } /// Sets the Kafka topic to send logs to. pub fn topic(mut self, topic: impl Into<String>) -> Self { self.writer_builder = self.writer_builder.topic(topic);...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_125_50
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs .brokers(brokers.iter().map(|s| s.to_string()).collect()); self } /// Sets the Kafka topic to send logs to. pub fn topic(mut self, topic: impl Into<String>) -> Self { self.writer_builder = self.writer_builder.topic(topic);...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_150_15
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs self } /// Sets the maximum number of messages to buffer in the producer's queue. pub fn queue_buffering_max_messages(mut self, size: usize) -> Self { self.writer_builder = self.writer_builder.queue_buffering_max_messages(size); ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_150_30
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs self } /// Sets the maximum number of messages to buffer in the producer's queue. pub fn queue_buffering_max_messages(mut self, size: usize) -> Self { self.writer_builder = self.writer_builder.queue_buffering_max_messages(size); ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_150_50
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs self } /// Sets the maximum number of messages to buffer in the producer's queue. pub fn queue_buffering_max_messages(mut self, size: usize) -> Self { self.writer_builder = self.writer_builder.queue_buffering_max_messages(size); ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 40, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_175_15
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs self } /// Adds a single static field that will be included in every log entry. pub fn add_static_field(mut self, key: String, value: serde_json::Value) -> Self { self.static_fields.insert(key, value); self } /// Buil...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_175_30
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs self } /// Adds a single static field that will be included in every log entry. pub fn add_static_field(mut self, key: String, value: serde_json::Value) -> Self { self.static_fields.insert(key, value); self } /// Buil...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-6998519293309712771_175_50
clm
snippet
// connector-service/backend/tracing-kafka/src/layer.rs self } /// Adds a single static field that will be included in every log entry. pub fn add_static_field(mut self, key: String, value: serde_json::Value) -> Self { self.static_fields.insert(key, value); self } /// Buil...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-7694171944301536872_0_15
clm
snippet
// connector-service/backend/tracing-kafka/src/builder.rs //! Builder pattern implementation for KafkaWriter use std::time::Duration; use super::writer::{KafkaWriter, KafkaWriterError}; /// Builder for creating a KafkaWriter with custom configuration #[derive(Debug, Clone, Default)] pub struct KafkaWriterBuilder { ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-7694171944301536872_0_30
clm
snippet
// connector-service/backend/tracing-kafka/src/builder.rs //! Builder pattern implementation for KafkaWriter use std::time::Duration; use super::writer::{KafkaWriter, KafkaWriterError}; /// Builder for creating a KafkaWriter with custom configuration #[derive(Debug, Clone, Default)] pub struct KafkaWriterBuilder { ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 30, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-7694171944301536872_0_50
clm
snippet
// connector-service/backend/tracing-kafka/src/builder.rs //! Builder pattern implementation for KafkaWriter use std::time::Duration; use super::writer::{KafkaWriter, KafkaWriterError}; /// Builder for creating a KafkaWriter with custom configuration #[derive(Debug, Clone, Default)] pub struct KafkaWriterBuilder { ...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 50, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...
connector-service_snippet_-7694171944301536872_25_15
clm
snippet
// connector-service/backend/tracing-kafka/src/builder.rs /// Sets the Kafka brokers to connect to pub fn brokers(mut self, brokers: Vec<String>) -> Self { self.brokers = Some(brokers); self } /// Sets the Kafka topic to send logs to pub fn topic(mut self, topic: impl Into<String>)...
{ "chunk": null, "crate": "tracing-kafka", "enum_name": null, "file_size": null, "for_type": null, "function_name": null, "is_async": null, "is_pub": null, "lines": 15, "method_name": null, "num_enums": null, "num_items": null, "num_structs": null, "repo": "connector-service", "start_line"...