id stringlengths 20 153 | type stringclasses 1
value | granularity stringclasses 14
values | content stringlengths 16 84.3k | metadata dict |
|---|---|---|---|---|
connector-service_snippet_-8421620542795988629_75_15 | clm | snippet | // connector-service/backend/connector-integration/src/utils/xml_utils.rs
// Return JSON as bytes
Ok(Bytes::from(json_string.into_bytes()))
}
/// Flattens a nested JSON structure, extracting values from "$text" fields
pub fn flatten_json_structure(json_value: Value) -> Value {
let mut flattened = Map::ne... | {
"chunk": null,
"crate": "connector-integration",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"sta... |
connector-service_snippet_-8421620542795988629_75_30 | clm | snippet | // connector-service/backend/connector-integration/src/utils/xml_utils.rs
// Return JSON as bytes
Ok(Bytes::from(json_string.into_bytes()))
}
/// Flattens a nested JSON structure, extracting values from "$text" fields
pub fn flatten_json_structure(json_value: Value) -> Value {
let mut flattened = Map::ne... | {
"chunk": null,
"crate": "connector-integration",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"sta... |
connector-service_snippet_-8421620542795988629_75_50 | clm | snippet | // connector-service/backend/connector-integration/src/utils/xml_utils.rs
// Return JSON as bytes
Ok(Bytes::from(json_string.into_bytes()))
}
/// Flattens a nested JSON structure, extracting values from "$text" fields
pub fn flatten_json_structure(json_value: Value) -> Value {
let mut flattened = Map::ne... | {
"chunk": null,
"crate": "connector-integration",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 44,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"sta... |
connector-service_snippet_-8421620542795988629_100_15 | clm | snippet | // connector-service/backend/connector-integration/src/utils/xml_utils.rs
if let Some(text_value) = value_obj.get("$text") {
// Extract the value from "$text" field
flattened.insert(key.clone(), text_value.clone());
} else if value_obj.is_empty() ... | {
"chunk": null,
"crate": "connector-integration",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"sta... |
connector-service_snippet_-8421620542795988629_100_30 | clm | snippet | // connector-service/backend/connector-integration/src/utils/xml_utils.rs
if let Some(text_value) = value_obj.get("$text") {
// Extract the value from "$text" field
flattened.insert(key.clone(), text_value.clone());
} else if value_obj.is_empty() ... | {
"chunk": null,
"crate": "connector-integration",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 19,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"sta... |
connector-service_snippet_-8421620542795988629_100_50 | clm | snippet | // connector-service/backend/connector-integration/src/utils/xml_utils.rs
if let Some(text_value) = value_obj.get("$text") {
// Extract the value from "$text" field
flattened.insert(key.clone(), text_value.clone());
} else if value_obj.is_empty() ... | {
"chunk": null,
"crate": "connector-integration",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 19,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"sta... |
connector-service_snippet_5501410502799828961_0_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/lib.rs
//! A Kafka tracing layer that integrates with the tracing ecosystem.
//!
//! This crate provides a simple way to send tracing logs to Kafka while maintaining
//! consistent JSON formatting through the log_utils infrastructure.
//!
//! # Examples
//! ```no_run
//! ... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_5501410502799828961_0_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/lib.rs
//! A Kafka tracing layer that integrates with the tracing ecosystem.
//!
//! This crate provides a simple way to send tracing logs to Kafka while maintaining
//! consistent JSON formatting through the log_utils infrastructure.
//!
//! # Examples
//! ```no_run
//! ... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_5501410502799828961_0_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/lib.rs
//! A Kafka tracing layer that integrates with the tracing ecosystem.
//!
//! This crate provides a simple way to send tracing logs to Kafka while maintaining
//! consistent JSON formatting through the log_utils infrastructure.
//!
//! # Examples
//! ```no_run
//! ... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_5501410502799828961_25_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/lib.rs
//!
//! ```no_run
//! use tracing_kafka::KafkaWriter;
//! use rdkafka::message::OwnedHeaders;
//!
//! let writer = KafkaWriter::new(
//! vec!["localhost:9092".to_string()],
//! "default-topic".to_string(),
//! None, None, None, None, None, None
//! ).ex... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_5501410502799828961_25_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/lib.rs
//!
//! ```no_run
//! use tracing_kafka::KafkaWriter;
//! use rdkafka::message::OwnedHeaders;
//!
//! let writer = KafkaWriter::new(
//! vec!["localhost:9092".to_string()],
//! "default-topic".to_string(),
//! None, None, None, None, None, None
//! ).ex... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_5501410502799828961_25_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/lib.rs
//!
//! ```no_run
//! use tracing_kafka::KafkaWriter;
//! use rdkafka::message::OwnedHeaders;
//!
//! let writer = KafkaWriter::new(
//! vec!["localhost:9092".to_string()],
//! "default-topic".to_string(),
//! None, None, None, None, None, None
//! ).ex... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 46,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_5501410502799828961_50_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/lib.rs
pub mod builder;
mod layer;
mod writer;
pub use layer::{KafkaLayer, KafkaLayerError};
pub use writer::{KafkaWriter, KafkaWriterError};
#[cfg(feature = "kafka-metrics")]
mod metrics;
/// Initializes the metrics for the tracing kafka.
/// This function should be c... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_5501410502799828961_50_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/lib.rs
pub mod builder;
mod layer;
mod writer;
pub use layer::{KafkaLayer, KafkaLayerError};
pub use writer::{KafkaWriter, KafkaWriterError};
#[cfg(feature = "kafka-metrics")]
mod metrics;
/// Initializes the metrics for the tracing kafka.
/// This function should be c... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 21,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_5501410502799828961_50_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/lib.rs
pub mod builder;
mod layer;
mod writer;
pub use layer::{KafkaLayer, KafkaLayerError};
pub use writer::{KafkaWriter, KafkaWriterError};
#[cfg(feature = "kafka-metrics")]
mod metrics;
/// Initializes the metrics for the tracing kafka.
/// This function should be c... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 21,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_0_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
//! Prometheus metrics for Kafka writer
use std::sync::LazyLock;
use prometheus::{register_int_counter, register_int_gauge, IntCounter, IntGauge};
/// Total number of logs successfully sent to Kafka
#[allow(clippy::expect_used)]
pub static KAFKA_LOGS_SENT: L... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_0_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
//! Prometheus metrics for Kafka writer
use std::sync::LazyLock;
use prometheus::{register_int_counter, register_int_gauge, IntCounter, IntGauge};
/// Total number of logs successfully sent to Kafka
#[allow(clippy::expect_used)]
pub static KAFKA_LOGS_SENT: L... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_0_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
//! Prometheus metrics for Kafka writer
use std::sync::LazyLock;
use prometheus::{register_int_counter, register_int_gauge, IntCounter, IntGauge};
/// Total number of logs successfully sent to Kafka
#[allow(clippy::expect_used)]
pub static KAFKA_LOGS_SENT: L... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_25_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
/// Current size of Kafka producer queue
#[allow(clippy::expect_used)]
pub static KAFKA_QUEUE_SIZE: LazyLock<IntGauge> = LazyLock::new(|| {
register_int_gauge!(
"kafka_producer_queue_size",
"Current size of Kafka producer queue"
)
.... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_25_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
/// Current size of Kafka producer queue
#[allow(clippy::expect_used)]
pub static KAFKA_QUEUE_SIZE: LazyLock<IntGauge> = LazyLock::new(|| {
register_int_gauge!(
"kafka_producer_queue_size",
"Current size of Kafka producer queue"
)
.... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_25_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
/// Current size of Kafka producer queue
#[allow(clippy::expect_used)]
pub static KAFKA_QUEUE_SIZE: LazyLock<IntGauge> = LazyLock::new(|| {
register_int_gauge!(
"kafka_producer_queue_size",
"Current size of Kafka producer queue"
)
.... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_50_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
"kafka_drops_msg_too_large_total",
"Total number of logs dropped due to message size exceeding limit"
)
.expect("Failed to register kafka_drops_msg_too_large_total metric")
});
/// Logs dropped due to timeout
#[allow(clippy::expect_used... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_50_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
"kafka_drops_msg_too_large_total",
"Total number of logs dropped due to message size exceeding limit"
)
.expect("Failed to register kafka_drops_msg_too_large_total metric")
});
/// Logs dropped due to timeout
#[allow(clippy::expect_used... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_50_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
"kafka_drops_msg_too_large_total",
"Total number of logs dropped due to message size exceeding limit"
)
.expect("Failed to register kafka_drops_msg_too_large_total metric")
});
/// Logs dropped due to timeout
#[allow(clippy::expect_used... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_75_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
/// Total number of audit events successfully sent to Kafka
#[allow(clippy::expect_used)]
pub static KAFKA_AUDIT_EVENTS_SENT: LazyLock<IntCounter> = LazyLock::new(|| {
register_int_counter!(
"kafka_audit_events_sent_total",
"Total number of... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_75_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
/// Total number of audit events successfully sent to Kafka
#[allow(clippy::expect_used)]
pub static KAFKA_AUDIT_EVENTS_SENT: LazyLock<IntCounter> = LazyLock::new(|| {
register_int_counter!(
"kafka_audit_events_sent_total",
"Total number of... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_75_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
/// Total number of audit events successfully sent to Kafka
#[allow(clippy::expect_used)]
pub static KAFKA_AUDIT_EVENTS_SENT: LazyLock<IntCounter> = LazyLock::new(|| {
register_int_counter!(
"kafka_audit_events_sent_total",
"Total number of... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_100_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
"kafka_audit_event_queue_size",
"Current size of Kafka audit event producer queue"
)
.expect("Failed to register kafka_audit_event_queue_size metric")
});
/// Audit events dropped due to queue full
#[allow(clippy::expect_used)]
pub stat... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_100_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
"kafka_audit_event_queue_size",
"Current size of Kafka audit event producer queue"
)
.expect("Failed to register kafka_audit_event_queue_size metric")
});
/// Audit events dropped due to queue full
#[allow(clippy::expect_used)]
pub stat... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_100_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
"kafka_audit_event_queue_size",
"Current size of Kafka audit event producer queue"
)
.expect("Failed to register kafka_audit_event_queue_size metric")
});
/// Audit events dropped due to queue full
#[allow(clippy::expect_used)]
pub stat... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_125_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
/// Audit events dropped due to timeout
#[allow(clippy::expect_used)]
pub static KAFKA_AUDIT_DROPS_TIMEOUT: LazyLock<IntCounter> = LazyLock::new(|| {
register_int_counter!(
"kafka_audit_drops_timeout_total",
"Total number of audit events dr... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_125_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
/// Audit events dropped due to timeout
#[allow(clippy::expect_used)]
pub static KAFKA_AUDIT_DROPS_TIMEOUT: LazyLock<IntCounter> = LazyLock::new(|| {
register_int_counter!(
"kafka_audit_drops_timeout_total",
"Total number of audit events dr... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_125_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
/// Audit events dropped due to timeout
#[allow(clippy::expect_used)]
pub static KAFKA_AUDIT_DROPS_TIMEOUT: LazyLock<IntCounter> = LazyLock::new(|| {
register_int_counter!(
"kafka_audit_drops_timeout_total",
"Total number of audit events dr... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 44,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_150_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
/// metric name), the application will panic immediately.
#[cfg(feature = "kafka-metrics")]
pub fn initialize_all_metrics() {
// Force evaluation of all lazy metrics to fail fast if registration fails.
let _ = &*KAFKA_LOGS_SENT;
let _ = &*KAFKA_LOGS... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_150_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
/// metric name), the application will panic immediately.
#[cfg(feature = "kafka-metrics")]
pub fn initialize_all_metrics() {
// Force evaluation of all lazy metrics to fail fast if registration fails.
let _ = &*KAFKA_LOGS_SENT;
let _ = &*KAFKA_LOGS... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 19,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_863436229918585844_150_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/metrics.rs
/// metric name), the application will panic immediately.
#[cfg(feature = "kafka-metrics")]
pub fn initialize_all_metrics() {
// Force evaluation of all lazy metrics to fail fast if registration fails.
let _ = &*KAFKA_LOGS_SENT;
let _ = &*KAFKA_LOGS... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 19,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_0_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
//! Kafka writer implementation for sending formatted log messages to Kafka.
use std::{
io::{self, Write},
sync::Arc,
time::Duration,
};
use rdkafka::{
config::ClientConfig,
error::{KafkaError, RDKafkaErrorCode},
message::OwnedHeaders,
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_0_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
//! Kafka writer implementation for sending formatted log messages to Kafka.
use std::{
io::{self, Write},
sync::Arc,
time::Duration,
};
use rdkafka::{
config::ClientConfig,
error::{KafkaError, RDKafkaErrorCode},
message::OwnedHeaders,
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_0_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
//! Kafka writer implementation for sending formatted log messages to Kafka.
use std::{
io::{self, Write},
sync::Arc,
time::Duration,
};
use rdkafka::{
config::ClientConfig,
error::{KafkaError, RDKafkaErrorCode},
message::OwnedHeaders,
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_25_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
/// A `ProducerContext` that handles delivery callbacks to increment metrics.
#[derive(Clone)]
struct MetricsProducerContext;
impl ClientContext for MetricsProducerContext {}
impl ProducerContext for MetricsProducerContext {
type DeliveryOpaque = Box<Kafka... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_25_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
/// A `ProducerContext` that handles delivery callbacks to increment metrics.
#[derive(Clone)]
struct MetricsProducerContext;
impl ClientContext for MetricsProducerContext {}
impl ProducerContext for MetricsProducerContext {
type DeliveryOpaque = Box<Kafka... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_25_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
/// A `ProducerContext` that handles delivery callbacks to increment metrics.
#[derive(Clone)]
struct MetricsProducerContext;
impl ClientContext for MetricsProducerContext {}
impl ProducerContext for MetricsProducerContext {
type DeliveryOpaque = Box<Kafka... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_50_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
match (message_type, &kafka_error) {
(
KafkaMessageType::Event,
KafkaError::MessageProduction(RDKafkaErrorCode::QueueFull),
) => {
KAFKA_AUDIT_DROPS_QUEUE_FUL... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_50_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
match (message_type, &kafka_error) {
(
KafkaMessageType::Event,
KafkaError::MessageProduction(RDKafkaErrorCode::QueueFull),
) => {
KAFKA_AUDIT_DROPS_QUEUE_FUL... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_50_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
match (message_type, &kafka_error) {
(
KafkaMessageType::Event,
KafkaError::MessageProduction(RDKafkaErrorCode::QueueFull),
) => {
KAFKA_AUDIT_DROPS_QUEUE_FUL... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_75_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
) => {
KAFKA_DROPS_QUEUE_FULL.inc();
}
(
KafkaMessageType::Log,
KafkaError::MessageProduction(RDKafkaErrorCode::MessageSizeTooLarge),
) =>... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_75_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
) => {
KAFKA_DROPS_QUEUE_FULL.inc();
}
(
KafkaMessageType::Log,
KafkaError::MessageProduction(RDKafkaErrorCode::MessageSizeTooLarge),
) =>... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_75_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
) => {
KAFKA_DROPS_QUEUE_FULL.inc();
}
(
KafkaMessageType::Log,
KafkaError::MessageProduction(RDKafkaErrorCode::MessageSizeTooLarge),
) =>... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_100_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
enum KafkaMessageType {
Event,
Log,
}
/// Kafka writer that implements std::io::Write for seamless integration with tracing
#[derive(Clone)]
pub struct KafkaWriter {
producer: Arc<ThreadedProducer<MetricsProducerContext>>,
topic: String,
}
impl... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_100_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
enum KafkaMessageType {
Event,
Log,
}
/// Kafka writer that implements std::io::Write for seamless integration with tracing
#[derive(Clone)]
pub struct KafkaWriter {
producer: Arc<ThreadedProducer<MetricsProducerContext>>,
topic: String,
}
impl... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_100_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
enum KafkaMessageType {
Event,
Log,
}
/// Kafka writer that implements std::io::Write for seamless integration with tracing
#[derive(Clone)]
pub struct KafkaWriter {
producer: Arc<ThreadedProducer<MetricsProducerContext>>,
topic: String,
}
impl... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_125_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
topic: String,
batch_size: Option<usize>,
linger_ms: Option<u64>,
queue_buffering_max_messages: Option<usize>,
queue_buffering_max_kbytes: Option<usize>,
reconnect_backoff_min_ms: Option<u64>,
reconnect_bac... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_125_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
topic: String,
batch_size: Option<usize>,
linger_ms: Option<u64>,
queue_buffering_max_messages: Option<usize>,
queue_buffering_max_kbytes: Option<usize>,
reconnect_backoff_min_ms: Option<u64>,
reconnect_bac... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_125_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
topic: String,
batch_size: Option<usize>,
linger_ms: Option<u64>,
queue_buffering_max_messages: Option<usize>,
queue_buffering_max_kbytes: Option<usize>,
reconnect_backoff_min_ms: Option<u64>,
reconnect_bac... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_150_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
}
if let Some(ms) = linger_ms {
config.set("linger.ms", ms.to_string());
}
let producer: ThreadedProducer<MetricsProducerContext> = config
.create_with_context(MetricsProducerContext)
.map_err(... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_150_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
}
if let Some(ms) = linger_ms {
config.set("linger.ms", ms.to_string());
}
let producer: ThreadedProducer<MetricsProducerContext> = config
.create_with_context(MetricsProducerContext)
.map_err(... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_150_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
}
if let Some(ms) = linger_ms {
config.set("linger.ms", ms.to_string());
}
let producer: ThreadedProducer<MetricsProducerContext> = config
.create_with_context(MetricsProducerContext)
.map_err(... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_175_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
key: Option<&str>,
payload: &[u8],
headers: Option<OwnedHeaders>,
) -> Result<(), KafkaError> {
#[cfg(feature = "kafka-metrics")]
{
let queue_size = self.producer.in_flight_count();
KAFKA_AUDIT_... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_175_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
key: Option<&str>,
payload: &[u8],
headers: Option<OwnedHeaders>,
) -> Result<(), KafkaError> {
#[cfg(feature = "kafka-metrics")]
{
let queue_size = self.producer.in_flight_count();
KAFKA_AUDIT_... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_175_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
key: Option<&str>,
payload: &[u8],
headers: Option<OwnedHeaders>,
) -> Result<(), KafkaError> {
#[cfg(feature = "kafka-metrics")]
{
let queue_size = self.producer.in_flight_count();
KAFKA_AUDIT_... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_200_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
}
match self.producer.send(record) {
Ok(_) => Ok(()),
Err((kafka_error, _)) => {
#[cfg(feature = "kafka-metrics")]
{
KAFKA_AUDIT_EVENTS_DROPPED.inc();
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_200_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
}
match self.producer.send(record) {
Ok(_) => Ok(()),
Err((kafka_error, _)) => {
#[cfg(feature = "kafka-metrics")]
{
KAFKA_AUDIT_EVENTS_DROPPED.inc();
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_200_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
}
match self.producer.send(record) {
Ok(_) => Ok(()),
Err((kafka_error, _)) => {
#[cfg(feature = "kafka-metrics")]
{
KAFKA_AUDIT_EVENTS_DROPPED.inc();
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_225_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
pub fn builder() -> crate::builder::KafkaWriterBuilder {
crate::builder::KafkaWriterBuilder::new()
}
}
impl Write for KafkaWriter {
fn write(&mut self, buf: &[u8]) -> io::Result<usize> {
#[cfg(feature = "kafka-metrics")]
{
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_225_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
pub fn builder() -> crate::builder::KafkaWriterBuilder {
crate::builder::KafkaWriterBuilder::new()
}
}
impl Write for KafkaWriter {
fn write(&mut self, buf: &[u8]) -> io::Result<usize> {
#[cfg(feature = "kafka-metrics")]
{
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_225_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
pub fn builder() -> crate::builder::KafkaWriterBuilder {
crate::builder::KafkaWriterBuilder::new()
}
}
impl Write for KafkaWriter {
fn write(&mut self, buf: &[u8]) -> io::Result<usize> {
#[cfg(feature = "kafka-metrics")]
{
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_250_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
KAFKA_LOGS_DROPPED.inc();
match &kafka_error {
KafkaError::MessageProduction(RDKafkaErrorCode::QueueFull) => {
KAFKA_DROPS_QUEUE_FULL.inc();
}
_ ... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_250_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
KAFKA_LOGS_DROPPED.inc();
match &kafka_error {
KafkaError::MessageProduction(RDKafkaErrorCode::QueueFull) => {
KAFKA_DROPS_QUEUE_FULL.inc();
}
_ ... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_250_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
KAFKA_LOGS_DROPPED.inc();
match &kafka_error {
KafkaError::MessageProduction(RDKafkaErrorCode::QueueFull) => {
KAFKA_DROPS_QUEUE_FULL.inc();
}
_ ... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_275_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
/// Errors that can occur when creating or using a KafkaWriter.
#[derive(Debug, thiserror::Error)]
pub enum KafkaWriterError {
#[error("Failed to create Kafka producer: {0}")]
ProducerCreation(KafkaError),
#[error("Failed to fetch Kafka metadata: {0}... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_275_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
/// Errors that can occur when creating or using a KafkaWriter.
#[derive(Debug, thiserror::Error)]
pub enum KafkaWriterError {
#[error("Failed to create Kafka producer: {0}")]
ProducerCreation(KafkaError),
#[error("Failed to fetch Kafka metadata: {0}... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-8840925468162898005_275_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/writer.rs
/// Errors that can occur when creating or using a KafkaWriter.
#[derive(Debug, thiserror::Error)]
pub enum KafkaWriterError {
#[error("Failed to create Kafka producer: {0}")]
ProducerCreation(KafkaError),
#[error("Failed to fetch Kafka metadata: {0}... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_0_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
//! Kafka layer implementation that reuses log_utils formatting.
use std::{
collections::{HashMap, HashSet},
time::Duration,
};
use log_utils::{
AdditionalFieldsPlacement, JsonFormattingLayer, JsonFormattingLayerConfig, LoggerError,
};
use tracing::... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_0_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
//! Kafka layer implementation that reuses log_utils formatting.
use std::{
collections::{HashMap, HashSet},
time::Duration,
};
use log_utils::{
AdditionalFieldsPlacement, JsonFormattingLayer, JsonFormattingLayerConfig, LoggerError,
};
use tracing::... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_0_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
//! Kafka layer implementation that reuses log_utils formatting.
use std::{
collections::{HashMap, HashSet},
time::Duration,
};
use log_utils::{
AdditionalFieldsPlacement, JsonFormattingLayer, JsonFormattingLayerConfig, LoggerError,
};
use tracing::... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_25_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
impl KafkaLayer {
/// Creates a new builder for configuring a KafkaLayer.
pub fn builder() -> KafkaLayerBuilder {
KafkaLayerBuilder::new()
}
/// Creates a new KafkaLayer from a pre-configured KafkaWriter.
/// This is primarily used in... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_25_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
impl KafkaLayer {
/// Creates a new builder for configuring a KafkaLayer.
pub fn builder() -> KafkaLayerBuilder {
KafkaLayerBuilder::new()
}
/// Creates a new KafkaLayer from a pre-configured KafkaWriter.
/// This is primarily used in... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_25_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
impl KafkaLayer {
/// Creates a new builder for configuring a KafkaLayer.
pub fn builder() -> KafkaLayerBuilder {
KafkaLayerBuilder::new()
}
/// Creates a new KafkaLayer from a pre-configured KafkaWriter.
/// This is primarily used in... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_50_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
impl<S> Layer<S> for KafkaLayer
where
S: Subscriber + for<'lookup> tracing_subscriber::registry::LookupSpan<'lookup>,
{
fn on_event(&self, event: &tracing::Event<'_>, ctx: tracing_subscriber::layer::Context<'_, S>) {
self.inner.on_event(event, ct... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_50_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
impl<S> Layer<S> for KafkaLayer
where
S: Subscriber + for<'lookup> tracing_subscriber::registry::LookupSpan<'lookup>,
{
fn on_event(&self, event: &tracing::Event<'_>, ctx: tracing_subscriber::layer::Context<'_, S>) {
self.inner.on_event(event, ct... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_50_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
impl<S> Layer<S> for KafkaLayer
where
S: Subscriber + for<'lookup> tracing_subscriber::registry::LookupSpan<'lookup>,
{
fn on_event(&self, event: &tracing::Event<'_>, ctx: tracing_subscriber::layer::Context<'_, S>) {
self.inner.on_event(event, ct... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_75_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
fn on_close(&self, id: tracing::span::Id, ctx: tracing_subscriber::layer::Context<'_, S>) {
self.inner.on_close(id, ctx);
}
}
impl KafkaLayer {
/// Boxes the layer, making it easier to compose with other layers.
pub fn boxed<S>(self) -> ... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_75_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
fn on_close(&self, id: tracing::span::Id, ctx: tracing_subscriber::layer::Context<'_, S>) {
self.inner.on_close(id, ctx);
}
}
impl KafkaLayer {
/// Boxes the layer, making it easier to compose with other layers.
pub fn boxed<S>(self) -> ... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_75_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
fn on_close(&self, id: tracing::span::Id, ctx: tracing_subscriber::layer::Context<'_, S>) {
self.inner.on_close(id, ctx);
}
}
impl KafkaLayer {
/// Boxes the layer, making it easier to compose with other layers.
pub fn boxed<S>(self) -> ... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_100_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
#[error("Missing brokers configuration")]
MissingBrokers,
#[error("Missing topic configuration")]
MissingTopic,
}
/// Builder for creating a KafkaLayer with custom configuration.
#[derive(Debug, Clone, Default)]
pub struct KafkaLayerBuilder {
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_100_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
#[error("Missing brokers configuration")]
MissingBrokers,
#[error("Missing topic configuration")]
MissingTopic,
}
/// Builder for creating a KafkaLayer with custom configuration.
#[derive(Debug, Clone, Default)]
pub struct KafkaLayerBuilder {
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_100_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
#[error("Missing brokers configuration")]
MissingBrokers,
#[error("Missing topic configuration")]
MissingTopic,
}
/// Builder for creating a KafkaLayer with custom configuration.
#[derive(Debug, Clone, Default)]
pub struct KafkaLayerBuilder {
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_125_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
.brokers(brokers.iter().map(|s| s.to_string()).collect());
self
}
/// Sets the Kafka topic to send logs to.
pub fn topic(mut self, topic: impl Into<String>) -> Self {
self.writer_builder = self.writer_builder.topic(topic);... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_125_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
.brokers(brokers.iter().map(|s| s.to_string()).collect());
self
}
/// Sets the Kafka topic to send logs to.
pub fn topic(mut self, topic: impl Into<String>) -> Self {
self.writer_builder = self.writer_builder.topic(topic);... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_125_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
.brokers(brokers.iter().map(|s| s.to_string()).collect());
self
}
/// Sets the Kafka topic to send logs to.
pub fn topic(mut self, topic: impl Into<String>) -> Self {
self.writer_builder = self.writer_builder.topic(topic);... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_150_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
self
}
/// Sets the maximum number of messages to buffer in the producer's queue.
pub fn queue_buffering_max_messages(mut self, size: usize) -> Self {
self.writer_builder = self.writer_builder.queue_buffering_max_messages(size);
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_150_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
self
}
/// Sets the maximum number of messages to buffer in the producer's queue.
pub fn queue_buffering_max_messages(mut self, size: usize) -> Self {
self.writer_builder = self.writer_builder.queue_buffering_max_messages(size);
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_150_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
self
}
/// Sets the maximum number of messages to buffer in the producer's queue.
pub fn queue_buffering_max_messages(mut self, size: usize) -> Self {
self.writer_builder = self.writer_builder.queue_buffering_max_messages(size);
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 40,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_175_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
self
}
/// Adds a single static field that will be included in every log entry.
pub fn add_static_field(mut self, key: String, value: serde_json::Value) -> Self {
self.static_fields.insert(key, value);
self
}
/// Buil... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_175_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
self
}
/// Adds a single static field that will be included in every log entry.
pub fn add_static_field(mut self, key: String, value: serde_json::Value) -> Self {
self.static_fields.insert(key, value);
self
}
/// Buil... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-6998519293309712771_175_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/layer.rs
self
}
/// Adds a single static field that will be included in every log entry.
pub fn add_static_field(mut self, key: String, value: serde_json::Value) -> Self {
self.static_fields.insert(key, value);
self
}
/// Buil... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-7694171944301536872_0_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/builder.rs
//! Builder pattern implementation for KafkaWriter
use std::time::Duration;
use super::writer::{KafkaWriter, KafkaWriterError};
/// Builder for creating a KafkaWriter with custom configuration
#[derive(Debug, Clone, Default)]
pub struct KafkaWriterBuilder {
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-7694171944301536872_0_30 | clm | snippet | // connector-service/backend/tracing-kafka/src/builder.rs
//! Builder pattern implementation for KafkaWriter
use std::time::Duration;
use super::writer::{KafkaWriter, KafkaWriterError};
/// Builder for creating a KafkaWriter with custom configuration
#[derive(Debug, Clone, Default)]
pub struct KafkaWriterBuilder {
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 30,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-7694171944301536872_0_50 | clm | snippet | // connector-service/backend/tracing-kafka/src/builder.rs
//! Builder pattern implementation for KafkaWriter
use std::time::Duration;
use super::writer::{KafkaWriter, KafkaWriterError};
/// Builder for creating a KafkaWriter with custom configuration
#[derive(Debug, Clone, Default)]
pub struct KafkaWriterBuilder {
... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 50,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
connector-service_snippet_-7694171944301536872_25_15 | clm | snippet | // connector-service/backend/tracing-kafka/src/builder.rs
/// Sets the Kafka brokers to connect to
pub fn brokers(mut self, brokers: Vec<String>) -> Self {
self.brokers = Some(brokers);
self
}
/// Sets the Kafka topic to send logs to
pub fn topic(mut self, topic: impl Into<String>)... | {
"chunk": null,
"crate": "tracing-kafka",
"enum_name": null,
"file_size": null,
"for_type": null,
"function_name": null,
"is_async": null,
"is_pub": null,
"lines": 15,
"method_name": null,
"num_enums": null,
"num_items": null,
"num_structs": null,
"repo": "connector-service",
"start_line"... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.