|
|
use std::{debug_assert, io::Write, sync::Arc, thread::JoinHandle, time::Duration}; |
|
|
|
|
|
use crossbeam_channel::{Receiver, RecvTimeoutError, Sender, TryRecvError, bounded, unbounded}; |
|
|
use crossbeam_utils::CachePadded; |
|
|
use parking_lot::{Mutex, MutexGuard}; |
|
|
use thread_local::ThreadLocal; |
|
|
|
|
|
type ThreadLocalState = CachePadded<Mutex<Option<TraceInfoBuffer>>>; |
|
|
|
|
|
|
|
|
|
|
|
const THREAD_LOCAL_INITIAL_BUFFER_SIZE: usize = 1024 * 1024; |
|
|
|
|
|
const WRITE_BUFFER_SIZE: usize = 100 * 1024 * 1024; |
|
|
|
|
|
struct TraceInfoBuffer { |
|
|
buffer: Vec<u8>, |
|
|
} |
|
|
|
|
|
impl TraceInfoBuffer { |
|
|
fn new(capacity: usize) -> Self { |
|
|
Self { |
|
|
buffer: Vec::with_capacity(capacity), |
|
|
} |
|
|
} |
|
|
|
|
|
fn push(&mut self, data: u8) { |
|
|
self.buffer.push(data); |
|
|
} |
|
|
|
|
|
fn extend(&mut self, data: &[u8]) { |
|
|
self.buffer.extend_from_slice(data); |
|
|
} |
|
|
|
|
|
fn clear(&mut self) { |
|
|
self.buffer.clear(); |
|
|
} |
|
|
} |
|
|
|
|
|
#[derive(Clone)] |
|
|
pub struct TraceWriter { |
|
|
data_tx: Sender<Option<TraceInfoBuffer>>, |
|
|
return_rx: Receiver<TraceInfoBuffer>, |
|
|
thread_locals: Arc<ThreadLocal<ThreadLocalState>>, |
|
|
} |
|
|
|
|
|
impl TraceWriter { |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
pub fn new<W: Write + Send + 'static>(mut writer: W) -> (Self, TraceWriterGuard) { |
|
|
let (data_tx, data_rx) = unbounded::<Option<TraceInfoBuffer>>(); |
|
|
let (return_tx, return_rx) = bounded::<TraceInfoBuffer>(1024); |
|
|
let thread_locals: Arc<ThreadLocal<ThreadLocalState>> = Default::default(); |
|
|
|
|
|
let trace_writer = Self { |
|
|
data_tx: data_tx.clone(), |
|
|
return_rx: return_rx.clone(), |
|
|
thread_locals: thread_locals.clone(), |
|
|
}; |
|
|
|
|
|
fn steal_from_thread_locals( |
|
|
thread_locals: &Arc<ThreadLocal<ThreadLocalState>>, |
|
|
stolen_buffers: &mut Vec<TraceInfoBuffer>, |
|
|
) { |
|
|
for state in thread_locals.iter() { |
|
|
let mut buffer = state.lock(); |
|
|
if let Some(buffer) = buffer.take() { |
|
|
stolen_buffers.push(buffer); |
|
|
} |
|
|
} |
|
|
} |
|
|
|
|
|
let handle: std::thread::JoinHandle<()> = std::thread::spawn(move || { |
|
|
let _ = writer.write(b"TRACEv0"); |
|
|
let mut buf = Vec::with_capacity(WRITE_BUFFER_SIZE); |
|
|
let mut stolen_buffers = Vec::new(); |
|
|
let mut should_exit = false; |
|
|
'outer: loop { |
|
|
if !buf.is_empty() { |
|
|
let _ = writer.write_all(&buf); |
|
|
let _ = writer.flush(); |
|
|
buf.clear(); |
|
|
} |
|
|
|
|
|
let recv = if should_exit { |
|
|
Ok(None) |
|
|
} else { |
|
|
data_rx.recv_timeout(Duration::from_secs(1)) |
|
|
}; |
|
|
|
|
|
let mut data = match recv { |
|
|
Ok(Some(data)) => data, |
|
|
result => { |
|
|
if result.is_ok() { |
|
|
|
|
|
should_exit = true; |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
steal_from_thread_locals(&thread_locals, &mut stolen_buffers); |
|
|
if let Some(data) = stolen_buffers.pop() { |
|
|
data |
|
|
} else { |
|
|
match result { |
|
|
Ok(Some(_)) => unreachable!(), |
|
|
Ok(None) | Err(RecvTimeoutError::Disconnected) => { |
|
|
|
|
|
break 'outer; |
|
|
} |
|
|
Err(RecvTimeoutError::Timeout) => { |
|
|
|
|
|
continue; |
|
|
} |
|
|
} |
|
|
} |
|
|
} |
|
|
}; |
|
|
if data.buffer.len() > buf.capacity() { |
|
|
let _ = writer.write_all(&data.buffer); |
|
|
} else { |
|
|
buf.extend_from_slice(&data.buffer); |
|
|
} |
|
|
data.clear(); |
|
|
let _ = return_tx.try_send(data); |
|
|
loop { |
|
|
let recv = stolen_buffers.pop().map(Some).ok_or(()).or_else(|_| { |
|
|
if should_exit { |
|
|
Ok(None) |
|
|
} else { |
|
|
data_rx.try_recv() |
|
|
} |
|
|
}); |
|
|
match recv { |
|
|
Ok(Some(mut data)) => { |
|
|
let data_buffer = &data.buffer; |
|
|
if data_buffer.is_empty() { |
|
|
break 'outer; |
|
|
} |
|
|
if buf.len() + data_buffer.len() > buf.capacity() { |
|
|
let _ = writer.write_all(&buf); |
|
|
buf.clear(); |
|
|
if data_buffer.len() > buf.capacity() { |
|
|
let _ = writer.write_all(data_buffer); |
|
|
} else { |
|
|
buf.extend_from_slice(data_buffer); |
|
|
} |
|
|
} else { |
|
|
buf.extend_from_slice(data_buffer); |
|
|
} |
|
|
data.clear(); |
|
|
let _ = return_tx.try_send(data); |
|
|
} |
|
|
Ok(None) | Err(TryRecvError::Disconnected) => { |
|
|
should_exit = true; |
|
|
break; |
|
|
} |
|
|
Err(TryRecvError::Empty) => { |
|
|
break; |
|
|
} |
|
|
} |
|
|
} |
|
|
} |
|
|
drop(writer); |
|
|
}); |
|
|
|
|
|
let guard = TraceWriterGuard { |
|
|
data_tx: Some(data_tx), |
|
|
return_rx: Some(return_rx), |
|
|
handle: Some(handle), |
|
|
}; |
|
|
(trace_writer, guard) |
|
|
} |
|
|
|
|
|
fn send(&self, data: TraceInfoBuffer) { |
|
|
debug_assert!(!data.buffer.is_empty()); |
|
|
let _ = self.data_tx.send(Some(data)); |
|
|
} |
|
|
|
|
|
fn get_empty_buffer(&self, capacity: usize) -> TraceInfoBuffer { |
|
|
self.return_rx |
|
|
.try_recv() |
|
|
.ok() |
|
|
.unwrap_or_else(|| TraceInfoBuffer::new(capacity)) |
|
|
} |
|
|
|
|
|
pub fn start_write(&self) -> WriteGuard<'_> { |
|
|
let thread_local_buffer = self.thread_locals.get_or_default(); |
|
|
let buffer = thread_local_buffer.lock(); |
|
|
WriteGuard::new(buffer, self) |
|
|
} |
|
|
} |
|
|
|
|
|
pub struct TraceWriterGuard { |
|
|
data_tx: Option<Sender<Option<TraceInfoBuffer>>>, |
|
|
return_rx: Option<Receiver<TraceInfoBuffer>>, |
|
|
handle: Option<JoinHandle<()>>, |
|
|
} |
|
|
|
|
|
impl Drop for TraceWriterGuard { |
|
|
fn drop(&mut self) { |
|
|
|
|
|
let _ = self.data_tx.take().unwrap().send(None); |
|
|
|
|
|
let return_rx = self.return_rx.take().unwrap(); |
|
|
while return_rx.recv().is_ok() {} |
|
|
|
|
|
let _ = self.handle.take().unwrap().join(); |
|
|
} |
|
|
} |
|
|
|
|
|
pub struct WriteGuard<'l> { |
|
|
|
|
|
buffer: MutexGuard<'l, Option<TraceInfoBuffer>>, |
|
|
trace_writer: &'l TraceWriter, |
|
|
} |
|
|
|
|
|
impl<'l> WriteGuard<'l> { |
|
|
fn new( |
|
|
mut buffer: MutexGuard<'l, Option<TraceInfoBuffer>>, |
|
|
trace_writer: &'l TraceWriter, |
|
|
) -> Self { |
|
|
|
|
|
if buffer.is_none() { |
|
|
*buffer = Some(trace_writer.get_empty_buffer(THREAD_LOCAL_INITIAL_BUFFER_SIZE)); |
|
|
}; |
|
|
Self { |
|
|
buffer, |
|
|
trace_writer, |
|
|
} |
|
|
} |
|
|
|
|
|
fn buffer(&mut self) -> &mut TraceInfoBuffer { |
|
|
|
|
|
unsafe { self.buffer.as_mut().unwrap_unchecked() } |
|
|
} |
|
|
|
|
|
pub fn push(&mut self, data: u8) { |
|
|
self.buffer().push(data); |
|
|
} |
|
|
|
|
|
pub fn extend(&mut self, data: &[u8]) { |
|
|
self.buffer().extend(data); |
|
|
} |
|
|
} |
|
|
|
|
|
impl Drop for WriteGuard<'_> { |
|
|
fn drop(&mut self) { |
|
|
if self.buffer().buffer.capacity() * 2 < self.buffer().buffer.len() * 3 { |
|
|
let capacity = self.buffer().buffer.capacity(); |
|
|
let new_buffer = self.trace_writer.get_empty_buffer(capacity); |
|
|
let buffer = std::mem::replace(self.buffer(), new_buffer); |
|
|
self.trace_writer.send(buffer); |
|
|
} |
|
|
} |
|
|
} |
|
|
|