repo
stringlengths
6
65
file_url
stringlengths
81
311
file_path
stringlengths
6
227
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 15:31:58
2026-01-04 20:25:31
truncated
bool
2 classes
Gelbpunkt/tokio-websockets
https://github.com/Gelbpunkt/tokio-websockets/blob/618599d15e9ee2a9a7191d081a56d90eae4a811a/src/proto/mod.rs
src/proto/mod.rs
//! This module contains a correct and complete implementation of [RFC6455](https://datatracker.ietf.org/doc/html/rfc6455). //! //! Any extensions are currently not implemented. #[cfg(any(feature = "client", feature = "server"))] pub(crate) use self::types::Role; pub use self::{ error::ProtocolError, stream::WebSocketStream, types::{CloseCode, Config, Limits, Message, Payload}, }; mod codec; mod error; mod stream; mod types;
rust
MIT
618599d15e9ee2a9a7191d081a56d90eae4a811a
2026-01-04T20:19:32.035255Z
false
Gelbpunkt/tokio-websockets
https://github.com/Gelbpunkt/tokio-websockets/blob/618599d15e9ee2a9a7191d081a56d90eae4a811a/tests/pr102_regression.rs
tests/pr102_regression.rs
#![cfg(all(feature = "client", feature = "server"))] use std::time::Duration; use futures_util::{SinkExt, StreamExt}; use tokio::{io::duplex, time::timeout}; use tokio_websockets::{ClientBuilder, Message, ServerBuilder}; #[tokio::test] async fn test_pr102_regression() { let (tx, rx) = duplex(64); // Create a server that sends a close message and immediately closes the // connection let mut server = ServerBuilder::new().serve(rx); // Don't use ws.close() here to avoid a graceful handshake server.send(Message::close(None, "")).await.unwrap(); // This will close the connection drop(server); // Create a client and make sure the close frame from the server was sent and // the connection closed let mut client = ClientBuilder::new().take_over(tx); // Receive the close frame, this will queue the close ack frame assert!(client.next().await.unwrap().unwrap().is_close()); // The close ack frame is queued, the connection is closed, before PR 102 this // would trigger an infinite loop in close() // For the regression test, fail the test if this doesn't close within a second assert!( timeout(Duration::from_secs(1), client.close()) .await .is_ok() ); }
rust
MIT
618599d15e9ee2a9a7191d081a56d90eae4a811a
2026-01-04T20:19:32.035255Z
false
Gelbpunkt/tokio-websockets
https://github.com/Gelbpunkt/tokio-websockets/blob/618599d15e9ee2a9a7191d081a56d90eae4a811a/tests/issue106.rs
tests/issue106.rs
#![cfg(feature = "client")] use std::{ io::Result, pin::Pin, sync::{ Arc, atomic::{AtomicUsize, Ordering}, }, task::{Context, Poll, Wake}, }; use futures_util::{SinkExt, StreamExt}; use tokio::io::{AsyncRead, AsyncWrite, ReadBuf}; use tokio_websockets::{ClientBuilder, Message}; struct CountWaker(AtomicUsize); impl Wake for CountWaker { fn wake(self: Arc<Self>) { self.0.fetch_add(1, Ordering::Relaxed); } } struct MockStream; impl AsyncRead for MockStream { fn poll_read( self: Pin<&mut Self>, _: &mut Context<'_>, _: &mut ReadBuf<'_>, ) -> Poll<Result<()>> { Poll::Ready(Ok(())) } } impl AsyncWrite for MockStream { // We can track the amount of flushes called on the WebSocketStream by counting // the writes to the mock stream, since flushing the WebSocketStream writes // pending data to the mock stream. fn poll_write(self: Pin<&mut Self>, cx: &mut Context<'_>, _: &[u8]) -> Poll<Result<usize>> { cx.waker().wake_by_ref(); Poll::Pending } fn poll_flush(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<Result<()>> { Poll::Pending } fn poll_shutdown(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<Result<()>> { Poll::Pending } } /// Test that calling poll_next always attempts to flush the stream. #[test] fn main() { let mut ws = ClientBuilder::new().take_over(MockStream); // queue a message to trigger a flush on next read let _ = ws.start_send_unpin(Message::text("message")); let count1 = Arc::new(CountWaker(AtomicUsize::new(0))); let count2 = Arc::new(CountWaker(AtomicUsize::new(0))); let waker1 = count1.clone().into(); let waker2 = count2.clone().into(); // queued message triggers a flush let _ = ws.poll_flush_unpin(&mut Context::from_waker(&waker1)); // there's still messages queued, last flush's waker is reused let _ = ws.poll_next_unpin(&mut Context::from_waker(&waker2)); assert_eq!(count1.0.load(Ordering::Relaxed), 2); assert_eq!(count2.0.load(Ordering::Relaxed), 0); }
rust
MIT
618599d15e9ee2a9a7191d081a56d90eae4a811a
2026-01-04T20:19:32.035255Z
false
Gelbpunkt/tokio-websockets
https://github.com/Gelbpunkt/tokio-websockets/blob/618599d15e9ee2a9a7191d081a56d90eae4a811a/tests/pr117_regression.rs
tests/pr117_regression.rs
#![cfg(all(feature = "client", feature = "server"))] use futures_util::{SinkExt, StreamExt, stream}; use tokio::io::{AsyncRead, AsyncWrite, duplex}; use tokio_websockets::{ClientBuilder, Message, ServerBuilder}; const NUM_MSG: usize = 1024; const MESSAGE: &str = "test message"; #[tokio::test] async fn test_pr117_regression() { let (tx, rx) = duplex(64); tokio::join!(server(rx), client(tx)); } async fn server<S>(stream: S) where S: AsyncRead + AsyncWrite + Unpin, { let server = ServerBuilder::new().serve(stream); let messages = server .inspect(|message| { let message = message.as_ref().unwrap(); assert!(matches!(message.as_text(), Some(MESSAGE))); }) .collect::<Vec<_>>() .await; assert_eq!(messages.len(), NUM_MSG); } async fn client<S>(stream: S) where S: AsyncRead + AsyncWrite + Unpin, { let mut client = ClientBuilder::new().take_over(stream); let mut messages = stream::iter((0..NUM_MSG).map(|_| Ok(Message::text(MESSAGE)))); client.send_all(&mut messages).await.unwrap(); }
rust
MIT
618599d15e9ee2a9a7191d081a56d90eae4a811a
2026-01-04T20:19:32.035255Z
false
Gelbpunkt/tokio-websockets
https://github.com/Gelbpunkt/tokio-websockets/blob/618599d15e9ee2a9a7191d081a56d90eae4a811a/tests/cancellation_safety.rs
tests/cancellation_safety.rs
#![cfg(feature = "server")] use std::{ io, pin::Pin, task::{Context, Poll}, }; use futures_util::{FutureExt, StreamExt, task::noop_waker_ref}; use tokio::io::{AsyncRead, AsyncWrite, ReadBuf}; use tokio_websockets::ServerBuilder; struct SuperSlow { buf: Vec<u8>, delays: u8, } impl SuperSlow { pub fn new() -> Self { let buf = b"\x89\x84\xe47D\xa4\xe55G\xa0".to_vec(); Self { buf, delays: 0 } } } impl AsyncRead for SuperSlow { fn poll_read( mut self: Pin<&mut Self>, cx: &mut Context<'_>, buf: &mut ReadBuf<'_>, ) -> Poll<io::Result<()>> { if self.delays == 0 { self.delays += 1; let a: Vec<u8> = self.buf.splice(..4, []).collect::<Vec<_>>(); Pin::new(&mut a.as_slice()).poll_read(cx, buf) } else if self.delays == 1 { self.delays += 1; cx.waker().wake_by_ref(); Poll::Pending } else { let a: Vec<u8> = self.buf.splice(.., []).collect::<Vec<_>>(); Pin::new(&mut a.as_slice()).poll_read(cx, buf) } } } impl AsyncWrite for SuperSlow { fn is_write_vectored(&self) -> bool { false } fn poll_flush(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<Result<(), io::Error>> { Poll::Ready(Ok(())) } fn poll_shutdown(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<Result<(), io::Error>> { Poll::Ready(Ok(())) } fn poll_write( mut self: Pin<&mut Self>, cx: &mut Context<'_>, buf: &[u8], ) -> Poll<Result<usize, io::Error>> { if self.delays != 10 { cx.waker().wake_by_ref(); self.delays = 10; Poll::Pending } else { Poll::Ready(Ok(buf.len())) } } } const TO_SEND: &[u8] = &[1, 2, 3, 4]; #[tokio::test] async fn test_cancellation_safety() { let stream = SuperSlow::new(); let mut server = ServerBuilder::new().serve(stream); let mut cx = Context::from_waker(noop_waker_ref()); loop { // Cancellable futures should be possible to re-create at any time and resume as // if they were created once and then polled a few times if let Poll::Ready(val) = server.next().poll_unpin(&mut cx) { let msg = val.expect("eof").expect("err"); assert_eq!(&*msg.into_payload(), TO_SEND); break; } } }
rust
MIT
618599d15e9ee2a9a7191d081a56d90eae4a811a
2026-01-04T20:19:32.035255Z
false
Gelbpunkt/tokio-websockets
https://github.com/Gelbpunkt/tokio-websockets/blob/618599d15e9ee2a9a7191d081a56d90eae4a811a/tests/utf8_validation.rs
tests/utf8_validation.rs
// Autobahn does test all edge cases for UTF-8 validation - except one: // When a text message is split into multiple frames and the invalid UTF-8 is // part of a continuation frame, it only expects clients to fail fast after the // entire frame has been received. We should, however, fail immediately. #![cfg(feature = "server")] use bytes::{BufMut, Bytes, BytesMut}; use futures_util::StreamExt; use tokio::io::{AsyncWriteExt, duplex}; use tokio_websockets::{Error, ServerBuilder, proto::ProtocolError}; const MASK: [u8; 4] = [0, 0, 0, 0]; fn encode_frame(opcode: u8, payload: &[u8], payload_size: usize, is_final: bool) -> Bytes { let mut dst = BytesMut::new(); let initial_byte = (u8::from(is_final) << 7) + opcode; dst.put_u8(initial_byte); if u16::try_from(payload_size).is_err() { dst.put_u8(255); dst.put_u64(payload_size as u64); } else if payload_size > 125 { dst.put_u8(254); dst.put_u16(payload_size as u16); } else { dst.put_u8(payload_size as u8 + 128); } dst.extend_from_slice(&MASK); dst.extend_from_slice(payload); dst.freeze() } #[tokio::test] async fn test_utf8_fail_fast_incomplete_continuation() { let (one, mut two) = duplex(usize::MAX); let mut server = ServerBuilder::new().serve(one); // [240, 159, 152, 132] is a UTF-8 emoji (grinning face) let mut frame1_payload: Vec<u8> = std::iter::repeat_with(|| fastrand::alphanumeric() as u8) .take(4096) .collect(); // We omit the last byte of the emoji, since it is *already* invalid at the 0. // This should catch even more edge cases let frame3_payload = [159, 0]; let frame1 = encode_frame(1, &frame1_payload, 4096, false); frame1_payload[4095] = 240; // Second frame has a trailing partial UTF-8 codepoint let frame2 = encode_frame(0, &frame1_payload, 4096, false); let frame3 = encode_frame(0, &frame3_payload, 4096, false); // Pretend the rest of the payload will be written later two.write_all(&frame1).await.unwrap(); two.write_all(&frame2).await.unwrap(); two.write_all(&frame3).await.unwrap(); assert!(matches!( server.next().await, Some(Err(Error::Protocol(ProtocolError::InvalidUtf8))) )); }
rust
MIT
618599d15e9ee2a9a7191d081a56d90eae4a811a
2026-01-04T20:19:32.035255Z
false
Gelbpunkt/tokio-websockets
https://github.com/Gelbpunkt/tokio-websockets/blob/618599d15e9ee2a9a7191d081a56d90eae4a811a/fuzz/fuzz_targets/stream.rs
fuzz/fuzz_targets/stream.rs
#![no_main] extern crate tokio_websockets; use std::{ convert::TryFrom, future::Future, io, num::NonZeroUsize, pin::Pin, sync::{Arc, Mutex}, task::{Context, Poll}, }; use arbitrary::Arbitrary; use futures::{pin_mut, stream::StreamExt, FutureExt, SinkExt}; use libfuzzer_sys::fuzz_target; use tokio::io::{AsyncRead, AsyncWrite, ReadBuf}; use tokio_websockets::{CloseCode, Config, Limits, Message, WebSocketStream}; #[derive(Arbitrary, Debug)] enum ArbitraryMessage { Binary(Vec<u8>), Text(String), Ping(Vec<u8>), Pong(Vec<u8>), Close(u16, String), } impl From<ArbitraryMessage> for Message { fn from(message: ArbitraryMessage) -> Self { match message { ArbitraryMessage::Binary(payload) => Message::binary(payload), ArbitraryMessage::Text(payload) => Message::text(payload), ArbitraryMessage::Ping(mut payload) => { payload.truncate(125); Message::ping(payload) } ArbitraryMessage::Pong(mut payload) => { payload.truncate(125); Message::pong(payload) } ArbitraryMessage::Close(code, mut reason) => { for len in (119..=123).rev() { if reason.is_char_boundary(len) { reason.truncate(len); break; } } Message::close( CloseCode::try_from(code) .ok() .filter(|code| !code.is_reserved()), &reason, ) } } } } #[derive(Arbitrary, Debug)] struct Operation { kind: OperationKind, /// After polling this many times, the operation /// is canceled by dropping the future. max_polls: Option<u8>, } #[derive(Arbitrary, Debug)] enum OperationKind { Read, Write(ArbitraryMessage), Flush, Close, } #[derive(Arbitrary, Debug)] struct Workload { /// Fuzz the server, as opposed to the client. server: bool, data_to_read: Vec<u8>, io_behaviors: Vec<IoBehavior>, operations: Vec<Operation>, /// Zero would panic. frame_size: NonZeroUsize, /// Limit to a small int to avoid OOM. max_payload_len: u16, } struct DeterministicStream { /// The data that would be read by successful reads. data_to_read: Vec<u8>, /// The data written by successful writes. data_written: Arc<Mutex<Vec<u8>>>, /// Each stream IO will take one to decide how it behaves. io_behaviors: Vec<IoBehavior>, /// Return EOF. If false, returns pending when out of data. eof: bool, } #[derive(Arbitrary, Debug)] enum IoBehavior { /// Instantly read/write at most this many bytes, if applicable. Limit(usize), /// Instantly return an error. Error, /// Wake the waker immediately but return pending. Pending, } impl Default for IoBehavior { fn default() -> Self { Self::Limit(usize::MAX) } } /// Ready once limit is expired. struct PollLimiter { remaining_polls: Option<u8>, } impl Future for PollLimiter { type Output = (); fn poll(mut self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<Self::Output> { if let Some(remaining_polls) = &mut self.remaining_polls { if let Some(new) = remaining_polls.checked_sub(1) { *remaining_polls = new; Poll::Pending } else { Poll::Ready(()) } } else { Poll::Pending } } } fn new_stream( server: bool, config: Config, limits: Limits, stream: DeterministicStream, ) -> WebSocketStream<DeterministicStream> { if server { tokio_websockets::ServerBuilder::new() .config(config) .limits(limits) .serve(stream) } else { tokio_websockets::ClientBuilder::new() .config(config) .limits(limits) .take_over(stream) } } fn fuzz( Workload { server, data_to_read, io_behaviors, operations, frame_size, max_payload_len, }: Workload, ) { let data_written = Arc::new(Mutex::new(Vec::new())); let mut ws = new_stream( server, Config::default().frame_size(frame_size.get()), Limits::default().max_payload_len(Some(max_payload_len as _)), DeterministicStream { data_to_read, data_written: Arc::clone(&data_written), io_behaviors, eof: true, }, ); futures::executor::block_on(async move { for operation in operations { let future: Pin<Box<dyn Future<Output = ()>>> = match operation.kind { OperationKind::Read => Box::pin(ws.next().map(|_| ())), OperationKind::Write(message) => Box::pin(ws.feed(message.into()).map(|_| ())), OperationKind::Flush => Box::pin(ws.flush().map(|_| ())), OperationKind::Close => Box::pin(ws.close().map(|_| ())), }; let limit = PollLimiter { remaining_polls: operation.max_polls, }; pin_mut!(limit); tokio::select! { biased; _ = limit => {} _ = future => {} } } }); let data_written = std::mem::take(&mut *data_written.lock().unwrap()); // As a sanity check, make sure `data_written` is prefixed with 0 or more // *valid* messages, possibly followed by an incomplete *valid* message. let mut sanity_check = new_stream( !server, Default::default(), Default::default(), DeterministicStream { data_to_read: data_written.clone(), data_written: Default::default(), io_behaviors: Vec::new(), // Allow an incomplete message at the end. eof: false, }, ); futures::executor::block_on(async move { loop { let limit = PollLimiter { remaining_polls: Some(1), }; pin_mut!(limit); tokio::select! { biased; _ = limit => { // Pending means EOF; remaining message, if any, is incomplete. break; } result = sanity_check.next() => { // Any complete messages must be valid. if let Some(result) = result { result.unwrap_or_else(|e| panic!("{} {:?}", e, data_written)); } else { // Received close. break; } } } } }); } fuzz_target!(|workload: Workload| { fuzz(workload); }); impl AsyncRead for DeterministicStream { fn poll_read( mut self: Pin<&mut Self>, cx: &mut Context<'_>, buf: &mut ReadBuf<'_>, ) -> Poll<io::Result<()>> { match self.io_behaviors.pop().unwrap_or_default() { IoBehavior::Limit(limit) => { if !self.eof && self.data_to_read.is_empty() { cx.waker().wake_by_ref(); return Poll::Pending; } let end = buf.remaining().min(limit).min(self.data_to_read.len()); let remainder = self.data_to_read.split_off(end); buf.put_slice(&self.data_to_read); self.data_to_read = remainder; Poll::Ready(Ok(())) } IoBehavior::Error => Poll::Ready(Err(io::Error::new(io::ErrorKind::BrokenPipe, "???"))), IoBehavior::Pending => { cx.waker().wake_by_ref(); Poll::Pending } } } } impl AsyncWrite for DeterministicStream { fn poll_write( mut self: Pin<&mut Self>, cx: &mut Context<'_>, buf: &[u8], ) -> Poll<Result<usize, io::Error>> { match self.io_behaviors.pop().unwrap_or_default() { IoBehavior::Limit(limit) => { let amount = buf.len().min(limit); self.data_written .lock() .unwrap() .extend_from_slice(&buf[..amount]); Poll::Ready(Ok(amount)) } IoBehavior::Error => Poll::Ready(Err(io::Error::new(io::ErrorKind::BrokenPipe, "???"))), IoBehavior::Pending => { cx.waker().wake_by_ref(); Poll::Pending } } } fn poll_flush(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Result<(), io::Error>> { match self.io_behaviors.pop().unwrap_or_default() { IoBehavior::Limit(_) => Poll::Ready(Ok(())), IoBehavior::Error => Poll::Ready(Err(io::Error::new(io::ErrorKind::BrokenPipe, "???"))), IoBehavior::Pending => { cx.waker().wake_by_ref(); Poll::Pending } } } fn poll_shutdown( mut self: Pin<&mut Self>, cx: &mut Context<'_>, ) -> Poll<Result<(), io::Error>> { match self.io_behaviors.pop().unwrap_or_default() { IoBehavior::Limit(_) => Poll::Ready(Ok(())), IoBehavior::Error => Poll::Ready(Err(io::Error::new( io::ErrorKind::Unsupported, "unsupported", ))), IoBehavior::Pending => { cx.waker().wake_by_ref(); Poll::Pending } } } }
rust
MIT
618599d15e9ee2a9a7191d081a56d90eae4a811a
2026-01-04T20:19:32.035255Z
false
Gelbpunkt/tokio-websockets
https://github.com/Gelbpunkt/tokio-websockets/blob/618599d15e9ee2a9a7191d081a56d90eae4a811a/examples/utf8_benchmark_client.rs
examples/utf8_benchmark_client.rs
// This is a benchmark for utf-8 validation in tokio-websockets. // In order to properly be able to benchmark a WebSocket library, this client // must not use a WebSocket library. In the end, we want to benchmark the // server, not the client. // // The client sends a single WebSocket message over and over again. This message // is split into a configurable amount of frames and sent in a configurable // amount of chops. // It expects to receive the message back. Benchmark performance is measured in // messages sent by the server per second. use std::{ env, io::{self, Read, Write}, os::unix::net::UnixStream, time::Instant, }; use bytes::{BufMut, Bytes, BytesMut}; const DELIMITER: char = '+'; // No-op mask const MASK: [u8; 4] = [0, 0, 0, 0]; fn encode_message(data: Vec<u8>, frame_size: usize) -> Bytes { let mut dst = BytesMut::new(); let mut chunks = data.chunks(frame_size).peekable(); let mut next_chunk = Some(chunks.next().unwrap_or_default()); let mut chunk_number = 0; while let Some(chunk) = next_chunk { let opcode_value = if chunk_number == 0 { 1 } else { 0 }; let is_final = chunks.peek().is_none(); let chunk_size = chunk.len(); let initial_byte = (u8::from(is_final) << 7) + opcode_value; dst.put_u8(initial_byte); if u16::try_from(chunk_size).is_err() { dst.put_u8(255); dst.put_u64(chunk_size as u64); } else if chunk_size > 125 { dst.put_u8(254); dst.put_u16(chunk_size as u16); } else { dst.put_u8(chunk_size as u8 + 128); } dst.extend_from_slice(&MASK); dst.extend_from_slice(chunk); next_chunk = chunks.next(); chunk_number += 1; } dst.freeze() } fn main() -> io::Result<()> { let message_size: usize = env::var("MSG_SIZE") .ok() .and_then(|s| s.parse().ok()) .unwrap_or(64 * 1024 * 1024); let frame_size: usize = env::var("FRAME_SIZE") .ok() .and_then(|s| s.parse().ok()) .unwrap_or(4 * 1024); let chop_size = env::var("CHOP_SIZE") .ok() .and_then(|s| s.parse().ok()) .unwrap_or(1024); let mut payload: String = std::iter::repeat_with(fastrand::alphanumeric) .take(message_size - 1) .collect(); payload.push(DELIMITER); let encoded_message = encode_message(payload.into_bytes(), frame_size); // We use Unix sockets because they are a lot faster than TCP and have no // buffering let mut stream = UnixStream::connect("/tmp/tokio-websockets.sock")?; let start = Instant::now(); let mut messages_received = 0; // This has to be lower than encoded_message since we don't know how big the // client chunks the frames let mut buf = vec![0; message_size]; // We expect the server to assume that this is a websocket connection (i.e. skip // the handshake) loop { // Now just write the message in chops for chop in encoded_message.chunks(chop_size) { stream.write_all(chop)?; } loop { let n = stream.read(&mut buf)?; if n == 0 { panic!("should never happen"); } let last_byte_read = unsafe { buf.get_unchecked(n - 1) }; if *last_byte_read == DELIMITER as u8 { break; } } messages_received += 1; if messages_received % 100 == 0 { let time_taken = Instant::now().duration_since(start); let msg_per_sec = messages_received as f64 / time_taken.as_secs_f64(); println!("{messages_received} messages received: {msg_per_sec} msg/s"); } } }
rust
MIT
618599d15e9ee2a9a7191d081a56d90eae4a811a
2026-01-04T20:19:32.035255Z
false
Gelbpunkt/tokio-websockets
https://github.com/Gelbpunkt/tokio-websockets/blob/618599d15e9ee2a9a7191d081a56d90eae4a811a/examples/utf8_benchmark_server.rs
examples/utf8_benchmark_server.rs
use std::fs::remove_file; use futures_util::{SinkExt, StreamExt}; use tokio::net::UnixListener; use tokio_websockets::{Error, Limits, ServerBuilder}; #[tokio::main(flavor = "current_thread")] async fn main() -> Result<(), Error> { // The socket might still exist from an earlier run! let _ = remove_file("/tmp/tokio-websockets.sock"); let listener = UnixListener::bind("/tmp/tokio-websockets.sock")?; let (stream, _) = listener.accept().await?; let mut ws = ServerBuilder::new() .limits(Limits::unlimited()) .serve(stream); while let Some(Ok(msg)) = ws.next().await { ws.send(msg).await?; } Ok(()) }
rust
MIT
618599d15e9ee2a9a7191d081a56d90eae4a811a
2026-01-04T20:19:32.035255Z
false
Gelbpunkt/tokio-websockets
https://github.com/Gelbpunkt/tokio-websockets/blob/618599d15e9ee2a9a7191d081a56d90eae4a811a/examples/autobahn_server.rs
examples/autobahn_server.rs
use std::net::SocketAddr; use futures_util::{SinkExt, StreamExt}; use tokio::net::{TcpListener, TcpStream}; use tokio_websockets::{Error, Limits, ServerBuilder}; async fn accept_connection(stream: TcpStream) { if let Err(e) = handle_connection(stream).await { match e { Error::Protocol(_) => (), err => eprintln!("Error processing connection: {err:?}"), } } } async fn handle_connection(stream: TcpStream) -> Result<(), Error> { let (_request, mut ws_stream) = ServerBuilder::new() .limits(Limits::unlimited()) .accept(stream) .await?; while let Some(msg) = ws_stream.next().await { let msg = msg?; if msg.is_text() || msg.is_binary() { ws_stream.send(msg).await?; } } Ok(()) } #[tokio::main(flavor = "current_thread")] async fn main() { let addr: SocketAddr = ([127, 0, 0, 1], 9006).into(); let listener = TcpListener::bind(&addr).await.expect("Can't listen"); println!("Listening on: {addr}"); while let Ok((stream, _)) = listener.accept().await { tokio::spawn(accept_connection(stream)); } }
rust
MIT
618599d15e9ee2a9a7191d081a56d90eae4a811a
2026-01-04T20:19:32.035255Z
false
Gelbpunkt/tokio-websockets
https://github.com/Gelbpunkt/tokio-websockets/blob/618599d15e9ee2a9a7191d081a56d90eae4a811a/examples/echo_server.rs
examples/echo_server.rs
use std::net::SocketAddr; use futures_util::{SinkExt, StreamExt}; use tokio::net::TcpListener; use tokio_websockets::{Config, Error, Limits, ServerBuilder}; const PORT: u16 = 3000; async fn run() -> Result<(), Error> { let addr = SocketAddr::from(([127, 0, 0, 1], PORT)); let listener = TcpListener::bind(addr).await?; loop { let (conn, _) = listener.accept().await?; tokio::spawn(tokio::task::unconstrained(async move { let (_request, mut server) = ServerBuilder::new() .config(Config::default().frame_size(usize::MAX)) .limits(Limits::unlimited()) .accept(conn) .await .unwrap(); while let Some(Ok(item)) = server.next().await { if item.is_text() || item.is_binary() { server.send(item).await.unwrap(); } } })); } } fn main() -> Result<(), Error> { let rt = tokio::runtime::Builder::new_current_thread() .enable_io() .build() .unwrap(); rt.block_on(run()) }
rust
MIT
618599d15e9ee2a9a7191d081a56d90eae4a811a
2026-01-04T20:19:32.035255Z
false
Gelbpunkt/tokio-websockets
https://github.com/Gelbpunkt/tokio-websockets/blob/618599d15e9ee2a9a7191d081a56d90eae4a811a/examples/client.rs
examples/client.rs
use futures_util::{SinkExt, StreamExt}; use http::Uri; use tokio_websockets::{ClientBuilder, Error, Message}; #[tokio::main] async fn main() -> Result<(), Error> { let uri = Uri::from_static("ws://127.0.0.1:3000"); let (mut client, _) = ClientBuilder::from_uri(uri).connect().await?; client.send(Message::text("Hello, world!")).await?; while let Some(item) = client.next().await { println!("{item:?}"); } Ok(()) }
rust
MIT
618599d15e9ee2a9a7191d081a56d90eae4a811a
2026-01-04T20:19:32.035255Z
false
Gelbpunkt/tokio-websockets
https://github.com/Gelbpunkt/tokio-websockets/blob/618599d15e9ee2a9a7191d081a56d90eae4a811a/examples/native_tls_self_signed_client.rs
examples/native_tls_self_signed_client.rs
use futures_util::{SinkExt, StreamExt}; use http::Uri; use tokio_native_tls::native_tls::{Certificate, TlsConnector}; use tokio_websockets::{ClientBuilder, Error}; #[tokio::main] async fn main() -> Result<(), Error> { let uri = Uri::from_static("wss://127.0.0.1:8080"); let bytes = std::fs::read("certs/localhost.crt")?; let cert = Certificate::from_pem(&bytes)?; let connector = TlsConnector::builder().add_root_certificate(cert).build()?; let connector = tokio_websockets::Connector::NativeTls(connector.into()); let (mut client, _) = ClientBuilder::from_uri(uri) .connector(&connector) .connect() .await?; let msg = client.next().await; println!("Got message: {msg:?}"); client.close().await?; Ok(()) }
rust
MIT
618599d15e9ee2a9a7191d081a56d90eae4a811a
2026-01-04T20:19:32.035255Z
false
Gelbpunkt/tokio-websockets
https://github.com/Gelbpunkt/tokio-websockets/blob/618599d15e9ee2a9a7191d081a56d90eae4a811a/examples/server.rs
examples/server.rs
use futures_util::{SinkExt, StreamExt}; use tokio::net::TcpListener; use tokio_websockets::{Error, ServerBuilder}; #[tokio::main] async fn main() -> Result<(), Error> { let listener = TcpListener::bind("127.0.0.1:3000").await?; loop { let (conn, _) = listener.accept().await?; let (_request, mut server) = ServerBuilder::new().accept(conn).await?; while let Some(Ok(item)) = server.next().await { println!("Received: {item:?}"); server.send(item).await?; } } }
rust
MIT
618599d15e9ee2a9a7191d081a56d90eae4a811a
2026-01-04T20:19:32.035255Z
false
Gelbpunkt/tokio-websockets
https://github.com/Gelbpunkt/tokio-websockets/blob/618599d15e9ee2a9a7191d081a56d90eae4a811a/examples/rustls_server.rs
examples/rustls_server.rs
use std::{ fs::File, io::{self, BufReader}, net::SocketAddr, sync::Arc, }; use futures_util::SinkExt; use rustls_pemfile::{certs, pkcs8_private_keys}; use rustls_pki_types::{CertificateDer, PrivateKeyDer}; use tokio::net::TcpListener; use tokio_rustls::{TlsAcceptor, rustls}; use tokio_websockets::Message; const PATH_TO_CERT: &str = "certs/localhost.crt"; const PATH_TO_KEY: &str = "certs/localhost.key"; fn load_certs(path: &str) -> io::Result<Vec<CertificateDer<'static>>> { certs(&mut BufReader::new(File::open(path)?)).collect() } fn load_key(path: &str) -> io::Result<PrivateKeyDer<'static>> { pkcs8_private_keys(&mut BufReader::new(File::open(path)?)) .next() .unwrap() .map(Into::into) } #[tokio::main] async fn main() -> io::Result<()> { let addr = SocketAddr::from(([0, 0, 0, 0], 8080)); let certs = load_certs(PATH_TO_CERT)?; let key = load_key(PATH_TO_KEY)?; let config = rustls::ServerConfig::builder() .with_no_client_auth() .with_single_cert(certs, key) .map_err(|err| io::Error::new(io::ErrorKind::InvalidInput, err))?; let acceptor = TlsAcceptor::from(Arc::new(config)); let listener = TcpListener::bind(&addr).await?; loop { let (stream, _) = listener.accept().await?; let acceptor = acceptor.clone(); let fut = async move { let stream = acceptor.accept(stream).await?; let (_request, mut ws) = tokio_websockets::ServerBuilder::new() .accept(stream) .await?; // From here, do what you want with it ws.send(Message::text(String::from("Hello, world!"))) .await?; ws.close().await?; Ok(()) as Result<(), tokio_websockets::Error> }; tokio::spawn(async move { if let Err(err) = fut.await { eprintln!("{err:?}"); } }); } }
rust
MIT
618599d15e9ee2a9a7191d081a56d90eae4a811a
2026-01-04T20:19:32.035255Z
false
Gelbpunkt/tokio-websockets
https://github.com/Gelbpunkt/tokio-websockets/blob/618599d15e9ee2a9a7191d081a56d90eae4a811a/examples/autobahn_client.rs
examples/autobahn_client.rs
use std::str::FromStr; use futures_util::{SinkExt, StreamExt}; use http::Uri; use tokio_websockets::{ClientBuilder, Connector, Error, Limits}; async fn get_case_count() -> Result<u32, Error> { let uri = Uri::from_static("ws://localhost:9001/getCaseCount"); let (mut stream, _) = ClientBuilder::from_uri(uri) .connector(&Connector::Plain) .connect() .await?; let msg = stream.next().await.unwrap()?; stream.close().await.unwrap(); Ok(msg.as_text().unwrap().parse::<u32>().unwrap()) } async fn update_reports() -> Result<(), Error> { let uri = Uri::from_static("ws://localhost:9001/updateReports?agent=tokio-websockets"); let (mut stream, _) = ClientBuilder::from_uri(uri) .connector(&Connector::Plain) .connect() .await?; stream.close().await?; Ok(()) } async fn run_test(case: u32) -> Result<(), Error> { println!("Running test case {case}"); let uri = Uri::from_str(&format!( "ws://localhost:9001/runCase?case={case}&agent=tokio-websockets", )) .unwrap(); let (mut stream, _) = ClientBuilder::from_uri(uri) .limits(Limits::unlimited()) .connector(&Connector::Plain) .connect() .await?; while let Some(msg) = stream.next().await { let msg = msg?; if msg.is_text() || msg.is_binary() { stream.send(msg).await?; } } Ok(()) } #[tokio::main(flavor = "current_thread")] async fn main() -> Result<(), Error> { let total = get_case_count().await?; println!("Running {total} tests"); for case in 1..=total { if let Err(e) = run_test(case).await { match e { Error::Protocol(_) => {} _ => eprintln!("Testcase failed: {e:?}"), } }; } update_reports().await?; Ok(()) }
rust
MIT
618599d15e9ee2a9a7191d081a56d90eae4a811a
2026-01-04T20:19:32.035255Z
false
ha-shine/rustagram
https://github.com/ha-shine/rustagram/blob/37d54cbb1f2ca26f8b1c4e1e3e6c226ae0241297/src/producer.rs
src/producer.rs
extern crate image; extern crate rustagram; use rustagram::filters::{RustagramFilter}; use rustagram::filters::FilterType::*; fn main() { let img = image::open("test.jpg").unwrap(); let out = img.to_rgba().apply_filter(NineTeenSeventySeven); out.save("output/NineTeenSeventySeven.jpg").unwrap(); let out = img.to_rgba().apply_filter(Aden); out.save("output/Aden.jpg").unwrap(); let out = img.to_rgba().apply_filter(Brannan); out.save("output/Brannan.jpg").unwrap(); let out = img.to_rgba().apply_filter(Brooklyn); out.save("output/Brooklyn.jpg").unwrap(); let out = img.to_rgba().apply_filter(Clarendon); out.save("output/Clarendon.jpg").unwrap(); let out = img.to_rgba().apply_filter(Earlybird); out.save("output/Earlybird.jpg").unwrap(); let out = img.to_rgba().apply_filter(Gingham); out.save("output/Gingham.jpg").unwrap(); let out = img.to_rgba().apply_filter(Hudson); out.save("output/Hudson.jpg").unwrap(); let out = img.to_rgba().apply_filter(Inkwell); out.save("output/Inkwell.jpg").unwrap(); let out = img.to_rgba().apply_filter(Kelvin); out.save("output/Kelvin.jpg").unwrap(); let out = img.to_rgba().apply_filter(Lark); out.save("output/Lark.jpg").unwrap(); let out = img.to_rgba().apply_filter(Lofi); out.save("output/Lofi.jpg").unwrap(); let out = img.to_rgba().apply_filter(Maven); out.save("output/Maven.jpg").unwrap(); let out = img.to_rgba().apply_filter(Mayfair); out.save("output/Mayfair.jpg").unwrap(); let out = img.to_rgba().apply_filter(Moon); out.save("output/Moon.jpg").unwrap(); let out = img.to_rgba().apply_filter(Nashville); out.save("output/Nashville.jpg").unwrap(); let out = img.to_rgba().apply_filter(Reyes); out.save("output/Reyes.jpg").unwrap(); let out = img.to_rgba().apply_filter(Rise); out.save("output/Rise.jpg").unwrap(); let out = img.to_rgba().apply_filter(Slumber); out.save("output/Slumber.jpg").unwrap(); let out = img.to_rgba().apply_filter(Stinson); out.save("output/Stinson.jpg").unwrap(); let out = img.to_rgba().apply_filter(Toaster); out.save("output/Toaster.jpg").unwrap(); let out = img.to_rgba().apply_filter(Valencia); out.save("output/Valencia.jpg").unwrap(); let out = img.to_rgba().apply_filter(Walden); out.save("output/Walden.jpg").unwrap(); }
rust
Unlicense
37d54cbb1f2ca26f8b1c4e1e3e6c226ae0241297
2026-01-04T20:19:29.999195Z
false
ha-shine/rustagram
https://github.com/ha-shine/rustagram/blob/37d54cbb1f2ca26f8b1c4e1e3e6c226ae0241297/src/lib.rs
src/lib.rs
extern crate image; pub mod filters; mod rustaops; use filters::{FilterType}; pub fn validate_filter_type(filter: &str, filter_strings: &Vec<&str>, filter_types: &Vec<FilterType>) -> Result<FilterType, &'static str> { let search_result = filter_strings.iter().enumerate().find(|f| &filter == f.1); match search_result { Some((i,_)) => Ok(filter_types[i].clone()), None => Err("Invalid filter type") } }
rust
Unlicense
37d54cbb1f2ca26f8b1c4e1e3e6c226ae0241297
2026-01-04T20:19:29.999195Z
false
ha-shine/rustagram
https://github.com/ha-shine/rustagram/blob/37d54cbb1f2ca26f8b1c4e1e3e6c226ae0241297/src/filters.rs
src/filters.rs
extern crate image; use image::{RgbaImage, ConvertBuffer}; use image::imageops; use rustaops; #[derive(Clone)] pub enum FilterType { NineTeenSeventySeven, Aden, Brannan, Brooklyn, Clarendon, Earlybird, Gingham, Hudson, Inkwell, Kelvin, Lark, Lofi, Maven, Mayfair, Moon, Nashville, Reyes, Rise, Slumber, Stinson, Toaster, Valencia, Walden } pub trait RustagramFilter { fn apply_filter(&self, ft: FilterType) -> Self; } impl RustagramFilter for RgbaImage { fn apply_filter(&self, ft: FilterType) -> Self { match ft { FilterType::NineTeenSeventySeven => apply_1977(&self), FilterType::Aden => apply_aden(&self), FilterType::Brannan => apply_brannan(&self), FilterType::Brooklyn => apply_brooklyn(&self), FilterType::Clarendon => apply_clarendon(&self), FilterType::Earlybird => apply_earlybird(&self), FilterType::Gingham => apply_gingham(&self), FilterType::Hudson => apply_hudson(&self), FilterType::Inkwell => apply_inkwell(&self), FilterType::Kelvin => apply_kelvin(&self), FilterType::Lark => apply_lark(&self), FilterType::Lofi => apply_lofi(&self), FilterType::Maven => apply_maven(&self), FilterType::Mayfair => apply_mayfair(&self), FilterType::Moon => apply_moon(&self), FilterType::Nashville => apply_nashville(&self), FilterType::Reyes => apply_reyes(&self), FilterType::Rise => apply_rise(&self), FilterType::Slumber => apply_slumber(&self), FilterType::Stinson => apply_stinson(&self), FilterType::Toaster => apply_toaster(&self), FilterType::Valencia => apply_valencia(&self), FilterType::Walden => apply_walden(&self) } } } pub fn apply_1977(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let contrasted = imageops::contrast(img, 10.0); let brightened = rustaops::brighten_by_percent(&contrasted, 10.0); let saturated = rustaops::saturate(&brightened, 30.0); let foreground = rustaops::fill_with_channels(width, height, &[243,106,188,76]); let out = rustaops::blend_screen(&saturated, &foreground); out } pub fn apply_aden(img: &RgbaImage) -> RgbaImage { let huerotated = imageops::huerotate(img, -20); let contrasted = imageops::contrast(&huerotated, -10.0); let saturated = rustaops::saturate(&contrasted, -20.0); let brightened = rustaops::brighten_by_percent(&saturated, 20.0); let out = rustaops::restore_transparency(&brightened); out } pub fn apply_brannan(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let with_sepia = rustaops::sepia(img, 20.0); let contrasted = imageops::contrast(&with_sepia, 20.0); let foreground = rustaops::fill_with_channels(width, height, &[161,44,199,59]); let out = rustaops::blend_lighten(&foreground, &contrasted); out } pub fn apply_brooklyn(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let contrasted = imageops::contrast(img, -10.0); let brightened = rustaops::brighten_by_percent(&contrasted, 10.0); let foreground = rustaops::fill_with_channels(width, height, &[168,223,193,150]); let background = rustaops::restore_transparency(&brightened); let out = rustaops::blend_overlay(&foreground, &background); out } pub fn apply_clarendon(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let contrasted = imageops::contrast(img, 20.0); let saturated = rustaops::saturate(&contrasted, 35.0); let foreground = rustaops::fill_with_channels(width, height, &[127,187,227,101]); let out = rustaops::blend_overlay(&foreground, &saturated); out } pub fn apply_earlybird(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let contrasted = imageops::contrast(img, -10.0); let with_sepia = rustaops::sepia(&contrasted, 5.0); let foreground = rustaops::fill_with_channels(width, height, &[208,186,142,150]); let out = rustaops::blend_overlay(&with_sepia, &foreground); let out = rustaops::restore_transparency(&out); out } pub fn apply_gingham(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let brightened = rustaops::brighten_by_percent(img, 5.0); let background = imageops::huerotate(&brightened, -10); let foreground = rustaops::fill_with_channels(width, height, &[230, 230, 230, 255]); let out = rustaops::blend_soft_light(&foreground, &background); out } pub fn apply_hudson(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let brightened = rustaops::brighten_by_percent(img, 50.0); let constrasted = imageops::contrast(&brightened, -10.0); let saturated = rustaops::saturate(&constrasted, 10.0); let foreground = rustaops::fill_with_channels(width, height, &[166, 177, 255, 208]); let blended = rustaops::blend_multiply(&foreground, &saturated); let out = rustaops::restore_transparency(&blended); out } pub fn apply_inkwell(img: &RgbaImage) -> RgbaImage { let with_sepia = rustaops::sepia(img, 30.0); let contrasted = imageops::contrast(&with_sepia, 10.0); let brightened = rustaops::brighten_by_percent(&contrasted, 10.0); let out = imageops::grayscale(&brightened); ConvertBuffer::convert(&out) } pub fn apply_kelvin(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let foreground = rustaops::fill_with_channels(width, height, &[56, 44, 52, 255]); let color_dodged = rustaops::blend_color_dodge(img, &foreground); let foreground = rustaops::fill_with_channels(width, height, &[183, 125, 33, 255]); let out = rustaops::blend_overlay(&foreground, &color_dodged); out } pub fn apply_lark(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let contrasted = imageops::contrast(img, -10.0); let foreground = rustaops::fill_with_channels(width, height, &[34, 37, 63, 255]); let color_dodged = rustaops::blend_color_dodge(&contrasted, &foreground); let foreground = rustaops::fill_with_channels(width, height, &[242, 242, 242, 204]); let out = rustaops::blend_darken(&foreground, &color_dodged); out } pub fn apply_lofi(img: &RgbaImage) -> RgbaImage { let saturated = rustaops::saturate(img, 10.0); let out = imageops::contrast(&saturated, 50.0); out } pub fn apply_maven(img: &RgbaImage) -> RgbaImage { let with_sepia = rustaops::sepia(img, 25.0); let brightened = rustaops::brighten_by_percent(&with_sepia, -0.05); let contrasted = imageops::contrast(&brightened, -0.05); let out = rustaops::saturate(&contrasted, 50.0); out } pub fn apply_mayfair(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let contrasted = imageops::contrast(img, 10.0); let saturated = rustaops::saturate(&contrasted, 10.0); let foreground = rustaops::fill_with_channels(width, height, &[255,200,200,153]); let out = rustaops::blend_overlay(&foreground, &saturated); out } pub fn apply_moon(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let contrasted = imageops::contrast(img, 10.0); let brightened = rustaops::brighten_by_percent(&contrasted, 10.0); let foreground = rustaops::fill_with_channels(width, height, &[160,160,160,255]); let soft_light = rustaops::blend_soft_light(&foreground, &brightened); let foreground = rustaops::fill_with_channels(width, height, &[56,56,56,255]); let lighten = rustaops::blend_lighten(&foreground, &soft_light); let out = imageops::grayscale(&lighten); ConvertBuffer::convert(&out) } pub fn apply_nashville(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let with_sepia = rustaops::sepia(img, 2.0); let contrasted = imageops::contrast(&with_sepia, 20.0); let brightened = rustaops::brighten_by_percent(&contrasted, 5.0); let saturated = rustaops::saturate(&brightened, 20.0); let foreground = rustaops::fill_with_channels(width, height, &[247,176,153,243]); let darkened = rustaops::blend_darken(&foreground, &saturated); let foreground = rustaops::fill_with_channels(width, height, &[0,70,150,230]); let out = rustaops::blend_lighten(&foreground, &darkened); out } pub fn apply_reyes(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let with_sepia = rustaops::sepia(img, 22.0); let brightened = rustaops::brighten_by_percent(&with_sepia, 10.0); let contrast = imageops::contrast(&brightened, -15.0); let saturated = rustaops::saturate(&contrast, -25.0); let foreground = rustaops::fill_with_channels(width, height, &[239, 205, 173, 10]); let out = rustaops::over(&foreground, &saturated); out } pub fn apply_rise(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let brightened = rustaops::brighten_by_percent(img, 5.0); let with_sepia = rustaops::sepia(&brightened, 5.0); let contrast = imageops::contrast(&with_sepia, -10.0); let saturated = rustaops::saturate(&contrast, -10.0); let foreground = rustaops::fill_with_channels(width, height, &[236, 205, 169, 240]); let multiply = rustaops::blend_multiply(&foreground, &saturated); let foreground = rustaops::fill_with_channels(width, height, &[232, 197, 152, 10]); let overlaid = rustaops::blend_overlay(&foreground, &multiply); let out = rustaops::over(&overlaid, img); out } pub fn apply_slumber(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let saturated = rustaops::saturate(img, -34.0); let brightened = rustaops::brighten_by_percent(&saturated, 5.0); let foreground = rustaops::fill_with_channels(width, height, &[69, 41, 12, 102]); let lightened = rustaops::blend_lighten(&foreground, &brightened); let foreground = rustaops::fill_with_channels(width, height, &[125, 105, 24, 128]); let out = rustaops::blend_soft_light(&foreground, &lightened); out } pub fn apply_stinson(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let contrasted = imageops::contrast(img, -25.0); let saturated = rustaops::saturate(&contrasted, -15.0); let brightened = rustaops::brighten_by_percent(&saturated, 15.0); let foreground = rustaops::fill_with_channels(width, height, &[240, 149, 128, 51]); let out = rustaops::blend_soft_light(&foreground, &brightened); out } pub fn apply_toaster(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let contrasted = imageops::contrast(img, 20.0); let brightened = rustaops::brighten_by_percent(&contrasted, -10.0); let foreground = rustaops::fill_with_channels(width, height, &[128, 78, 15, 140]); let out = rustaops::blend_screen(&foreground, &brightened); out } pub fn apply_valencia(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let contrasted = imageops::contrast(img, 8.0); let brightened = rustaops::brighten_by_percent(&contrasted, 8.0); let sepia = rustaops::sepia(&brightened, 8.0); let foreground = rustaops::fill_with_channels(width, height, &[58, 3, 57, 128]); let out = rustaops::blend_exclusion(&foreground, &sepia); out } pub fn apply_walden(img: &RgbaImage) -> RgbaImage { let (width, height) = img.dimensions(); let brightened = rustaops::brighten_by_percent(img, 10.0); let huerotated = imageops::huerotate(&brightened, -10); let saturated = rustaops::saturate(&huerotated, 60.0); let sepia = rustaops::sepia(&saturated, 5.0); let foreground = rustaops::fill_with_channels(width, height, &[0, 88, 244, 77]); let out = rustaops::blend_screen(&foreground, &sepia); out }
rust
Unlicense
37d54cbb1f2ca26f8b1c4e1e3e6c226ae0241297
2026-01-04T20:19:29.999195Z
false
ha-shine/rustagram
https://github.com/ha-shine/rustagram/blob/37d54cbb1f2ca26f8b1c4e1e3e6c226ae0241297/src/main.rs
src/main.rs
extern crate image; extern crate rustagram; #[macro_use] extern crate clap; use rustagram::filters::{RustagramFilter}; use rustagram::filters::FilterType::*; use std::process; fn main() { let filter_strings = vec![ "1977", "aden", "brannan", "brooklyn", "clarendon", "earlybird", "gingham", "hudson", "inkwell", "kelvin", "lark", "lofi", "maven", "mayfair", "moon", "nashville", "reyes", "rise", "slumber", "stinson", "toaster", "valencia", "walden" ]; let filters = vec![ NineTeenSeventySeven, Aden, Brannan, Brooklyn, Clarendon, Earlybird, Gingham, Hudson, Inkwell, Kelvin, Lark, Lofi, Maven, Mayfair, Moon, Nashville, Reyes, Rise, Slumber, Stinson, Toaster, Valencia, Walden ]; let filter_arg = &format!("Filter name: {}", filter_strings.join(", ")); let matches = clap_app!(myapp => (version: "1.0") (author: "Htet Aung Shine <h@shine.rocks>") (about: "Apply instagram filters to you photos") (@arg OUTPUT: -o --out "Output file name") (@arg INPUT: +required "Path to the input image file") (@arg FILTER: +required filter_arg) ).get_matches(); let output = matches.value_of("OUTPUT").unwrap_or("output.jpg"); let input = matches.value_of("INPUT").unwrap(); let filter = matches.value_of("FILTER").unwrap(); let filter_type = match rustagram::validate_filter_type(filter, &filter_strings, &filters) { Ok(item) => item, Err(msg) => { eprintln!("{}", msg); process::exit(1); } }; let img = image::open(input).unwrap(); let out = img.to_rgba().apply_filter(filter_type); out.save(output).unwrap(); }
rust
Unlicense
37d54cbb1f2ca26f8b1c4e1e3e6c226ae0241297
2026-01-04T20:19:29.999195Z
false
ha-shine/rustagram
https://github.com/ha-shine/rustagram/blob/37d54cbb1f2ca26f8b1c4e1e3e6c226ae0241297/src/rustaops/mod.rs
src/rustaops/mod.rs
extern crate image; use image::{GenericImage, ImageBuffer, Pixel, Rgba}; use image::math::utils::clamp; mod blend; pub fn brighten_by_percent<I, P>(image: &I, value: f32) -> ImageBuffer<P, Vec<u8>> where I: GenericImage<Pixel=P>, P: Pixel<Subpixel=u8> + 'static { let (width, height) = image.dimensions(); let mut out = ImageBuffer::new(width, height); let max = u8::max_value(); let max: f32 = max as f32; let percent = (value + 100.0) / 100.0; for y in 0..height { for x in 0..width { let e = image.get_pixel(x, y).map_with_alpha(|b| { let c: f32 = b as f32; let d = clamp(c * percent, 0.0, max); d as u8 }, |alpha| alpha); out.put_pixel(x, y, e); } } out } pub fn sepia<I>(image: &I, intensity: f32) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { let (width, height) = image.dimensions(); let depth = 20; let mut out = ImageBuffer::new(width, height); let percent = intensity / 100.0; for (x, y, pixel) in out.enumerate_pixels_mut() { let channels = image.get_pixel(x, y).data; let mut r = channels[0] as u16; let mut g = channels[1] as u16; let mut b = channels[2] as u16; let gray = (r + g + b) / 3; r = r + (depth * 2); g = g + depth; b = gray; if r > 255 { r = 255 } if g > 255 { g = 255 } if b > 255 { b = 255 } let f = b as f32; b = (f - (f * percent)) as u16; if b > 255 { b = 255 } *pixel = Rgba([r as u8, g as u8, b as u8, channels[3]]); } out } pub fn fill_with_channels(width: u32, height: u32, channels: &[u8; 4]) -> ImageBuffer<Rgba<u8>, Vec<u8>> { let a = channels[3] as f32; let r = ((channels[0] as f32) * a / 255.0) as u8; let g = ((channels[1] as f32) * a / 255.0) as u8; let b = ((channels[2] as f32) * a / 255.0) as u8; let fill = [r, g, b, channels[3]]; let mut out = ImageBuffer::new(width, height); for y in 0..height { for x in 0..width { out.put_pixel(x, y, *Rgba::from_slice(&fill)); } } out } pub fn restore_transparency<I>(image: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { let (width, height) = image.dimensions(); let mut out = ImageBuffer::new(width, height); for y in 0..height { for x in 0..width { let mut e = image.get_pixel(x, y).data; e[3] = 255; out.put_pixel(x, y, *Rgba::from_slice(&e)); } } out } pub fn over<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { let (width, height) = foreground.dimensions(); let mut out = ImageBuffer::new(width, height); for (x, y, pixel) in out.enumerate_pixels_mut() { let fg_data = foreground.get_pixel(x, y).data; let bg_data = background.get_pixel(x, y).data; let final_alpha = blend::compute_final_alpha(&fg_data, &bg_data); let mut final_data = [0; 4]; final_data[3] = final_alpha; for i in 0..3 { let fg_c = fg_data[i] as f32 / 255.0; let bg_c = bg_data[i] as f32 / 255.0; let final_c = fg_c + bg_c * (1.0 - fg_c); final_data[i] = (final_c * 255.0) as u8; } *pixel = Rgba(final_data); } out } #[allow(dead_code)] pub fn blend_screen<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { process_blend(foreground, background, &blend::blend_screen) } #[allow(dead_code)] pub fn blend_soft_light<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { process_blend(foreground, background, &blend::blend_soft_light) } #[allow(dead_code)] pub fn blend_overlay<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { process_blend(foreground, background, &blend::blend_overlay) } #[allow(dead_code)] pub fn blend_color_dodge<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { process_blend(foreground, background, &blend::blend_color_dodge) } #[allow(dead_code)] pub fn blend_darken<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { process_blend(foreground, background, &blend::blend_darken) } #[allow(dead_code)] pub fn blend_lighten<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { process_blend(foreground, background, &blend::blend_lighten) } #[allow(dead_code)] pub fn blend_multiply<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { process_blend(foreground, background, &blend::blend_multiply) } #[allow(dead_code)] pub fn blend_color_burn<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { process_blend(foreground, background, &blend::blend_color_burn) } #[allow(dead_code)] pub fn blend_linear_burn<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { process_blend(foreground, background, &blend::blend_linear_burn) } #[allow(dead_code)] pub fn blend_linear_dodge<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { process_blend(foreground, background, &blend::blend_linear_dodge) } #[allow(dead_code)] pub fn blend_hard_light<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { process_blend(foreground, background, &blend::blend_hard_light) } #[allow(dead_code)] pub fn blend_vivid_light<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { process_blend(foreground, background, &blend::blend_vivid_light) } #[allow(dead_code)] pub fn blend_linear_light<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { process_blend(foreground, background, &blend::blend_linear_light) } #[allow(dead_code)] pub fn blend_pin_light<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { process_blend(foreground, background, &blend::blend_pin_light) } #[allow(dead_code)] pub fn blend_difference<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { process_blend(foreground, background, &blend::blend_difference) } #[allow(dead_code)] pub fn blend_exclusion<I>(foreground: &I, background: &I) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { process_blend(foreground, background, &blend::blend_exclusion) } fn process_blend<I>(foreground: &I, background: &I, f: &Fn(u8, u8) -> u8) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { let (width, height) = foreground.dimensions(); let mut out = ImageBuffer::new(width, height); for (x, y, pixel) in out.enumerate_pixels_mut() { let fg_data = foreground.get_pixel(x, y).data; let bg_data = background.get_pixel(x, y).data; let final_r = f(fg_data[0], bg_data[0]); let final_g = f(fg_data[1], bg_data[1]); let final_b = f(fg_data[2], bg_data[2]); let final_alpha = blend::compute_final_alpha(&fg_data, &bg_data); *pixel = Rgba([final_r, final_g, final_b, final_alpha]); } out } fn saturate_value(s: f32, percent: f32) -> f32 { let mut s = s; if percent >= 0.0 { let interval = 1.0 - s; s = s + percent * interval * s; } else { s = s + percent * s; } s } fn rgb_to_hls(rgba: &[u8; 4]) -> [f32; 3] { let r = rgba[0] as f32 / 255.0; let g = rgba[1] as f32 / 255.0; let b = rgba[2] as f32 / 255.0; let max = float_max(r, g, b); let min = float_min(r, g, b); let mut hue = 0.0; let mut saturation = 0.0; let lumination = (max + min) / 2.0; if max == min { return [hue, lumination, saturation] } let delta = max - min; if lumination < 0.5 { saturation = delta / (max + min); } else { saturation = delta / (2.0 - max - min); } if r == max { hue = (g - b) / delta; } else if g == max { hue = 2.0 + (b - r) / delta; } else { hue = 4.0 + (r - g) / delta; } hue /= 6.0; if hue < 0.0 { hue += 1.0; } return [hue, lumination, saturation] } pub fn saturate<I>(image: &I, value: f32) -> ImageBuffer<Rgba<u8>, Vec<u8>> where I: GenericImage<Pixel=Rgba<u8>> { let (width, height) = image.dimensions(); let mut out = ImageBuffer::new(width, height); let percent = value / 100.0; for (x, y, pixel) in out.enumerate_pixels_mut() { let data = image.get_pixel(x, y).data; let mut hls = rgb_to_hls(&data); hls[2] = saturate_value(hls[2], percent); let rgb = hls_to_rgb(&hls, data[3]); *pixel = Rgba(rgb); } out } fn hls_to_rgb(hsl: &[f32; 3], alpha: u8) -> [u8; 4] { let (r,g,b,m1,m2); let hue = hsl[0]; let lumination = hsl[1]; let saturation = hsl[2]; if saturation == 0.0 { r = lumination; g = lumination; b = lumination; } else { if lumination <= 0.5 { m2 = lumination * (1.0 + saturation); } else { m2 = lumination + saturation - lumination * saturation; } m1 = 2.0 * lumination - m2; r = hue_to_rgb(m1, m2, hue + (1.0/3.0)); g = hue_to_rgb(m1, m2, hue); b = hue_to_rgb(m1, m2, hue - (1.0/3.0)); } let red = (r * 255.0) as u8; let green = (g * 255.0) as u8; let blue = (b * 255.0) as u8; [red, green, blue, alpha] } fn hue_to_rgb(m1: f32, m2: f32, hue: f32) -> f32 { let mut hue = hue; if hue < 0.0 { hue += 1.0; } else if hue > 1.0 { hue -= 1.0; } if (6.0 * hue) < 1.0 { return m1 + (m2 - m1) * hue * 6.0 } else if (2.0 * hue) < 1.0 { return m2 } else if (3.0 * hue) < 2.0 { return m1 + (m2 - m1) * ((2.0/3.0) - hue) * 6.0 } else { return m1 } } fn float_max(a: f32, b: f32, c: f32) -> f32 { if a >= b && a >= c { a } else if b >= a && b >= c { b } else { c } } fn float_min(a: f32, b: f32, c: f32) -> f32 { if a <= b && a <= c { a } else if b <= a && b <= c { b } else { c } }
rust
Unlicense
37d54cbb1f2ca26f8b1c4e1e3e6c226ae0241297
2026-01-04T20:19:29.999195Z
false
ha-shine/rustagram
https://github.com/ha-shine/rustagram/blob/37d54cbb1f2ca26f8b1c4e1e3e6c226ae0241297/src/rustaops/blend.rs
src/rustaops/blend.rs
use std::cmp; #[allow(dead_code)] pub fn compute_final_alpha(fg: &[u8; 4], bg: &[u8; 4]) -> u8 { let fg_alpha = fg[3] as f32 / 255.0; let bg_alpha = bg[3] as f32 / 255.0; let final_alpha = fg_alpha + bg_alpha * (1.0 - fg_alpha); (final_alpha * 255.0) as u8 } #[allow(dead_code)] pub fn blend_screen(x1: u8, x2: u8) -> u8 { let x1: u16 = x1 as u16; let x2: u16 = x2 as u16; (255 - ((255 - x1).wrapping_mul(255 - x2)>>8)) as u8 } #[allow(dead_code)] pub fn blend_darken(x1: u8, x2: u8) -> u8 { cmp::min(x1, x2) } #[allow(dead_code)] pub fn blend_lighten(x1: u8, x2: u8) -> u8 { cmp::max(x1, x2) } #[allow(dead_code)] pub fn blend_multiply(x1: u8, x2: u8) -> u8 { let x1 = x1 as u16; let x2 = x2 as u16; ((x1 * x2) / 255) as u8 } #[allow(dead_code)] pub fn blend_color_burn(x1: u8, x2: u8) -> u8 { if x2 == 0 { x2 } else { let x1 = x1 as u16; let x2 = x2 as u16; let max = 255 as u16; let rhs = max.wrapping_sub(((max - x1)<<8)/x2); if rhs > 0 { rhs as u8 } else { 0 } } } #[allow(dead_code)] pub fn blend_subtract(x1: u8, x2: u8) -> u8 { let x1 = x1 as u16; let x2 = x2 as u16; let lhs = x1 + x2; if lhs < 255 { 0 } else { (lhs - 255) as u8 } } #[allow(dead_code)] pub fn blend_linear_burn(x1: u8, x2: u8) -> u8 { blend_subtract(x2, x1) } #[allow(dead_code)] pub fn blend_color_dodge(x1: u8, x2: u8) -> u8 { if x2 == 255 { x2 } else { let x1: u16 = x1 as u16; let x2: u16 = x2 as u16; let rhs = (x1<<8)/(255-x2); if 255 < rhs { 255 } else { rhs as u8 } } } #[allow(dead_code)] pub fn blend_add(x1: u8, x2: u8) -> u8 { let rhs = x1.wrapping_add(x2); rhs } #[allow(dead_code)] pub fn blend_linear_dodge(x1: u8, x2: u8) -> u8 { blend_add(x2, x1) } #[allow(dead_code)] pub fn blend_overlay(x1: u8, x2: u8) -> u8 { let x1 = x1 as u16; let x2 = x2 as u16; if x2 < 128 { (2 * x1 * x2 / 255) as u8 } else { (255 - 2 * (255 - x1) * (255 - x2) / 255) as u8 } } #[allow(dead_code)] pub fn blend_soft_light(x1: u8, x2: u8) -> u8 { let f2 = x2 as f32; let shifted = (x1>>1) as f32; if x2 < 128 { ((2.0*((shifted)+64.0))*(f2/255.0)) as u8 } else { (255.0-(2.0*(255.0-((shifted)+64.0))*(255.0-f2)/255.0)) as u8 } } #[allow(dead_code)] pub fn blend_hard_light(x1: u8, x2: u8) -> u8 { blend_overlay(x2, x1) } #[allow(dead_code)] pub fn blend_vivid_light(x1: u8, x2: u8) -> u8 { if x2 < 128 { blend_color_burn(x1, x2 * 2) } else { blend_color_dodge(x1, 2 * (x2 - 128)) } } #[allow(dead_code)] pub fn blend_linear_light(x1: u8, x2: u8) -> u8 { if x2 < 128 { blend_linear_burn(x1, 2 * x2) } else { blend_linear_dodge(x1, 2 * (x2 - 128)) } } #[allow(dead_code)] pub fn blend_pin_light(x1: u8, x2: u8) -> u8 { if x2 < 128 { blend_darken(x1, 2 * x2) } else { blend_lighten(x1, 2 * (x2 - 128)) } } #[allow(dead_code)] pub fn blend_difference(x1: u8, x2: u8) -> u8 { let x1 = x1 as i16; let x2 = x2 as i16; (x1 - x2).abs() as u8 } #[allow(dead_code)] pub fn blend_exclusion(x1: u8, x2: u8) -> u8 { let x1 = x1 as u32; let x2 = x2 as u32; (x1 + x2 - 2 * x1 * x2 / 255) as u8 }
rust
Unlicense
37d54cbb1f2ca26f8b1c4e1e3e6c226ae0241297
2026-01-04T20:19:29.999195Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/node/build.rs
node/build.rs
use substrate_build_script_utils::{generate_cargo_keys, rerun_if_git_head_changed}; fn main() { generate_cargo_keys(); rerun_if_git_head_changed(); }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/node/src/command.rs
node/src/command.rs
use contracts_parachain_runtime::Block; use cumulus_primitives_core::ParaId; use frame_benchmarking_cli::{BenchmarkCmd, SUBSTRATE_REFERENCE_HARDWARE}; use log::info; use sc_cli::{ ChainSpec, CliConfiguration, DefaultConfigurationValues, ImportParams, KeystoreParams, NetworkParams, Result, RpcEndpoint, SharedParams, SubstrateCli, }; use sc_service::config::{BasePath, PrometheusConfig}; use crate::{ chain_spec, cli::{Cli, RelayChainCli, Subcommand}, service::{dev, new_partial}, }; fn load_spec(id: &str) -> std::result::Result<Box<dyn ChainSpec>, String> { Ok(match id { "" | "dev" => Box::new(chain_spec::dev::development_config()?), "contracts-parachain-local" => Box::new(chain_spec::local_testnet_config()), path => Box::new(chain_spec::ChainSpec::from_json_file(std::path::PathBuf::from(path))?), }) } impl SubstrateCli for Cli { fn impl_name() -> String { "Substrate Contracts Node".into() } fn impl_version() -> String { env!("SUBSTRATE_CLI_IMPL_VERSION").into() } fn description() -> String { env!("CARGO_PKG_DESCRIPTION").into() } fn author() -> String { env!("CARGO_PKG_AUTHORS").into() } fn support_url() -> String { "https://github.com/paritytech/substrate-contracts-node/issues/new".into() } fn copyright_start_year() -> i32 { 2021 } fn load_spec(&self, id: &str) -> std::result::Result<Box<dyn sc_service::ChainSpec>, String> { load_spec(id) } } impl SubstrateCli for RelayChainCli { fn impl_name() -> String { "Substrate Contracts Node".into() } fn impl_version() -> String { env!("SUBSTRATE_CLI_IMPL_VERSION").into() } fn description() -> String { env!("CARGO_PKG_DESCRIPTION").into() } fn author() -> String { env!("CARGO_PKG_AUTHORS").into() } fn support_url() -> String { "https://github.com/paritytech/substrate-contracts-node/issues/new".into() } fn copyright_start_year() -> i32 { 2020 } fn load_spec(&self, id: &str) -> std::result::Result<Box<dyn sc_service::ChainSpec>, String> { polkadot_cli::Cli::from_iter([RelayChainCli::executable_name()].iter()).load_spec(id) } } macro_rules! construct_async_run { (|$components:ident, $cli:ident, $cmd:ident, $config:ident| $( $code:tt )* ) => {{ let runner = $cli.create_runner($cmd)?; runner.async_run(|$config| { let $components = new_partial(&$config)?; let task_manager = $components.task_manager; { $( $code )* }.map(|v| (v, task_manager)) }) }} } /// Parse command line arguments into service configuration. pub fn run() -> Result<()> { let mut cli = Cli::from_args(); if cli.run.base.shared_params.chain.is_none() { log::debug!("forcing dev mode"); cli.run.base.shared_params.dev = true; } // remove block production noise and output contracts debug buffer by default if cli.run.base.shared_params.log.is_empty() { cli.run.base.shared_params.log = vec![ "runtime::contracts=debug".into(), "sc_cli=info".into(), "sc_rpc_server=info".into(), "warn".into(), ]; } match &cli.subcommand { Some(Subcommand::BuildSpec(cmd)) => { let runner = cli.create_runner(cmd)?; runner.sync_run(|config| cmd.run(config.chain_spec, config.network)) }, Some(Subcommand::CheckBlock(cmd)) => { construct_async_run!(|components, cli, cmd, config| { Ok(cmd.run(components.client, components.import_queue)) }) }, Some(Subcommand::ExportBlocks(cmd)) => { construct_async_run!(|components, cli, cmd, config| { Ok(cmd.run(components.client, config.database)) }) }, Some(Subcommand::ExportState(cmd)) => { construct_async_run!(|components, cli, cmd, config| { Ok(cmd.run(components.client, config.chain_spec)) }) }, Some(Subcommand::ImportBlocks(cmd)) => { construct_async_run!(|components, cli, cmd, config| { Ok(cmd.run(components.client, components.import_queue)) }) }, Some(Subcommand::Revert(cmd)) => { construct_async_run!(|components, cli, cmd, config| { Ok(cmd.run(components.client, components.backend, None)) }) }, Some(Subcommand::PurgeChain(cmd)) => { let runner = cli.create_runner(cmd)?; runner.sync_run(|config| { let polkadot_cli = RelayChainCli::new( &config, [RelayChainCli::executable_name()].iter().chain(cli.relay_chain_args.iter()), ); let polkadot_config = SubstrateCli::create_configuration( &polkadot_cli, &polkadot_cli, config.tokio_handle.clone(), ) .map_err(|err| format!("Relay chain argument error: {}", err))?; cmd.run(config, polkadot_config) }) }, Some(Subcommand::ExportGenesisHead(cmd)) => { let runner = cli.create_runner(cmd)?; runner.sync_run(|config| { let partials = new_partial(&config)?; cmd.run(partials.client) }) }, Some(Subcommand::ExportGenesisWasm(cmd)) => { let runner = cli.create_runner(cmd)?; runner.sync_run(|_config| { let spec = cli.load_spec(&cmd.shared_params.chain.clone().unwrap_or_default())?; cmd.run(&*spec) }) }, Some(Subcommand::Benchmark(cmd)) => { let runner = cli.create_runner(cmd)?; // Switch on the concrete benchmark sub-command- match cmd { BenchmarkCmd::Pallet(cmd) => if cfg!(feature = "runtime-benchmarks") { runner.sync_run(|config| { cmd.run_with_spec::<sp_runtime::traits::HashingFor<Block>, ()>(Some( config.chain_spec, )) }) } else { Err("Benchmarking wasn't enabled when building the node. \ You can enable it with `--features runtime-benchmarks`." .into()) }, BenchmarkCmd::Block(cmd) => runner.sync_run(|config| { let partials = new_partial(&config)?; cmd.run(partials.client) }), #[cfg(not(feature = "runtime-benchmarks"))] BenchmarkCmd::Storage(_) => Err(sc_cli::Error::Input( "Compile with --features=runtime-benchmarks \ to enable storage benchmarks." .into(), )), #[cfg(feature = "runtime-benchmarks")] BenchmarkCmd::Storage(cmd) => runner.sync_run(|config| { let partials = new_partial(&config)?; let db = partials.backend.expose_db(); let storage = partials.backend.expose_storage(); cmd.run(config, partials.client.clone(), db, storage) }), BenchmarkCmd::Machine(cmd) => runner.sync_run(|config| cmd.run(&config, SUBSTRATE_REFERENCE_HARDWARE.clone())), // NOTE: this allows the Client to leniently implement // new benchmark commands without requiring a companion MR. #[allow(unreachable_patterns)] _ => Err("Benchmarking sub-command unsupported".into()), } }, None => { let runner = cli.create_runner(&cli.run.normalize())?; let collator_options = cli.run.collator_options(); runner.run_node_until_exit(|config| async move { if config.chain_spec.name() == "Development" { return dev::new_full::<sc_network::NetworkWorker<_, _>>( config, cli.finalize_delay_sec.into(), ) .map_err(sc_cli::Error::Service); } let hwbench = (!cli.no_hardware_benchmarks) .then_some(config.database.path().map(|database_path| { let _ = std::fs::create_dir_all(database_path); sc_sysinfo::gather_hwbench( Some(database_path), &SUBSTRATE_REFERENCE_HARDWARE, ) })) .flatten(); let para_id = chain_spec::Extensions::try_get(&*config.chain_spec) .map(|e| e.para_id) .ok_or("Could not find parachain ID in chain-spec.")?; let polkadot_cli = RelayChainCli::new( &config, [RelayChainCli::executable_name()].iter().chain(cli.relay_chain_args.iter()), ); let id = ParaId::from(para_id); let tokio_handle = config.tokio_handle.clone(); let polkadot_config = SubstrateCli::create_configuration(&polkadot_cli, &polkadot_cli, tokio_handle) .map_err(|err| format!("Relay chain argument error: {}", err))?; info!("Is collating: {}", if config.role.is_authority() { "yes" } else { "no" }); crate::service::start_parachain_node( config, polkadot_config, collator_options, id, hwbench, ) .await .map(|r| r.0) .map_err(Into::into) }) }, } } impl DefaultConfigurationValues for RelayChainCli { fn p2p_listen_port() -> u16 { 30334 } fn rpc_listen_port() -> u16 { 9945 } fn prometheus_listen_port() -> u16 { 9616 } } impl CliConfiguration<Self> for RelayChainCli { fn shared_params(&self) -> &SharedParams { self.base.base.shared_params() } fn import_params(&self) -> Option<&ImportParams> { self.base.base.import_params() } fn network_params(&self) -> Option<&NetworkParams> { self.base.base.network_params() } fn keystore_params(&self) -> Option<&KeystoreParams> { self.base.base.keystore_params() } fn base_path(&self) -> Result<Option<BasePath>> { Ok(self .shared_params() .base_path()? .or_else(|| self.base_path.clone().map(Into::into))) } fn rpc_addr(&self, default_listen_port: u16) -> Result<Option<Vec<RpcEndpoint>>> { self.base.base.rpc_addr(default_listen_port) } fn prometheus_config( &self, default_listen_port: u16, chain_spec: &Box<dyn ChainSpec>, ) -> Result<Option<PrometheusConfig>> { self.base.base.prometheus_config(default_listen_port, chain_spec) } fn init<F>(&self, _support_url: &String, _impl_version: &String, _logger_hook: F) -> Result<()> where F: FnOnce(&mut sc_cli::LoggerBuilder), { unreachable!("PolkadotCli is never initialized; qed"); } fn chain_id(&self, is_dev: bool) -> Result<String> { let chain_id = self.base.base.chain_id(is_dev)?; Ok(if chain_id.is_empty() { self.chain_id.clone().unwrap_or_default() } else { chain_id }) } fn role(&self, is_dev: bool) -> Result<sc_service::Role> { self.base.base.role(is_dev) } fn transaction_pool(&self, is_dev: bool) -> Result<sc_service::config::TransactionPoolOptions> { self.base.base.transaction_pool(is_dev) } fn trie_cache_maximum_size(&self) -> Result<Option<usize>> { self.base.base.trie_cache_maximum_size() } fn rpc_methods(&self) -> Result<sc_service::config::RpcMethods> { self.base.base.rpc_methods() } fn rpc_max_connections(&self) -> Result<u32> { self.base.base.rpc_max_connections() } fn rpc_cors(&self, is_dev: bool) -> Result<Option<Vec<String>>> { self.base.base.rpc_cors(is_dev) } fn default_heap_pages(&self) -> Result<Option<u64>> { self.base.base.default_heap_pages() } fn force_authoring(&self) -> Result<bool> { self.base.base.force_authoring() } fn disable_grandpa(&self) -> Result<bool> { self.base.base.disable_grandpa() } fn max_runtime_instances(&self) -> Result<Option<usize>> { self.base.base.max_runtime_instances() } fn announce_block(&self) -> Result<bool> { self.base.base.announce_block() } fn telemetry_endpoints( &self, chain_spec: &Box<dyn ChainSpec>, ) -> Result<Option<sc_telemetry::TelemetryEndpoints>> { self.base.base.telemetry_endpoints(chain_spec) } fn node_name(&self) -> Result<String> { self.base.base.node_name() } }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/node/src/chain_spec.rs
node/src/chain_spec.rs
pub mod dev; use contracts_parachain_runtime::{AccountId, AuraId, Signature, EXISTENTIAL_DEPOSIT}; use cumulus_primitives_core::ParaId; use sc_chain_spec::{ChainSpecExtension, ChainSpecGroup}; use sc_service::ChainType; use serde::{Deserialize, Serialize}; use sp_core::{sr25519, Pair, Public}; use sp_runtime::traits::{IdentifyAccount, Verify}; /// Specialized `ChainSpec` for the normal parachain runtime. pub type ChainSpec = sc_service::GenericChainSpec<Extensions>; /// The default XCM version to set in genesis config. const SAFE_XCM_VERSION: u32 = xcm::prelude::XCM_VERSION; /// Helper function to generate a crypto pair from seed pub fn get_from_seed<TPublic: Public>(seed: &str) -> <TPublic::Pair as Pair>::Public { TPublic::Pair::from_string(&format!("//{}", seed), None) .expect("static values are valid; qed") .public() } /// The extensions for the [`ChainSpec`]. #[derive(Debug, Clone, PartialEq, Serialize, Deserialize, ChainSpecGroup, ChainSpecExtension)] #[serde(deny_unknown_fields)] pub struct Extensions { /// The relay chain of the Parachain. #[serde(alias = "relayChain", alias = "RelayChain")] pub relay_chain: String, /// The id of the Parachain. #[serde(alias = "paraId", alias = "ParaId")] pub para_id: u32, } impl Extensions { /// Try to get the extension from the given `ChainSpec`. pub fn try_get(chain_spec: &dyn sc_service::ChainSpec) -> Option<&Self> { sc_chain_spec::get_extension(chain_spec.extensions()) } } type AccountPublic = <Signature as Verify>::Signer; /// Generate collator keys from seed. /// /// This function's return type must always match the session keys of the chain in tuple format. pub fn get_collator_keys_from_seed(seed: &str) -> AuraId { get_from_seed::<AuraId>(seed) } /// Helper function to generate an account ID from seed pub fn get_account_id_from_seed<TPublic: Public>(seed: &str) -> AccountId where AccountPublic: From<<TPublic::Pair as Pair>::Public>, { AccountPublic::from(get_from_seed::<TPublic>(seed)).into_account() } /// Generate the session keys from individual elements. /// /// The input must be a tuple of individual keys (a single arg for now since we have just one key). pub fn template_session_keys(keys: AuraId) -> contracts_parachain_runtime::SessionKeys { contracts_parachain_runtime::SessionKeys { aura: keys } } pub fn local_testnet_config() -> ChainSpec { // Give your base currency a unit name and decimal places let mut properties = sc_chain_spec::Properties::new(); properties.insert("tokenSymbol".into(), "UNIT".into()); properties.insert("tokenDecimals".into(), 12.into()); properties.insert("ss58Format".into(), 42.into()); #[allow(deprecated)] ChainSpec::builder( contracts_parachain_runtime::WASM_BINARY .expect("WASM binary was not built, please build it!"), Extensions { relay_chain: "rococo-local".into(), // You MUST set this to the correct network! para_id: 1000, }, ) .with_name("Local Testnet") .with_id("local_testnet") .with_chain_type(ChainType::Local) .with_genesis_config_patch(testnet_genesis( // initial collators. vec![ ( get_account_id_from_seed::<sr25519::Public>("Alice"), get_collator_keys_from_seed("Alice"), ), ( get_account_id_from_seed::<sr25519::Public>("Bob"), get_collator_keys_from_seed("Bob"), ), ], vec![ get_account_id_from_seed::<sr25519::Public>("Alice"), get_account_id_from_seed::<sr25519::Public>("Bob"), get_account_id_from_seed::<sr25519::Public>("Charlie"), get_account_id_from_seed::<sr25519::Public>("Dave"), get_account_id_from_seed::<sr25519::Public>("Eve"), get_account_id_from_seed::<sr25519::Public>("Ferdie"), get_account_id_from_seed::<sr25519::Public>("Alice//stash"), get_account_id_from_seed::<sr25519::Public>("Bob//stash"), get_account_id_from_seed::<sr25519::Public>("Charlie//stash"), get_account_id_from_seed::<sr25519::Public>("Dave//stash"), get_account_id_from_seed::<sr25519::Public>("Eve//stash"), get_account_id_from_seed::<sr25519::Public>("Ferdie//stash"), ], get_account_id_from_seed::<sr25519::Public>("Alice"), 1000.into(), )) .with_protocol_id("contracts-local") .with_properties(properties) .build() } fn testnet_genesis( invulnerables: Vec<(AccountId, AuraId)>, endowed_accounts: Vec<AccountId>, root: AccountId, id: ParaId, ) -> serde_json::Value { serde_json::json!({ "balances": { "balances": endowed_accounts.iter().cloned().map(|k| (k, 1u64 << 60)).collect::<Vec<_>>(), }, "parachainInfo": { "parachainId": id, }, "collatorSelection": { "invulnerables": invulnerables.iter().cloned().map(|(acc, _)| acc).collect::<Vec<_>>(), "candidacyBond": EXISTENTIAL_DEPOSIT * 16, }, "session": { "keys": invulnerables .into_iter() .map(|(acc, aura)| { ( acc.clone(), // account id acc, // validator id template_session_keys(aura), // session keys ) }) .collect::<Vec<_>>(), }, "polkadotXcm": { "safeXcmVersion": Some(SAFE_XCM_VERSION), }, "sudo": { "key": Some(root) } }) }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/node/src/cli.rs
node/src/cli.rs
use std::path::PathBuf; /// Sub-commands supported by the collator. #[derive(Debug, clap::Subcommand)] pub enum Subcommand { /// Build a chain specification. BuildSpec(sc_cli::BuildSpecCmd), /// Validate blocks. CheckBlock(sc_cli::CheckBlockCmd), /// Export blocks. ExportBlocks(sc_cli::ExportBlocksCmd), /// Export the state of a given block into a chain spec. ExportState(sc_cli::ExportStateCmd), /// Import blocks. ImportBlocks(sc_cli::ImportBlocksCmd), /// Revert the chain to a previous state. Revert(sc_cli::RevertCmd), /// Remove the whole chain. PurgeChain(cumulus_client_cli::PurgeChainCmd), /// Export the genesis state of the parachain. #[command(alias = "export-genesis-state")] ExportGenesisHead(cumulus_client_cli::ExportGenesisHeadCommand), /// Export the genesis wasm of the parachain. ExportGenesisWasm(cumulus_client_cli::ExportGenesisWasmCommand), /// Sub-commands concerned with benchmarking. /// The pallet benchmarking moved to the `pallet` sub-command. #[command(subcommand)] Benchmark(frame_benchmarking_cli::BenchmarkCmd), } const AFTER_HELP_EXAMPLE: &str = color_print::cstr!( r#"<bold><underline>Examples:</></> <bold>parachain-contracts-node build-spec --disable-default-bootnode > plain-parachain-chainspec.json</> Export a chainspec for a local testnet in json format. <bold>parachain-contracts-node --chain plain-parachain-chainspec.json --tmp -- --chain rococo-local</> Launch a full node with chain specification loaded from plain-parachain-chainspec.json. <bold>parachain-contracts-node</> Launch a full node with default parachain <italic>local-testnet</> and relay chain <italic>rococo-local</>. <bold>parachain-contracts-node --collator</> Launch a collator with default parachain <italic>local-testnet</> and relay chain <italic>rococo-local</>. "# ); #[derive(Debug, clap::Parser)] #[command( propagate_version = true, args_conflicts_with_subcommands = true, subcommand_negates_reqs = true )] #[clap(after_help = AFTER_HELP_EXAMPLE)] pub struct Cli { #[command(subcommand)] pub subcommand: Option<Subcommand>, #[command(flatten)] pub run: cumulus_client_cli::RunCmd, /// Disable automatic hardware benchmarks. /// /// By default these benchmarks are automatically ran at startup and measure /// the CPU speed, the memory bandwidth and the disk speed. /// /// The results are then printed out in the logs, and also sent as part of /// telemetry, if telemetry is enabled. #[arg(long)] pub no_hardware_benchmarks: bool, /// Relay chain arguments #[arg(raw = true)] pub relay_chain_args: Vec<String>, /// The number of seconds to delay before finalizing blocks. #[arg(long, default_value_t = 1)] pub finalize_delay_sec: u8, } #[derive(Debug)] pub struct RelayChainCli { /// The actual relay chain cli object. pub base: polkadot_cli::RunCmd, /// Optional chain id that should be passed to the relay chain. pub chain_id: Option<String>, /// The base path that should be used by the relay chain. pub base_path: Option<PathBuf>, } impl RelayChainCli { /// Parse the relay chain CLI parameters using the para chain `Configuration`. pub fn new<'a>( para_config: &sc_service::Configuration, relay_chain_args: impl Iterator<Item = &'a String>, ) -> Self { let extension = crate::chain_spec::Extensions::try_get(&*para_config.chain_spec); let chain_id = extension.map(|e| e.relay_chain.clone()); let base_path = para_config.base_path.path().join("polkadot"); Self { base_path: Some(base_path), chain_id, base: clap::Parser::parse_from(relay_chain_args), } } }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/node/src/service.rs
node/src/service.rs
//! Service and ServiceFactory implementation. Specialized wrapper over substrate service. pub mod dev; // std use std::{sync::Arc, time::Duration}; use cumulus_client_cli::CollatorOptions; // Local Runtime Types use contracts_parachain_runtime::{ opaque::{Block, Hash}, RuntimeApi, }; // Cumulus Imports use cumulus_client_collator::service::CollatorService; use cumulus_client_consensus_aura::collators::lookahead::{self as aura, Params as AuraParams}; use cumulus_client_consensus_common::ParachainBlockImport as TParachainBlockImport; use cumulus_client_consensus_proposer::Proposer; use cumulus_client_service::{ build_network, build_relay_chain_interface, prepare_node_config, start_relay_chain_tasks, BuildNetworkParams, CollatorSybilResistance, DARecoveryProfile, ParachainHostFunctions, StartRelayChainTasksParams, }; use cumulus_primitives_core::{ relay_chain::{CollatorPair, ValidationCode}, ParaId, }; use cumulus_relay_chain_interface::{OverseerHandle, RelayChainInterface}; // Substrate Imports use frame_benchmarking_cli::SUBSTRATE_REFERENCE_HARDWARE; use prometheus_endpoint::Registry; use sc_client_api::Backend; use sc_consensus::ImportQueue; use sc_executor::{HeapAllocStrategy, WasmExecutor, DEFAULT_HEAP_ALLOC_STRATEGY}; use sc_network::NetworkBlock; use sc_service::{Configuration, PartialComponents, TFullBackend, TFullClient, TaskManager}; use sc_telemetry::{Telemetry, TelemetryHandle, TelemetryWorker, TelemetryWorkerHandle}; use sc_transaction_pool_api::OffchainTransactionPoolFactory; use sp_keystore::KeystorePtr; type ParachainExecutor = WasmExecutor<ParachainHostFunctions>; type ParachainClient = TFullClient<Block, RuntimeApi, ParachainExecutor>; type ParachainBackend = TFullBackend<Block>; type ParachainBlockImport = TParachainBlockImport<Block, Arc<ParachainClient>, ParachainBackend>; /// Assembly of PartialComponents (enough to run chain ops subcommands) pub type Service = PartialComponents< ParachainClient, ParachainBackend, (), sc_consensus::DefaultImportQueue<Block>, sc_transaction_pool::FullPool<Block, ParachainClient>, (ParachainBlockImport, Option<Telemetry>, Option<TelemetryWorkerHandle>), >; /// Starts a `ServiceBuilder` for a full service. /// /// Use this macro if you don't actually need the full service, but just the builder in order to /// be able to perform chain operations. pub fn new_partial(config: &Configuration) -> Result<Service, sc_service::Error> { let telemetry = config .telemetry_endpoints .clone() .filter(|x| !x.is_empty()) .map(|endpoints| -> Result<_, sc_telemetry::Error> { let worker = TelemetryWorker::new(16)?; let telemetry = worker.handle().new_telemetry(endpoints); Ok((worker, telemetry)) }) .transpose()?; let heap_pages = config .executor .default_heap_pages .map_or(DEFAULT_HEAP_ALLOC_STRATEGY, |h| HeapAllocStrategy::Static { extra_pages: h as _ }); let executor = ParachainExecutor::builder() .with_execution_method(config.executor.wasm_method) .with_onchain_heap_alloc_strategy(heap_pages) .with_offchain_heap_alloc_strategy(heap_pages) .with_max_runtime_instances(config.executor.max_runtime_instances) .with_runtime_cache_size(config.executor.runtime_cache_size) .build(); let (client, backend, keystore_container, task_manager) = sc_service::new_full_parts_record_import::<Block, RuntimeApi, _>( config, telemetry.as_ref().map(|(_, telemetry)| telemetry.handle()), executor, true, )?; let client = Arc::new(client); let telemetry_worker_handle = telemetry.as_ref().map(|(worker, _)| worker.handle()); let telemetry = telemetry.map(|(worker, telemetry)| { task_manager.spawn_handle().spawn("telemetry", None, worker.run()); telemetry }); let transaction_pool = sc_transaction_pool::BasicPool::new_full( config.transaction_pool.clone(), config.role.is_authority().into(), config.prometheus_registry(), task_manager.spawn_essential_handle(), client.clone(), ); let block_import = ParachainBlockImport::new(client.clone(), backend.clone()); let import_queue = build_import_queue( client.clone(), block_import.clone(), config, telemetry.as_ref().map(|telemetry| telemetry.handle()), &task_manager, ); Ok(PartialComponents { backend, client, import_queue, keystore_container, task_manager, transaction_pool, select_chain: (), other: (block_import, telemetry, telemetry_worker_handle), }) } /// Build the import queue for the parachain runtime. fn build_import_queue( client: Arc<ParachainClient>, block_import: ParachainBlockImport, config: &Configuration, telemetry: Option<TelemetryHandle>, task_manager: &TaskManager, ) -> sc_consensus::DefaultImportQueue<Block> { cumulus_client_consensus_aura::equivocation_import_queue::fully_verifying_import_queue::< sp_consensus_aura::sr25519::AuthorityPair, _, _, _, _, >( client, block_import, move |_, _| async move { let timestamp = sp_timestamp::InherentDataProvider::from_system_time(); Ok(timestamp) }, &task_manager.spawn_essential_handle(), config.prometheus_registry(), telemetry, ) } #[allow(clippy::too_many_arguments)] fn start_consensus( client: Arc<ParachainClient>, backend: Arc<ParachainBackend>, block_import: ParachainBlockImport, prometheus_registry: Option<&Registry>, telemetry: Option<TelemetryHandle>, task_manager: &TaskManager, relay_chain_interface: Arc<dyn RelayChainInterface>, transaction_pool: Arc<sc_transaction_pool::FullPool<Block, ParachainClient>>, keystore: KeystorePtr, relay_chain_slot_duration: Duration, para_id: ParaId, collator_key: CollatorPair, overseer_handle: OverseerHandle, announce_block: Arc<dyn Fn(Hash, Option<Vec<u8>>) + Send + Sync>, ) -> Result<(), sc_service::Error> { let proposer_factory = sc_basic_authorship::ProposerFactory::with_proof_recording( task_manager.spawn_handle(), client.clone(), transaction_pool, prometheus_registry, telemetry.clone(), ); let proposer = Proposer::new(proposer_factory); let collator_service = CollatorService::new( client.clone(), Arc::new(task_manager.spawn_handle()), announce_block, client.clone(), ); let params = AuraParams { create_inherent_data_providers: move |_, ()| async move { Ok(()) }, block_import, para_client: client.clone(), para_backend: backend, relay_client: relay_chain_interface, code_hash_provider: move |block_hash| { client.code_at(block_hash).ok().map(|c| ValidationCode::from(c).hash()) }, keystore, collator_key, para_id, overseer_handle, relay_chain_slot_duration, proposer, collator_service, authoring_duration: Duration::from_millis(2000), reinitialize: false, }; let fut = aura::run::<Block, sp_consensus_aura::sr25519::AuthorityPair, _, _, _, _, _, _, _, _>( params, ); task_manager.spawn_essential_handle().spawn("aura", None, fut); Ok(()) } /// Start a node with the given parachain `Configuration` and relay chain `Configuration`. #[sc_tracing::logging::prefix_logs_with("Parachain")] pub async fn start_parachain_node( parachain_config: Configuration, polkadot_config: Configuration, collator_options: CollatorOptions, para_id: ParaId, hwbench: Option<sc_sysinfo::HwBench>, ) -> sc_service::error::Result<(TaskManager, Arc<ParachainClient>)> { let parachain_config = prepare_node_config(parachain_config); let params = new_partial(&parachain_config)?; let (block_import, mut telemetry, telemetry_worker_handle) = params.other; let prometheus_registry = parachain_config.prometheus_registry().cloned(); let net_config = sc_network::config::FullNetworkConfiguration::< _, _, sc_network::NetworkWorker<Block, Hash>, >::new(&parachain_config.network, prometheus_registry.clone()); let client = params.client.clone(); let backend = params.backend.clone(); let mut task_manager = params.task_manager; let (relay_chain_interface, collator_key) = build_relay_chain_interface( polkadot_config, &parachain_config, telemetry_worker_handle, &mut task_manager, collator_options.clone(), hwbench.clone(), ) .await .map_err(|e| sc_service::Error::Application(Box::new(e) as Box<_>))?; let validator = parachain_config.role.is_authority(); let transaction_pool = params.transaction_pool.clone(); let import_queue_service = params.import_queue.service(); // NOTE: because we use Aura here explicitly, we can use `CollatorSybilResistance::Resistant` // when starting the network. let (network, system_rpc_tx, tx_handler_controller, start_network, sync_service) = build_network(BuildNetworkParams { parachain_config: &parachain_config, net_config, client: client.clone(), transaction_pool: transaction_pool.clone(), para_id, spawn_handle: task_manager.spawn_handle(), relay_chain_interface: relay_chain_interface.clone(), import_queue: params.import_queue, sybil_resistance_level: CollatorSybilResistance::Resistant, // because of Aura }) .await?; if parachain_config.offchain_worker.enabled { use futures::FutureExt; task_manager.spawn_handle().spawn( "offchain-workers-runner", "offchain-work", sc_offchain::OffchainWorkers::new(sc_offchain::OffchainWorkerOptions { runtime_api_provider: client.clone(), keystore: Some(params.keystore_container.keystore()), offchain_db: backend.offchain_storage(), transaction_pool: Some(OffchainTransactionPoolFactory::new( transaction_pool.clone(), )), network_provider: Arc::new(network.clone()), is_validator: parachain_config.role.is_authority(), enable_http_requests: false, custom_extensions: move |_| vec![], }) .run(client.clone(), task_manager.spawn_handle()) .boxed(), ); } let rpc_builder = { let client = client.clone(); let transaction_pool = transaction_pool.clone(); Box::new(move |_| { let deps = crate::rpc::FullDeps { client: client.clone(), pool: transaction_pool.clone() }; crate::rpc::create_full(deps).map_err(Into::into) }) }; sc_service::spawn_tasks(sc_service::SpawnTasksParams { rpc_builder, client: client.clone(), transaction_pool: transaction_pool.clone(), task_manager: &mut task_manager, config: parachain_config, keystore: params.keystore_container.keystore(), backend: backend.clone(), network, sync_service: sync_service.clone(), system_rpc_tx, tx_handler_controller, telemetry: telemetry.as_mut(), })?; if let Some(hwbench) = hwbench { sc_sysinfo::print_hwbench(&hwbench); // Here you can check whether the hardware meets your chains' requirements. Putting a link // in there and swapping out the requirements for your own are probably a good idea. The // requirements for a para-chain are dictated by its relay-chain. match SUBSTRATE_REFERENCE_HARDWARE.check_hardware(&hwbench, false) { Err(err) if validator => { log::warn!( "⚠️ The hardware does not meet the minimal requirements {} for role 'Authority'.", err ); }, _ => {}, } if let Some(ref mut telemetry) = telemetry { let telemetry_handle = telemetry.handle(); task_manager.spawn_handle().spawn( "telemetry_hwbench", None, sc_sysinfo::initialize_hwbench_telemetry(telemetry_handle, hwbench), ); } } let announce_block = { let sync_service = sync_service.clone(); Arc::new(move |hash, data| sync_service.announce_block(hash, data)) }; let relay_chain_slot_duration = Duration::from_secs(6); let overseer_handle = relay_chain_interface .overseer_handle() .map_err(|e| sc_service::Error::Application(Box::new(e)))?; start_relay_chain_tasks(StartRelayChainTasksParams { client: client.clone(), announce_block: announce_block.clone(), para_id, relay_chain_interface: relay_chain_interface.clone(), task_manager: &mut task_manager, da_recovery_profile: if validator { DARecoveryProfile::Collator } else { DARecoveryProfile::FullNode }, import_queue: import_queue_service, relay_chain_slot_duration, recovery_handle: Box::new(overseer_handle.clone()), sync_service: sync_service.clone(), })?; if validator { start_consensus( client.clone(), backend, block_import, prometheus_registry.as_ref(), telemetry.as_ref().map(|t| t.handle()), &task_manager, relay_chain_interface, transaction_pool, params.keystore_container.keystore(), relay_chain_slot_duration, para_id, collator_key.expect("Command line arguments do not allow this. qed"), overseer_handle, announce_block, )?; } start_network.start_network(); Ok((task_manager, client)) }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/node/src/rpc.rs
node/src/rpc.rs
//! A collection of node-specific RPC methods. //! Substrate provides the `sc-rpc` crate, which defines the core RPC layer //! used by Substrate nodes. This file extends those RPC definitions with //! capabilities that are specific to this project's runtime configuration. #![warn(missing_docs)] use std::sync::Arc; use contracts_parachain_runtime::{opaque::Block, AccountId, Balance, Nonce}; use sc_transaction_pool_api::TransactionPool; use sp_api::ProvideRuntimeApi; use sp_block_builder::BlockBuilder; use sp_blockchain::{Error as BlockChainError, HeaderBackend, HeaderMetadata}; /// A type representing all RPC extensions. pub type RpcExtension = jsonrpsee::RpcModule<()>; /// Full client dependencies pub struct FullDeps<C, P> { /// The client instance to use. pub client: Arc<C>, /// Transaction pool instance. pub pool: Arc<P>, } /// Instantiate all RPC extensions. pub fn create_full<C, P>( deps: FullDeps<C, P>, ) -> Result<RpcExtension, Box<dyn std::error::Error + Send + Sync>> where C: ProvideRuntimeApi<Block> + HeaderBackend<Block> + HeaderMetadata<Block, Error = BlockChainError> + Send + Sync + 'static, C::Api: pallet_transaction_payment_rpc::TransactionPaymentRuntimeApi<Block, Balance>, C::Api: substrate_frame_rpc_system::AccountNonceApi<Block, AccountId, Nonce>, C::Api: BlockBuilder<Block>, P: TransactionPool + Sync + Send + 'static, { use pallet_transaction_payment_rpc::{TransactionPayment, TransactionPaymentApiServer}; use substrate_frame_rpc_system::{System, SystemApiServer}; let mut module = RpcExtension::new(()); let FullDeps { client, pool } = deps; module.merge(System::new(client.clone(), pool).into_rpc())?; module.merge(TransactionPayment::new(client).into_rpc())?; Ok(module) }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/node/src/main.rs
node/src/main.rs
//! Substrate Parachain Node Template CLI #![warn(missing_docs)] mod chain_spec; #[macro_use] mod service; mod cli; mod command; mod rpc; fn main() -> sc_cli::Result<()> { command::run() }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/node/src/service/dev.rs
node/src/service/dev.rs
//! Service and ServiceFactory implementation. Specialized wrapper over substrate service. use contracts_node_runtime::{self, opaque::Block, RuntimeApi}; use futures::FutureExt; use sc_client_api::Backend; use sc_service::{error::Error as ServiceError, Configuration, TaskManager}; use sc_telemetry::{Telemetry, TelemetryWorker}; use sc_transaction_pool_api::OffchainTransactionPoolFactory; use std::sync::Arc; pub(crate) type FullClient = sc_service::TFullClient< Block, RuntimeApi, sc_executor::WasmExecutor<sp_io::SubstrateHostFunctions>, >; type FullBackend = sc_service::TFullBackend<Block>; type FullSelectChain = sc_consensus::LongestChain<FullBackend, Block>; pub fn new_partial( config: &Configuration, ) -> Result< sc_service::PartialComponents< FullClient, FullBackend, FullSelectChain, sc_consensus::DefaultImportQueue<Block>, sc_transaction_pool::FullPool<Block, FullClient>, Option<Telemetry>, >, ServiceError, > { let telemetry = config .telemetry_endpoints .clone() .filter(|x| !x.is_empty()) .map(|endpoints| -> Result<_, sc_telemetry::Error> { let worker = TelemetryWorker::new(16)?; let telemetry = worker.handle().new_telemetry(endpoints); Ok((worker, telemetry)) }) .transpose()?; let executor = sc_service::new_wasm_executor(&config.executor); let (client, backend, keystore_container, task_manager) = sc_service::new_full_parts::<Block, RuntimeApi, _>( config, telemetry.as_ref().map(|(_, telemetry)| telemetry.handle()), executor, )?; let client = Arc::new(client); let select_chain = sc_consensus::LongestChain::new(backend.clone()); let telemetry = telemetry.map(|(worker, telemetry)| { task_manager.spawn_handle().spawn("telemetry", None, worker.run()); telemetry }); let import_queue = sc_consensus_manual_seal::import_queue( Box::new(client.clone()), &task_manager.spawn_essential_handle(), config.prometheus_registry(), ); let transaction_pool = sc_transaction_pool::BasicPool::new_full( config.transaction_pool.clone(), config.role.is_authority().into(), config.prometheus_registry(), task_manager.spawn_essential_handle(), client.clone(), ); Ok(sc_service::PartialComponents { client, backend, task_manager, import_queue, keystore_container, select_chain, transaction_pool, other: (telemetry), }) } pub fn new_full< N: sc_network::NetworkBackend<Block, <Block as sp_runtime::traits::Block>::Hash>, >( config: Configuration, finalize_delay_sec: u64, ) -> Result<TaskManager, ServiceError> { let sc_service::PartialComponents { client, backend, mut task_manager, import_queue, keystore_container, select_chain, transaction_pool, other: mut telemetry, } = new_partial(&config)?; let net_config = sc_network::config::FullNetworkConfiguration::< Block, <Block as sp_runtime::traits::Block>::Hash, N, >::new(&config.network, config.prometheus_config.as_ref().map(|cfg| cfg.registry.clone())); let metrics = N::register_notification_metrics(config.prometheus_registry()); let (network, system_rpc_tx, tx_handler_controller, network_starter, sync_service) = sc_service::build_network(sc_service::BuildNetworkParams { config: &config, client: client.clone(), transaction_pool: transaction_pool.clone(), spawn_handle: task_manager.spawn_handle(), import_queue, net_config, block_announce_validator_builder: None, warp_sync_config: None, block_relay: None, metrics, })?; if config.offchain_worker.enabled { task_manager.spawn_handle().spawn( "offchain-workers-runner", "offchain-worker", sc_offchain::OffchainWorkers::new(sc_offchain::OffchainWorkerOptions { runtime_api_provider: client.clone(), is_validator: config.role.is_authority(), keystore: Some(keystore_container.keystore()), offchain_db: backend.offchain_storage(), transaction_pool: Some(OffchainTransactionPoolFactory::new( transaction_pool.clone(), )), network_provider: Arc::new(network.clone()), enable_http_requests: true, custom_extensions: |_| vec![], }) .run(client.clone(), task_manager.spawn_handle()) .boxed(), ); } let prometheus_registry = config.prometheus_registry().cloned(); let rpc_extensions_builder = { let client = client.clone(); let pool = transaction_pool.clone(); Box::new(move |_| { let deps = crate::rpc::FullDeps { client: client.clone(), pool: pool.clone() }; crate::rpc::create_full(deps).map_err(Into::into) }) }; let _rpc_handlers = sc_service::spawn_tasks(sc_service::SpawnTasksParams { network, client: client.clone(), keystore: keystore_container.keystore(), task_manager: &mut task_manager, transaction_pool: transaction_pool.clone(), rpc_builder: rpc_extensions_builder, backend, system_rpc_tx, tx_handler_controller, sync_service, config, telemetry: telemetry.as_mut(), })?; let proposer = sc_basic_authorship::ProposerFactory::new( task_manager.spawn_handle(), client.clone(), transaction_pool.clone(), prometheus_registry.as_ref(), telemetry.as_ref().map(|x| x.handle()), ); let params = sc_consensus_manual_seal::InstantSealParams { block_import: client.clone(), env: proposer, client: client.clone(), pool: transaction_pool, select_chain, consensus_data_provider: None, create_inherent_data_providers: move |_, ()| async move { Ok(sp_timestamp::InherentDataProvider::from_system_time()) }, }; let authorship_future = sc_consensus_manual_seal::run_instant_seal(params); task_manager .spawn_essential_handle() .spawn_blocking("instant-seal", None, authorship_future); let delayed_finalize_params = sc_consensus_manual_seal::DelayedFinalizeParams { client, spawn_handle: task_manager.spawn_handle(), delay_sec: finalize_delay_sec, }; task_manager.spawn_essential_handle().spawn_blocking( "delayed_finalize", None, sc_consensus_manual_seal::run_delayed_finalize(delayed_finalize_params), ); network_starter.start_network(); Ok(task_manager) }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/node/src/chain_spec/dev.rs
node/src/chain_spec/dev.rs
use contracts_node_runtime::{AccountId, Signature, WASM_BINARY}; use sc_service::ChainType; use sp_core::{sr25519, Pair, Public}; use sp_runtime::traits::{IdentifyAccount, Verify}; /// Specialized `ChainSpec`. This is a specialization of the general Substrate ChainSpec type. pub type ChainSpec = sc_service::GenericChainSpec; /// Generate a crypto pair from seed. pub fn get_from_seed<TPublic: Public>(seed: &str) -> <TPublic::Pair as Pair>::Public { TPublic::Pair::from_string(&format!("//{}", seed), None) .expect("static values are valid; qed") .public() } type AccountPublic = <Signature as Verify>::Signer; /// Generate an account ID from seed. pub fn get_account_id_from_seed<TPublic: Public>(seed: &str) -> AccountId where AccountPublic: From<<TPublic::Pair as Pair>::Public>, { AccountPublic::from(get_from_seed::<TPublic>(seed)).into_account() } pub fn development_config() -> Result<ChainSpec, String> { Ok(ChainSpec::builder(WASM_BINARY.expect("Development wasm not available"), Default::default()) .with_name("Development") .with_id("dev") .with_chain_type(ChainType::Development) .with_genesis_config_patch(testnet_genesis( // Sudo account get_account_id_from_seed::<sr25519::Public>("Alice"), // Pre-funded accounts vec![ get_account_id_from_seed::<sr25519::Public>("Alice"), get_account_id_from_seed::<sr25519::Public>("Alice//stash"), get_account_id_from_seed::<sr25519::Public>("Bob"), get_account_id_from_seed::<sr25519::Public>("Bob//stash"), get_account_id_from_seed::<sr25519::Public>("Charlie"), get_account_id_from_seed::<sr25519::Public>("Charlie//stash"), get_account_id_from_seed::<sr25519::Public>("Dave"), get_account_id_from_seed::<sr25519::Public>("Dave//stash"), get_account_id_from_seed::<sr25519::Public>("Eve"), get_account_id_from_seed::<sr25519::Public>("Eve//stash"), get_account_id_from_seed::<sr25519::Public>("Ferdie"), get_account_id_from_seed::<sr25519::Public>("Ferdie//stash"), ], true, )) .build()) } /// Configure initial storage state for FRAME modules. fn testnet_genesis( root_key: AccountId, endowed_accounts: Vec<AccountId>, _enable_println: bool, ) -> serde_json::Value { serde_json::json!({ "balances": { // Configure endowed accounts with initial balance of 1 << 60. "balances": endowed_accounts.iter().cloned().map(|k| (k, 1u64 << 60)).collect::<Vec<_>>(), }, "sudo": { // Assign network admin rights. "key": Some(root_key), }, }) }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/runtime/build.rs
runtime/build.rs
fn main() { #[cfg(feature = "std")] { substrate_wasm_builder::WasmBuilder::new() .with_current_project() .export_heap_base() .import_memory() .build(); } }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/runtime/src/lib.rs
runtime/src/lib.rs
#![cfg_attr(not(feature = "std"), no_std)] // `construct_runtime!` does a lot of recursion and requires us to increase the limit to 256. #![recursion_limit = "256"] // Make the WASM binary available. #[cfg(feature = "std")] include!(concat!(env!("OUT_DIR"), "/wasm_binary.rs")); mod assets_config; mod contracts_config; mod revive_config; extern crate alloc; use alloc::vec::Vec; use frame_support::{ derive_impl, dispatch::DispatchClass, genesis_builder_helper::{build_state, get_preset}, }; use frame_system::limits::{BlockLength, BlockWeights}; use polkadot_runtime_common::SlowAdjustingFeeUpdate; use sp_api::impl_runtime_apis; use sp_core::{crypto::KeyTypeId, OpaqueMetadata}; use sp_runtime::{ create_runtime_str, generic, impl_opaque_keys, traits::{BlakeTwo256, Block as BlockT, IdentifyAccount, Verify}, transaction_validity::{TransactionSource, TransactionValidity}, ApplyExtrinsicResult, MultiSignature, }; #[cfg(feature = "std")] use sp_version::NativeVersion; use sp_version::RuntimeVersion; // A few exports that help ease life for downstream crates. pub use frame_support::{ construct_runtime, parameter_types, traits::{ AsEnsureOriginWithArg, ConstBool, ConstU128, ConstU32, ConstU8, KeyOwnerProofSystem, Randomness, StorageInfo, }, weights::{ constants::{ BlockExecutionWeight, ExtrinsicBaseWeight, RocksDbWeight, WEIGHT_REF_TIME_PER_SECOND, }, IdentityFee, Weight, }, StorageValue, }; pub use frame_system::Call as SystemCall; pub use pallet_balances::Call as BalancesCall; pub use pallet_timestamp::Call as TimestampCall; use pallet_transaction_payment::FungibleAdapter; #[cfg(any(feature = "std", test))] pub use sp_runtime::BuildStorage; pub use sp_runtime::{Perbill, Permill}; /// An index to a block. pub type BlockNumber = u32; /// Alias to 512-bit hash when used in the context of a transaction signature on the chain. pub type Signature = MultiSignature; /// Some way of identifying an account on the chain. We intentionally make it equivalent /// to the public key of our transaction signing scheme. pub type AccountId = <<Signature as Verify>::Signer as IdentifyAccount>::AccountId; /// Balance of an account. pub type Balance = u128; /// Index of a transaction in the chain. pub type Nonce = u32; /// A hash of some data used by the chain. pub type Hash = sp_core::H256; /// Opaque types. These are used by the CLI to instantiate machinery that don't need to know /// the specifics of the runtime. They can then be made to be agnostic over specific formats /// of data like extrinsics, allowing for them to continue syncing the network through upgrades /// to even the core data structures. pub mod opaque { use super::*; pub use sp_runtime::OpaqueExtrinsic as UncheckedExtrinsic; /// Opaque block header type. pub type Header = generic::Header<BlockNumber, BlakeTwo256>; /// Opaque block type. pub type Block = generic::Block<Header, UncheckedExtrinsic>; /// Opaque block identifier type. pub type BlockId = generic::BlockId<Block>; impl_opaque_keys! { pub struct SessionKeys {} } } // To learn more about runtime versioning, see: // https://docs.substrate.io/main-docs/build/upgrade#runtime-versioning #[sp_version::runtime_version] pub const VERSION: RuntimeVersion = RuntimeVersion { spec_name: create_runtime_str!("substrate-contracts-node"), impl_name: create_runtime_str!("substrate-contracts-node"), authoring_version: 1, // The version of the runtime specification. A full node will not attempt to use its native // runtime in substitute for the on-chain Wasm runtime unless all of `spec_name`, // `spec_version`, and `authoring_version` are the same between Wasm and native. // This value is set to 100 to notify Polkadot-JS App (https://polkadot.js.org/apps) to use // the compatible custom types. spec_version: 100, impl_version: 1, apis: RUNTIME_API_VERSIONS, transaction_version: 1, state_version: 1, }; /// The version information used to identify this runtime when compiled natively. #[cfg(feature = "std")] pub fn native_version() -> NativeVersion { NativeVersion { runtime_version: VERSION, can_author_with: Default::default() } } const NORMAL_DISPATCH_RATIO: Perbill = Perbill::from_percent(75); /// We assume that ~10% of the block weight is consumed by `on_initialize` handlers. /// This is used to limit the maximal weight of a single extrinsic. const AVERAGE_ON_INITIALIZE_RATIO: Perbill = Perbill::from_percent(10); /// We allow for 2 seconds of compute with a 6 second average block time, with maximum proof size. const MAXIMUM_BLOCK_WEIGHT: Weight = Weight::from_parts(WEIGHT_REF_TIME_PER_SECOND.saturating_mul(2), u64::MAX); // Prints debug output of the `revive` pallet to stdout if the node is // started with `-lruntime::revive=trace` or `-lruntime::contracts=debug`. const CONTRACTS_DEBUG_OUTPUT: pallet_contracts::DebugInfo = pallet_contracts::DebugInfo::UnsafeDebug; const CONTRACTS_EVENTS: pallet_contracts::CollectEvents = pallet_contracts::CollectEvents::UnsafeCollect; const REVIVE_DEBUG_OUTPUT: pallet_revive::DebugInfo = pallet_revive::DebugInfo::UnsafeDebug; const REVIVE_EVENTS: pallet_revive::CollectEvents = pallet_revive::CollectEvents::UnsafeCollect; // Unit = the base number of indivisible units for balances const MILLIUNIT: Balance = 1_000_000_000; pub const EXISTENTIAL_DEPOSIT: Balance = MILLIUNIT; impl pallet_insecure_randomness_collective_flip::Config for Runtime {} parameter_types! { pub const BlockHashCount: BlockNumber = 2400; pub const Version: RuntimeVersion = VERSION; // This part is copied from Substrate's `bin/node/runtime/src/lib.rs`. // The `RuntimeBlockLength` and `RuntimeBlockWeights` exist here because the // `DeletionWeightLimit` and `DeletionQueueDepth` depend on those to parameterize // the lazy contract deletion. pub RuntimeBlockLength: BlockLength = BlockLength::max_with_normal_ratio(5 * 1024 * 1024, NORMAL_DISPATCH_RATIO); pub RuntimeBlockWeights: BlockWeights = BlockWeights::builder() .base_block(BlockExecutionWeight::get()) .for_class(DispatchClass::all(), |weights| { weights.base_extrinsic = ExtrinsicBaseWeight::get(); }) .for_class(DispatchClass::Normal, |weights| { weights.max_total = Some(NORMAL_DISPATCH_RATIO * MAXIMUM_BLOCK_WEIGHT); }) .for_class(DispatchClass::Operational, |weights| { weights.max_total = Some(MAXIMUM_BLOCK_WEIGHT); // Operational transactions have some extra reserved space, so that they // are included even if block reached `MAXIMUM_BLOCK_WEIGHT`. weights.reserved = Some( MAXIMUM_BLOCK_WEIGHT - NORMAL_DISPATCH_RATIO * MAXIMUM_BLOCK_WEIGHT ); }) .avg_block_initialization(AVERAGE_ON_INITIALIZE_RATIO) .build_or_panic(); pub const SS58Prefix: u8 = 42; } // Configure FRAME pallets to include in runtime. #[derive_impl(frame_system::config_preludes::SolochainDefaultConfig)] impl frame_system::Config for Runtime { /// The block type for the runtime. type Block = Block; /// Block & extrinsics weights: base values and limits. type BlockWeights = RuntimeBlockWeights; /// The maximum length of a block (in bytes). type BlockLength = RuntimeBlockLength; /// The identifier used to distinguish between accounts. type AccountId = AccountId; /// The type for storing how many extrinsics an account has signed. type Nonce = Nonce; /// The type for hashing blocks and tries. type Hash = Hash; /// Maximum number of block number to block hash mappings to keep (oldest pruned first). type BlockHashCount = BlockHashCount; /// The weight of database operations that the runtime can invoke. type DbWeight = RocksDbWeight; /// Version of the runtime. type Version = Version; /// The data to be stored in an account. type AccountData = pallet_balances::AccountData<Balance>; /// This is used as an identifier of the chain. 42 is the generic substrate prefix. type SS58Prefix = SS58Prefix; type MaxConsumers = frame_support::traits::ConstU32<16>; } parameter_types! { pub const UncleGenerations: u32 = 0; } impl pallet_authorship::Config for Runtime { type FindAuthor = (); type EventHandler = (); } parameter_types! { pub const MinimumPeriod: u64 = 5; } impl pallet_timestamp::Config for Runtime { /// A timestamp: milliseconds since the unix epoch. type Moment = u64; type OnTimestampSet = (); type MinimumPeriod = MinimumPeriod; type WeightInfo = (); } parameter_types! { pub const MaxLocks: u32 = 50; pub const MaxReserves: u32 = 50; } impl pallet_balances::Config for Runtime { type MaxLocks = MaxLocks; type MaxReserves = MaxReserves; type ReserveIdentifier = [u8; 8]; /// The type for recording an account's balance. type Balance = Balance; /// The ubiquitous event type. type RuntimeEvent = RuntimeEvent; type DustRemoval = (); type ExistentialDeposit = ConstU128<EXISTENTIAL_DEPOSIT>; type AccountStore = System; type WeightInfo = pallet_balances::weights::SubstrateWeight<Runtime>; type FreezeIdentifier = (); type MaxFreezes = (); type RuntimeHoldReason = RuntimeHoldReason; type RuntimeFreezeReason = RuntimeFreezeReason; } impl pallet_transaction_payment::Config for Runtime { type RuntimeEvent = RuntimeEvent; type OnChargeTransaction = FungibleAdapter<Balances, ()>; type OperationalFeeMultiplier = ConstU8<5>; type WeightToFee = IdentityFee<Balance>; type LengthToFee = IdentityFee<Balance>; type FeeMultiplierUpdate = SlowAdjustingFeeUpdate<Self>; } impl pallet_sudo::Config for Runtime { type RuntimeEvent = RuntimeEvent; type RuntimeCall = RuntimeCall; type WeightInfo = pallet_sudo::weights::SubstrateWeight<Runtime>; } impl pallet_utility::Config for Runtime { type RuntimeEvent = RuntimeEvent; type RuntimeCall = RuntimeCall; type PalletsOrigin = OriginCaller; type WeightInfo = pallet_utility::weights::SubstrateWeight<Runtime>; } // Create the runtime by composing the FRAME pallets that were previously configured. #[frame_support::runtime] mod runtime { #[runtime::runtime] #[runtime::derive( RuntimeCall, RuntimeEvent, RuntimeError, RuntimeOrigin, RuntimeFreezeReason, RuntimeHoldReason, RuntimeSlashReason, RuntimeLockId, RuntimeTask )] pub struct Runtime; #[runtime::pallet_index(0)] pub type System = frame_system; #[runtime::pallet_index(1)] pub type RandomnessCollectiveFlip = pallet_insecure_randomness_collective_flip; #[runtime::pallet_index(2)] pub type Utility = pallet_utility; #[runtime::pallet_index(3)] pub type Timestamp = pallet_timestamp; #[runtime::pallet_index(4)] pub type Balances = pallet_balances; #[runtime::pallet_index(5)] pub type Authorship = pallet_authorship; #[runtime::pallet_index(6)] pub type TransactionPayment = pallet_transaction_payment; #[runtime::pallet_index(7)] pub type Sudo = pallet_sudo; #[runtime::pallet_index(8)] pub type Contracts = pallet_contracts; #[runtime::pallet_index(9)] pub type Revive = pallet_revive; #[runtime::pallet_index(10)] pub type Assets = pallet_assets; } /// The address format for describing accounts. pub type Address = sp_runtime::MultiAddress<AccountId, ()>; /// Block header type as expected by this runtime. pub type Header = generic::Header<BlockNumber, BlakeTwo256>; /// Block type as expected by this runtime. pub type Block = generic::Block<Header, UncheckedExtrinsic>; /// The SignedExtension to the basic transaction logic. pub type SignedExtra = ( frame_system::CheckNonZeroSender<Runtime>, frame_system::CheckSpecVersion<Runtime>, frame_system::CheckTxVersion<Runtime>, frame_system::CheckGenesis<Runtime>, frame_system::CheckEra<Runtime>, frame_system::CheckNonce<Runtime>, frame_system::CheckWeight<Runtime>, pallet_transaction_payment::ChargeTransactionPayment<Runtime>, ); /// The payload being signed in transactions. pub type SignedPayload = generic::SignedPayload<RuntimeCall, SignedExtra>; /// Unchecked extrinsic type as expected by this runtime. pub type UncheckedExtrinsic = generic::UncheckedExtrinsic<Address, RuntimeCall, Signature, SignedExtra>; /// Executive: handles dispatch to the various modules. pub type Executive = frame_executive::Executive< Runtime, Block, frame_system::ChainContext<Runtime>, Runtime, AllPalletsWithSystem, >; type EventRecord = frame_system::EventRecord< <Runtime as frame_system::Config>::RuntimeEvent, <Runtime as frame_system::Config>::Hash, >; impl_runtime_apis! { impl sp_api::Core<Block> for Runtime { fn version() -> RuntimeVersion { VERSION } fn execute_block(block: Block) { Executive::execute_block(block); } fn initialize_block(header: &<Block as BlockT>::Header) -> sp_runtime::ExtrinsicInclusionMode { Executive::initialize_block(header) } } impl sp_api::Metadata<Block> for Runtime { fn metadata() -> OpaqueMetadata { OpaqueMetadata::new(Runtime::metadata().into()) } fn metadata_at_version(version: u32) -> Option<OpaqueMetadata> { Runtime::metadata_at_version(version) } fn metadata_versions() -> Vec<u32> { Runtime::metadata_versions() } } impl sp_block_builder::BlockBuilder<Block> for Runtime { fn apply_extrinsic(extrinsic: <Block as BlockT>::Extrinsic) -> ApplyExtrinsicResult { Executive::apply_extrinsic(extrinsic) } fn finalize_block() -> <Block as BlockT>::Header { Executive::finalize_block() } fn inherent_extrinsics(data: sp_inherents::InherentData) -> Vec<<Block as BlockT>::Extrinsic> { data.create_extrinsics() } fn check_inherents( block: Block, data: sp_inherents::InherentData, ) -> sp_inherents::CheckInherentsResult { data.check_extrinsics(&block) } } impl sp_transaction_pool::runtime_api::TaggedTransactionQueue<Block> for Runtime { fn validate_transaction( source: TransactionSource, tx: <Block as BlockT>::Extrinsic, block_hash: <Block as BlockT>::Hash, ) -> TransactionValidity { Executive::validate_transaction(source, tx, block_hash) } } impl sp_offchain::OffchainWorkerApi<Block> for Runtime { fn offchain_worker(header: &<Block as BlockT>::Header) { Executive::offchain_worker(header) } } impl sp_session::SessionKeys<Block> for Runtime { fn generate_session_keys(seed: Option<Vec<u8>>) -> Vec<u8> { opaque::SessionKeys::generate(seed) } fn decode_session_keys( encoded: Vec<u8>, ) -> Option<Vec<(Vec<u8>, KeyTypeId)>> { opaque::SessionKeys::decode_into_raw_public_keys(&encoded) } } impl frame_system_rpc_runtime_api::AccountNonceApi<Block, AccountId, Nonce> for Runtime { fn account_nonce(account: AccountId) -> Nonce { System::account_nonce(account) } } impl pallet_transaction_payment_rpc_runtime_api::TransactionPaymentApi<Block, Balance> for Runtime { fn query_info( uxt: <Block as BlockT>::Extrinsic, len: u32, ) -> pallet_transaction_payment_rpc_runtime_api::RuntimeDispatchInfo<Balance> { TransactionPayment::query_info(uxt, len) } fn query_fee_details( uxt: <Block as BlockT>::Extrinsic, len: u32, ) -> pallet_transaction_payment::FeeDetails<Balance> { TransactionPayment::query_fee_details(uxt, len) } fn query_weight_to_fee(weight: Weight) -> Balance { TransactionPayment::weight_to_fee(weight) } fn query_length_to_fee(length: u32) -> Balance { TransactionPayment::length_to_fee(length) } } impl pallet_transaction_payment_rpc_runtime_api::TransactionPaymentCallApi<Block, Balance, RuntimeCall> for Runtime { fn query_call_info( call: RuntimeCall, len: u32, ) -> pallet_transaction_payment::RuntimeDispatchInfo<Balance> { TransactionPayment::query_call_info(call, len) } fn query_call_fee_details( call: RuntimeCall, len: u32, ) -> pallet_transaction_payment::FeeDetails<Balance> { TransactionPayment::query_call_fee_details(call, len) } fn query_weight_to_fee(weight: Weight) -> Balance { TransactionPayment::weight_to_fee(weight) } fn query_length_to_fee(length: u32) -> Balance { TransactionPayment::length_to_fee(length) } } impl pallet_contracts::ContractsApi<Block, AccountId, Balance, BlockNumber, Hash, EventRecord> for Runtime { fn call( origin: AccountId, dest: AccountId, value: Balance, gas_limit: Option<Weight>, storage_deposit_limit: Option<Balance>, input_data: Vec<u8>, ) -> pallet_contracts::ContractExecResult<Balance, EventRecord> { let gas_limit = gas_limit.unwrap_or(RuntimeBlockWeights::get().max_block); Contracts::bare_call( origin, dest, value, gas_limit, storage_deposit_limit, input_data, CONTRACTS_DEBUG_OUTPUT, CONTRACTS_EVENTS, pallet_contracts::Determinism::Enforced, ) } fn instantiate( origin: AccountId, value: Balance, gas_limit: Option<Weight>, storage_deposit_limit: Option<Balance>, code: pallet_contracts::Code<Hash>, data: Vec<u8>, salt: Vec<u8>, ) -> pallet_contracts::ContractInstantiateResult<AccountId, Balance, EventRecord> { let gas_limit = gas_limit.unwrap_or(RuntimeBlockWeights::get().max_block); Contracts::bare_instantiate( origin, value, gas_limit, storage_deposit_limit, code, data, salt, CONTRACTS_DEBUG_OUTPUT, CONTRACTS_EVENTS, ) } fn upload_code( origin: AccountId, code: Vec<u8>, storage_deposit_limit: Option<Balance>, determinism: pallet_contracts::Determinism, ) -> pallet_contracts::CodeUploadResult<Hash, Balance> { Contracts::bare_upload_code(origin, code, storage_deposit_limit, determinism) } fn get_storage( address: AccountId, key: Vec<u8>, ) -> pallet_contracts::GetStorageResult { Contracts::get_storage(address, key) } } impl pallet_revive::ReviveApi<Block, AccountId, Balance, BlockNumber, Hash, EventRecord> for Runtime { fn call( origin: AccountId, dest: AccountId, value: Balance, gas_limit: Option<Weight>, storage_deposit_limit: Option<Balance>, input_data: Vec<u8>, ) -> pallet_revive::ContractExecResult<Balance, EventRecord> { Revive::bare_call( RuntimeOrigin::signed(origin), dest, value, gas_limit.unwrap_or(RuntimeBlockWeights::get().max_block), storage_deposit_limit.unwrap_or(u128::MAX), input_data, REVIVE_DEBUG_OUTPUT, REVIVE_EVENTS, ) } fn instantiate( origin: AccountId, value: Balance, gas_limit: Option<Weight>, storage_deposit_limit: Option<Balance>, code: pallet_revive::Code<Hash>, data: Vec<u8>, salt: Vec<u8>, ) -> pallet_revive::ContractInstantiateResult<AccountId, Balance, EventRecord> { Revive::bare_instantiate( RuntimeOrigin::signed(origin), value, gas_limit.unwrap_or(RuntimeBlockWeights::get().max_block), storage_deposit_limit.unwrap_or(u128::MAX), code, data, salt, REVIVE_DEBUG_OUTPUT, REVIVE_EVENTS, ) } fn upload_code( origin: AccountId, code: Vec<u8>, storage_deposit_limit: Option<Balance>, ) -> pallet_revive::CodeUploadResult<Hash, Balance> { Revive::bare_upload_code( RuntimeOrigin::signed(origin), code, storage_deposit_limit.unwrap_or(u128::MAX), ) } fn get_storage( address: AccountId, key: Vec<u8>, ) -> pallet_revive::GetStorageResult { Revive::get_storage( address, key ) } } impl sp_genesis_builder::GenesisBuilder<Block> for Runtime { fn build_state(config: Vec<u8>) -> sp_genesis_builder::Result { build_state::<RuntimeGenesisConfig>(config) } fn get_preset(id: &Option<sp_genesis_builder::PresetId>) -> Option<Vec<u8>> { get_preset::<RuntimeGenesisConfig>(id, |_| None) } fn preset_names() -> Vec<sp_genesis_builder::PresetId> { Default::default() } } }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/runtime/src/contracts_config.rs
runtime/src/contracts_config.rs
use crate::{ Balance, Balances, BalancesCall, Perbill, RandomnessCollectiveFlip, Runtime, RuntimeCall, RuntimeEvent, RuntimeHoldReason, Timestamp, }; use frame_support::{ parameter_types, traits::{ConstBool, ConstU32}, }; use frame_system::EnsureSigned; pub enum AllowBalancesCall {} impl frame_support::traits::Contains<RuntimeCall> for AllowBalancesCall { fn contains(call: &RuntimeCall) -> bool { matches!(call, RuntimeCall::Balances(BalancesCall::transfer_allow_death { .. })) } } // Unit = the base number of indivisible units for balances const UNIT: Balance = 1_000_000_000_000; const MILLIUNIT: Balance = 1_000_000_000; const fn deposit(items: u32, bytes: u32) -> Balance { (items as Balance * UNIT + (bytes as Balance) * (5 * MILLIUNIT / 100)) / 10 } fn schedule<T: pallet_contracts::Config>() -> pallet_contracts::Schedule<T> { pallet_contracts::Schedule { limits: pallet_contracts::Limits { runtime_memory: 1024 * 1024 * 1024, validator_runtime_memory: 1024 * 1024 * 1024 * 2, ..Default::default() }, ..Default::default() } } parameter_types! { pub const DepositPerItem: Balance = deposit(1, 0); pub const DepositPerByte: Balance = deposit(0, 1); pub Schedule: pallet_contracts::Schedule<Runtime> = schedule::<Runtime>(); pub const DefaultDepositLimit: Balance = deposit(1024, 1024 * 1024); pub const CodeHashLockupDepositPercent: Perbill = Perbill::from_percent(0); pub const MaxDelegateDependencies: u32 = 32; } impl pallet_contracts::Config for Runtime { type Time = Timestamp; type Randomness = RandomnessCollectiveFlip; type Currency = Balances; type RuntimeEvent = RuntimeEvent; type RuntimeCall = RuntimeCall; /// The safest default is to allow no calls at all. /// /// Runtimes should whitelist dispatchables that are allowed to be called from contracts /// and make sure they are stable. Dispatchables exposed to contracts are not allowed to /// change because that would break already deployed contracts. The `RuntimeCall` structure /// itself is not allowed to change the indices of existing pallets, too. type CallFilter = AllowBalancesCall; type DepositPerItem = DepositPerItem; type DepositPerByte = DepositPerByte; type CallStack = [pallet_contracts::Frame<Self>; 23]; type WeightPrice = pallet_transaction_payment::Pallet<Self>; type WeightInfo = pallet_contracts::weights::SubstrateWeight<Self>; type ChainExtension = (); type Schedule = Schedule; type AddressGenerator = pallet_contracts::DefaultAddressGenerator; // This node is geared towards development and testing of contracts. // We decided to increase the default allowed contract size for this // reason (the default is `128 * 1024`). // // Our reasoning is that the error code `CodeTooLarge` is thrown // if a too-large contract is uploaded. We noticed that it poses // less friction during development when the requirement here is // just more lax. type MaxCodeLen = ConstU32<{ 256 * 1024 }>; type DefaultDepositLimit = DefaultDepositLimit; type MaxStorageKeyLen = ConstU32<128>; type MaxTransientStorageSize = ConstU32<{ 1024 * 1024 }>; type MaxDebugBufferLen = ConstU32<{ 2 * 1024 * 1024 }>; type UnsafeUnstableInterface = ConstBool<true>; type CodeHashLockupDepositPercent = CodeHashLockupDepositPercent; type MaxDelegateDependencies = MaxDelegateDependencies; type RuntimeHoldReason = RuntimeHoldReason; type Environment = (); type Debug = (); type ApiVersion = (); type Migrations = (); #[cfg(feature = "parachain")] type Xcm = pallet_xcm::Pallet<Self>; #[cfg(not(feature = "parachain"))] type Xcm = (); type UploadOrigin = EnsureSigned<Self::AccountId>; type InstantiateOrigin = EnsureSigned<Self::AccountId>; }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/runtime/src/assets_config.rs
runtime/src/assets_config.rs
use crate::{AccountId, Balance, Balances, Runtime, RuntimeEvent}; use frame_support::{ parameter_types, traits::{AsEnsureOriginWithArg, ConstU128, ConstU32}, }; use frame_system::EnsureSigned; pub const MILLICENTS: Balance = 1_000_000_000; pub const CENTS: Balance = 1_000 * MILLICENTS; // assume this is worth about a cent. pub const DOLLARS: Balance = 100 * CENTS; parameter_types! { pub const AssetDeposit: Balance = 100 * DOLLARS; pub const ApprovalDeposit: Balance = 1 * DOLLARS; pub const StringLimit: u32 = 50; pub const MetadataDepositBase: Balance = 10 * DOLLARS; pub const MetadataDepositPerByte: Balance = 1 * DOLLARS; } impl pallet_assets::Config for Runtime { type RuntimeEvent = RuntimeEvent; type Balance = u128; type AssetId = u32; type AssetIdParameter = codec::Compact<u32>; type CreateOrigin = AsEnsureOriginWithArg<EnsureSigned<AccountId>>; type Currency = Balances; type ForceOrigin = frame_system::EnsureRoot<AccountId>; type AssetDeposit = AssetDeposit; type AssetAccountDeposit = ConstU128<DOLLARS>; type MetadataDepositBase = MetadataDepositBase; type MetadataDepositPerByte = MetadataDepositPerByte; type ApprovalDeposit = ApprovalDeposit; type StringLimit = StringLimit; type Freezer = (); type Extra = (); type WeightInfo = pallet_assets::weights::SubstrateWeight<Runtime>; type RemoveItemsLimit = ConstU32<1000>; type CallbackHandle = (); #[cfg(feature = "runtime-benchmarks")] type BenchmarkHelper = (); }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/runtime/src/revive_config.rs
runtime/src/revive_config.rs
use crate::{ Balance, Balances, BalancesCall, Perbill, Runtime, RuntimeCall, RuntimeEvent, RuntimeHoldReason, Timestamp, }; use frame_support::{ parameter_types, traits::{ConstBool, ConstU32}, }; use frame_system::EnsureSigned; pub enum AllowBalancesCall {} impl frame_support::traits::Contains<RuntimeCall> for AllowBalancesCall { fn contains(call: &RuntimeCall) -> bool { matches!(call, RuntimeCall::Balances(BalancesCall::transfer_allow_death { .. })) } } // Unit = the base number of indivisible units for balances const UNIT: Balance = 1_000_000_000_000; const MILLIUNIT: Balance = 1_000_000_000; const fn deposit(items: u32, bytes: u32) -> Balance { (items as Balance * UNIT + (bytes as Balance) * (5 * MILLIUNIT / 100)) / 10 } parameter_types! { pub const DepositPerItem: Balance = deposit(1, 0); pub const DepositPerByte: Balance = deposit(0, 1); pub const DefaultDepositLimit: Balance = deposit(1024, 1024 * 1024); pub const CodeHashLockupDepositPercent: Perbill = Perbill::from_percent(0); pub const MaxDelegateDependencies: u32 = 32; } impl pallet_revive::Config for Runtime { type Time = Timestamp; type Currency = Balances; type RuntimeEvent = RuntimeEvent; type RuntimeCall = RuntimeCall; /// The safest default is to allow no calls at all. /// /// Runtimes should whitelist dispatchables that are allowed to be called from contracts /// and make sure they are stable. Dispatchables exposed to contracts are not allowed to /// change because that would break already deployed contracts. The `RuntimeCall` structure /// itself is not allowed to change the indices of existing pallets, too. type CallFilter = AllowBalancesCall; type DepositPerItem = DepositPerItem; type DepositPerByte = DepositPerByte; type WeightPrice = pallet_transaction_payment::Pallet<Self>; type WeightInfo = pallet_revive::weights::SubstrateWeight<Self>; type ChainExtension = (); type AddressGenerator = pallet_revive::DefaultAddressGenerator; type MaxCodeLen = ConstU32<{ 123 * 1024 }>; type RuntimeMemory = ConstU32<{ 128 * 1024 * 1024 }>; type PVFMemory = ConstU32<{ 512 * 1024 * 1024 }>; type UnsafeUnstableInterface = ConstBool<true>; type CodeHashLockupDepositPercent = CodeHashLockupDepositPercent; type RuntimeHoldReason = RuntimeHoldReason; type Debug = (); type Migrations = (); #[cfg(feature = "parachain")] type Xcm = pallet_xcm::Pallet<Self>; #[cfg(not(feature = "parachain"))] type Xcm = (); type UploadOrigin = EnsureSigned<Self::AccountId>; type InstantiateOrigin = EnsureSigned<Self::AccountId>; }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/parachain-runtime/build.rs
parachain-runtime/build.rs
#[cfg(feature = "std")] fn main() { substrate_wasm_builder::WasmBuilder::new() .with_current_project() .export_heap_base() .import_memory() .build() } /// The wasm builder is deactivated when compiling /// this crate for wasm to speed up the compilation. #[cfg(not(feature = "std"))] fn main() {}
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/parachain-runtime/src/lib.rs
parachain-runtime/src/lib.rs
#![cfg_attr(not(feature = "std"), no_std)] // `construct_runtime!` does a lot of recursion and requires us to increase the limit to 256. #![recursion_limit = "256"] // Make the WASM binary available. #[cfg(feature = "std")] include!(concat!(env!("OUT_DIR"), "/wasm_binary.rs")); mod assets_config; mod contracts_config; mod revive_config; mod weights; mod xcm_config; extern crate alloc; use alloc::vec::Vec; use cumulus_pallet_parachain_system::RelayNumberStrictlyIncreases; use polkadot_runtime_common::xcm_sender::NoPriceForMessageDelivery; use smallvec::smallvec; use sp_api::impl_runtime_apis; use sp_core::{crypto::KeyTypeId, OpaqueMetadata}; use sp_runtime::{ create_runtime_str, generic, impl_opaque_keys, traits::{BlakeTwo256, Block as BlockT, IdentifyAccount, Verify}, transaction_validity::{TransactionSource, TransactionValidity}, ApplyExtrinsicResult, MultiSignature, }; use cumulus_primitives_core::{AggregateMessageOrigin, ParaId}; use frame_support::{ derive_impl, dispatch::DispatchClass, genesis_builder_helper::{build_state, get_preset}, parameter_types, traits::{ConstBool, ConstU32, ConstU64, ConstU8, EitherOfDiverse, TransformOrigin}, weights::{ constants::WEIGHT_REF_TIME_PER_SECOND, ConstantMultiplier, Weight, WeightToFeeCoefficient, WeightToFeeCoefficients, WeightToFeePolynomial, }, PalletId, }; use frame_system::{ limits::{BlockLength, BlockWeights}, EnsureRoot, }; use pallet_xcm::{EnsureXcm, IsVoiceOfBody}; use parachains_common::message_queue::{NarrowOriginToSibling, ParaIdToSibling}; pub use sp_consensus_aura::sr25519::AuthorityId as AuraId; pub use sp_runtime::{MultiAddress, Perbill, Permill}; #[cfg(feature = "std")] use sp_version::NativeVersion; use sp_version::RuntimeVersion; use xcm_config::{RelayLocation, XcmOriginToTransactDispatchOrigin}; #[cfg(any(feature = "std", test))] pub use sp_runtime::BuildStorage; // Polkadot imports use polkadot_runtime_common::{BlockHashCount, SlowAdjustingFeeUpdate}; use weights::{BlockExecutionWeight, ExtrinsicBaseWeight, RocksDbWeight}; // XCM Imports use xcm::latest::prelude::BodyId; /// Alias to 512-bit hash when used in the context of a transaction signature on the chain. pub type Signature = MultiSignature; /// Some way of identifying an account on the chain. We intentionally make it equivalent /// to the public key of our transaction signing scheme. pub type AccountId = <<Signature as Verify>::Signer as IdentifyAccount>::AccountId; /// Balance of an account. pub type Balance = u128; /// Index of a transaction in the chain. pub type Nonce = u32; /// A hash of some data used by the chain. pub type Hash = sp_core::H256; /// An index to a block. pub type BlockNumber = u32; /// The address format for describing accounts. pub type Address = MultiAddress<AccountId, ()>; /// Block header type as expected by this runtime. pub type Header = generic::Header<BlockNumber, BlakeTwo256>; /// Block type as expected by this runtime. pub type Block = generic::Block<Header, UncheckedExtrinsic>; /// A Block signed with a Justification pub type SignedBlock = generic::SignedBlock<Block>; /// BlockId type as expected by this runtime. pub type BlockId = generic::BlockId<Block>; type ConsensusHook = cumulus_pallet_aura_ext::FixedVelocityConsensusHook< Runtime, RELAY_CHAIN_SLOT_DURATION_MILLIS, BLOCK_PROCESSING_VELOCITY, UNINCLUDED_SEGMENT_CAPACITY, >; /// The SignedExtension to the basic transaction logic. pub type SignedExtra = ( frame_system::CheckNonZeroSender<Runtime>, frame_system::CheckSpecVersion<Runtime>, frame_system::CheckTxVersion<Runtime>, frame_system::CheckGenesis<Runtime>, frame_system::CheckEra<Runtime>, frame_system::CheckNonce<Runtime>, frame_system::CheckWeight<Runtime>, pallet_transaction_payment::ChargeTransactionPayment<Runtime>, ); /// Unchecked extrinsic type as expected by this runtime. pub type UncheckedExtrinsic = generic::UncheckedExtrinsic<Address, RuntimeCall, Signature, SignedExtra>; /// Executive: handles dispatch to the various modules. pub type Executive = frame_executive::Executive< Runtime, Block, frame_system::ChainContext<Runtime>, Runtime, AllPalletsWithSystem, >; /// Handles converting a weight scalar to a fee value, based on the scale and granularity of the /// node's balance type. /// /// This should typically create a mapping between the following ranges: /// - `[0, MAXIMUM_BLOCK_WEIGHT]` /// - `[Balance::min, Balance::max]` /// /// Yet, it can be used for any other sort of change to weight-fee. Some examples being: /// - Setting it to `0` will essentially disable the weight fee. /// - Setting it to `1` will cause the literal `#[weight = x]` values to be charged. pub struct WeightToFee; impl WeightToFeePolynomial for WeightToFee { type Balance = Balance; fn polynomial() -> WeightToFeeCoefficients<Self::Balance> { // in Rococo, extrinsic base weight (smallest non-zero weight) is mapped to 1 MILLIUNIT: // We map to 1/10 of that, or 1/10 MILLIUNIT let p = MILLIUNIT / 10; let q = 100 * Balance::from(ExtrinsicBaseWeight::get().ref_time()); smallvec![WeightToFeeCoefficient { degree: 1, negative: false, coeff_frac: Perbill::from_rational(p % q, q), coeff_integer: p / q, }] } } /// Opaque types. These are used by the CLI to instantiate machinery that don't need to know /// the specifics of the runtime. They can then be made to be agnostic over specific formats /// of data like extrinsics, allowing for them to continue syncing the network through upgrades /// to even the core data structures. pub mod opaque { use super::*; use sp_runtime::{ generic, traits::{BlakeTwo256, Hash as HashT}, }; pub use sp_runtime::OpaqueExtrinsic as UncheckedExtrinsic; /// Opaque block header type. pub type Header = generic::Header<BlockNumber, BlakeTwo256>; /// Opaque block type. pub type Block = generic::Block<Header, UncheckedExtrinsic>; /// Opaque block identifier type. pub type BlockId = generic::BlockId<Block>; /// Opaque block hash type. pub type Hash = <BlakeTwo256 as HashT>::Output; } impl_opaque_keys! { pub struct SessionKeys { pub aura: Aura, } } #[sp_version::runtime_version] pub const VERSION: RuntimeVersion = RuntimeVersion { spec_name: create_runtime_str!("contracts-parachain"), impl_name: create_runtime_str!("contracts-parachain"), authoring_version: 1, spec_version: 1, impl_version: 0, apis: RUNTIME_API_VERSIONS, transaction_version: 1, state_version: 1, }; mod block_times { /// This determines the average expected block time that we are targeting. Blocks will be /// produced at a minimum duration defined by `SLOT_DURATION`. `SLOT_DURATION` is picked up by /// `pallet_timestamp` which is in turn picked up by `pallet_aura` to implement `fn /// slot_duration()`. /// /// Change this to adjust the block time. pub const MILLISECS_PER_BLOCK: u64 = 6000; // NOTE: Currently it is not possible to change the slot duration after the chain has started. // Attempting to do so will brick block production. pub const SLOT_DURATION: u64 = MILLISECS_PER_BLOCK; } pub use block_times::*; // Time is measured by number of blocks. pub const MINUTES: BlockNumber = 60_000 / (MILLISECS_PER_BLOCK as BlockNumber); pub const HOURS: BlockNumber = MINUTES * 60; pub const DAYS: BlockNumber = HOURS * 24; // Unit = the base number of indivisible units for balances pub const UNIT: Balance = 1_000_000_000_000; pub const MILLIUNIT: Balance = 1_000_000_000; pub const MICROUNIT: Balance = 1_000_000; /// The existential deposit. Set to 1/10 of the Connected Relay Chain. pub const EXISTENTIAL_DEPOSIT: Balance = MILLIUNIT; /// We assume that ~5% of the block weight is consumed by `on_initialize` handlers. This is /// used to limit the maximal weight of a single extrinsic. const AVERAGE_ON_INITIALIZE_RATIO: Perbill = Perbill::from_percent(5); /// We allow `Normal` extrinsics to fill up the block up to 75%, the rest can be used by /// `Operational` extrinsics. const NORMAL_DISPATCH_RATIO: Perbill = Perbill::from_percent(75); /// We allow for 2 seconds of compute with a 6 second average block time. const MAXIMUM_BLOCK_WEIGHT: Weight = Weight::from_parts( WEIGHT_REF_TIME_PER_SECOND.saturating_mul(2), cumulus_primitives_core::relay_chain::MAX_POV_SIZE as u64, ); mod async_backing_params { /// Maximum number of blocks simultaneously accepted by the Runtime, not yet included /// into the relay chain. pub(crate) const UNINCLUDED_SEGMENT_CAPACITY: u32 = 3; /// How many parachain blocks are processed by the relay chain per parent. Limits the /// number of blocks authored per slot. pub(crate) const BLOCK_PROCESSING_VELOCITY: u32 = 1; /// Relay chain slot duration, in milliseconds. pub(crate) const RELAY_CHAIN_SLOT_DURATION_MILLIS: u32 = 6000; } pub(crate) use async_backing_params::*; /// The version information used to identify this runtime when compiled natively. #[cfg(feature = "std")] pub fn native_version() -> NativeVersion { NativeVersion { runtime_version: VERSION, can_author_with: Default::default() } } parameter_types! { pub const Version: RuntimeVersion = VERSION; // This part is copied from Substrate's `bin/node/runtime/src/lib.rs`. // The `RuntimeBlockLength` and `RuntimeBlockWeights` exist here because the // `DeletionWeightLimit` and `DeletionQueueDepth` depend on those to parameterize // the lazy contract deletion. pub RuntimeBlockLength: BlockLength = BlockLength::max_with_normal_ratio(5 * 1024 * 1024, NORMAL_DISPATCH_RATIO); pub RuntimeBlockWeights: BlockWeights = BlockWeights::builder() .base_block(BlockExecutionWeight::get()) .for_class(DispatchClass::all(), |weights| { weights.base_extrinsic = ExtrinsicBaseWeight::get(); }) .for_class(DispatchClass::Normal, |weights| { weights.max_total = Some(NORMAL_DISPATCH_RATIO * MAXIMUM_BLOCK_WEIGHT); }) .for_class(DispatchClass::Operational, |weights| { weights.max_total = Some(MAXIMUM_BLOCK_WEIGHT); // Operational transactions have some extra reserved space, so that they // are included even if block reached `MAXIMUM_BLOCK_WEIGHT`. weights.reserved = Some( MAXIMUM_BLOCK_WEIGHT - NORMAL_DISPATCH_RATIO * MAXIMUM_BLOCK_WEIGHT ); }) .avg_block_initialization(AVERAGE_ON_INITIALIZE_RATIO) .build_or_panic(); pub const SS58Prefix: u16 = 42; } /// The default types are being injected by [`derive_impl`](`frame_support::derive_impl`) from /// [`ParaChainDefaultConfig`](`struct@frame_system::config_preludes::ParaChainDefaultConfig`), /// but overridden as needed. #[derive_impl(frame_system::config_preludes::ParaChainDefaultConfig)] impl frame_system::Config for Runtime { /// The identifier used to distinguish between accounts. type AccountId = AccountId; /// The index type for storing how many extrinsics an account has signed. type Nonce = Nonce; /// The type for hashing blocks and tries. type Hash = Hash; /// The block type. type Block = Block; /// Maximum number of block number to block hash mappings to keep (oldest pruned first). type BlockHashCount = BlockHashCount; /// Runtime version. type Version = Version; /// The data to be stored in an account. type AccountData = pallet_balances::AccountData<Balance>; /// The weight of database operations that the runtime can invoke. type DbWeight = RocksDbWeight; /// Block & extrinsics weights: base values and limits. type BlockWeights = RuntimeBlockWeights; /// The maximum length of a block (in bytes). type BlockLength = RuntimeBlockLength; /// This is used as an identifier of the chain. 42 is the generic substrate prefix. type SS58Prefix = SS58Prefix; /// The action to take on a Runtime Upgrade type OnSetCode = cumulus_pallet_parachain_system::ParachainSetCode<Self>; type MaxConsumers = frame_support::traits::ConstU32<16>; } impl pallet_timestamp::Config for Runtime { /// A timestamp: milliseconds since the unix epoch. type Moment = u64; type OnTimestampSet = Aura; type MinimumPeriod = ConstU64<{ SLOT_DURATION / 2 }>; type WeightInfo = (); } impl pallet_authorship::Config for Runtime { type FindAuthor = pallet_session::FindAccountFromAuthorIndex<Self, Aura>; type EventHandler = (CollatorSelection,); } parameter_types! { pub const ExistentialDeposit: Balance = EXISTENTIAL_DEPOSIT; } impl pallet_balances::Config for Runtime { type MaxLocks = ConstU32<50>; /// The type for recording an account's balance. type Balance = Balance; /// The ubiquitous event type. type RuntimeEvent = RuntimeEvent; type DustRemoval = (); type ExistentialDeposit = ExistentialDeposit; type AccountStore = System; type WeightInfo = pallet_balances::weights::SubstrateWeight<Runtime>; type MaxReserves = ConstU32<50>; type ReserveIdentifier = [u8; 8]; type RuntimeHoldReason = RuntimeHoldReason; type RuntimeFreezeReason = RuntimeFreezeReason; type FreezeIdentifier = (); type MaxFreezes = ConstU32<50>; } parameter_types! { /// Relay Chain `TransactionByteFee` / 10 pub const TransactionByteFee: Balance = 10 * MICROUNIT; } impl pallet_transaction_payment::Config for Runtime { type RuntimeEvent = RuntimeEvent; type OnChargeTransaction = pallet_transaction_payment::FungibleAdapter<Balances, ()>; type WeightToFee = WeightToFee; type LengthToFee = ConstantMultiplier<Balance, TransactionByteFee>; type FeeMultiplierUpdate = SlowAdjustingFeeUpdate<Self>; type OperationalFeeMultiplier = ConstU8<5>; } impl pallet_sudo::Config for Runtime { type RuntimeEvent = RuntimeEvent; type RuntimeCall = RuntimeCall; type WeightInfo = (); } parameter_types! { pub const ReservedXcmpWeight: Weight = MAXIMUM_BLOCK_WEIGHT.saturating_div(4); pub const ReservedDmpWeight: Weight = MAXIMUM_BLOCK_WEIGHT.saturating_div(4); pub const RelayOrigin: AggregateMessageOrigin = AggregateMessageOrigin::Parent; } impl cumulus_pallet_parachain_system::Config for Runtime { type WeightInfo = (); type RuntimeEvent = RuntimeEvent; type OnSystemEvent = (); type SelfParaId = parachain_info::Pallet<Runtime>; type OutboundXcmpMessageSource = XcmpQueue; type DmpQueue = frame_support::traits::EnqueueWithOrigin<MessageQueue, RelayOrigin>; type ReservedDmpWeight = ReservedDmpWeight; type XcmpMessageHandler = XcmpQueue; type ReservedXcmpWeight = ReservedXcmpWeight; type CheckAssociatedRelayNumber = RelayNumberStrictlyIncreases; type ConsensusHook = ConsensusHook; } impl parachain_info::Config for Runtime {} parameter_types! { pub MessageQueueServiceWeight: Weight = Perbill::from_percent(35) * RuntimeBlockWeights::get().max_block; } impl pallet_message_queue::Config for Runtime { type RuntimeEvent = RuntimeEvent; type WeightInfo = (); #[cfg(feature = "runtime-benchmarks")] type MessageProcessor = pallet_message_queue::mock_helpers::NoopMessageProcessor< cumulus_primitives_core::AggregateMessageOrigin, >; #[cfg(not(feature = "runtime-benchmarks"))] type MessageProcessor = xcm_builder::ProcessXcmMessage< AggregateMessageOrigin, xcm_executor::XcmExecutor<xcm_config::XcmConfig>, RuntimeCall, >; type Size = u32; // The XCMP queue pallet is only ever able to handle the `Sibling(ParaId)` origin: type QueueChangeHandler = NarrowOriginToSibling<XcmpQueue>; type QueuePausedQuery = NarrowOriginToSibling<XcmpQueue>; type HeapSize = ConstU32<{ 64 * 1024 }>; type MaxStale = ConstU32<8>; type ServiceWeight = MessageQueueServiceWeight; type IdleMaxServiceWeight = (); } impl cumulus_pallet_aura_ext::Config for Runtime {} impl cumulus_pallet_xcmp_queue::Config for Runtime { type RuntimeEvent = RuntimeEvent; type ChannelInfo = ParachainSystem; type VersionWrapper = (); // Enqueue XCMP messages from siblings for later processing. type XcmpQueue = TransformOrigin<MessageQueue, AggregateMessageOrigin, ParaId, ParaIdToSibling>; type MaxInboundSuspended = ConstU32<1_000>; type MaxActiveOutboundChannels = ConstU32<128>; type MaxPageSize = ConstU32<{ 1 << 16 }>; type ControllerOrigin = EnsureRoot<AccountId>; type ControllerOriginConverter = XcmOriginToTransactDispatchOrigin; type WeightInfo = (); type PriceForSiblingDelivery = NoPriceForMessageDelivery<ParaId>; } parameter_types! { pub const Period: u32 = 6 * HOURS; pub const Offset: u32 = 0; } impl pallet_session::Config for Runtime { type RuntimeEvent = RuntimeEvent; type ValidatorId = <Self as frame_system::Config>::AccountId; // we don't have stash and controller, thus we don't need the convert as well. type ValidatorIdOf = pallet_collator_selection::IdentityCollator; type ShouldEndSession = pallet_session::PeriodicSessions<Period, Offset>; type NextSessionRotation = pallet_session::PeriodicSessions<Period, Offset>; type SessionManager = CollatorSelection; // Essentially just Aura, but let's be pedantic. type SessionHandler = <SessionKeys as sp_runtime::traits::OpaqueKeys>::KeyTypeIdProviders; type Keys = SessionKeys; type WeightInfo = (); } impl pallet_aura::Config for Runtime { type AuthorityId = AuraId; type DisabledValidators = (); type MaxAuthorities = ConstU32<100_000>; type AllowMultipleBlocksPerSlot = ConstBool<false>; type SlotDuration = pallet_aura::MinimumPeriodTimesTwo<Self>; } parameter_types! { pub const PotId: PalletId = PalletId(*b"PotStake"); pub const SessionLength: BlockNumber = 6 * HOURS; // StakingAdmin pluralistic body. pub const StakingAdminBodyId: BodyId = BodyId::Defense; } /// We allow root and the StakingAdmin to execute privileged collator selection operations. pub type CollatorSelectionUpdateOrigin = EitherOfDiverse< EnsureRoot<AccountId>, EnsureXcm<IsVoiceOfBody<RelayLocation, StakingAdminBodyId>>, >; impl pallet_collator_selection::Config for Runtime { type RuntimeEvent = RuntimeEvent; type Currency = Balances; type UpdateOrigin = CollatorSelectionUpdateOrigin; type PotId = PotId; type MaxCandidates = ConstU32<100>; type MinEligibleCollators = ConstU32<4>; type MaxInvulnerables = ConstU32<20>; // should be a multiple of session or things will get inconsistent type KickThreshold = Period; type ValidatorId = <Self as frame_system::Config>::AccountId; type ValidatorIdOf = pallet_collator_selection::IdentityCollator; type ValidatorRegistration = Session; type WeightInfo = (); } pub use pallet_balances::Call as BalancesCall; impl pallet_insecure_randomness_collective_flip::Config for Runtime {} impl pallet_utility::Config for Runtime { type RuntimeEvent = RuntimeEvent; type RuntimeCall = RuntimeCall; type PalletsOrigin = OriginCaller; type WeightInfo = pallet_utility::weights::SubstrateWeight<Runtime>; } // Create the runtime by composing the FRAME pallets that were previously configured. #[frame_support::runtime] mod runtime { #[runtime::runtime] #[runtime::derive( RuntimeCall, RuntimeEvent, RuntimeError, RuntimeOrigin, RuntimeFreezeReason, RuntimeHoldReason, RuntimeSlashReason, RuntimeLockId, RuntimeTask )] pub struct Runtime; // Order should match with Runtime defined in runtime/src/lib.rs #[runtime::pallet_index(0)] pub type System = frame_system; #[runtime::pallet_index(1)] pub type RandomnessCollectiveFlip = pallet_insecure_randomness_collective_flip; #[runtime::pallet_index(2)] pub type Utility = pallet_utility; #[runtime::pallet_index(3)] pub type Timestamp = pallet_timestamp; #[runtime::pallet_index(4)] pub type Balances = pallet_balances; #[runtime::pallet_index(5)] pub type Authorship = pallet_authorship; #[runtime::pallet_index(6)] pub type TransactionPayment = pallet_transaction_payment; #[runtime::pallet_index(7)] pub type Sudo = pallet_sudo; #[runtime::pallet_index(8)] pub type Contracts = pallet_contracts; #[runtime::pallet_index(9)] pub type Revive = pallet_revive; #[runtime::pallet_index(10)] pub type Assets = pallet_assets; // Parachain support. #[runtime::pallet_index(11)] pub type ParachainSystem = cumulus_pallet_parachain_system; #[runtime::pallet_index(12)] pub type ParachainInfo = parachain_info; // Collator support. The order of these 4 are important and shall not change. #[runtime::pallet_index(13)] pub type CollatorSelection = pallet_collator_selection; #[runtime::pallet_index(14)] pub type Session = pallet_session; #[runtime::pallet_index(15)] pub type Aura = pallet_aura; #[runtime::pallet_index(16)] pub type AuraExt = cumulus_pallet_aura_ext; // XCM helpers. #[runtime::pallet_index(17)] pub type XcmpQueue = cumulus_pallet_xcmp_queue; #[runtime::pallet_index(18)] pub type PolkadotXcm = pallet_xcm; #[runtime::pallet_index(19)] pub type CumulusXcm = cumulus_pallet_xcm; #[runtime::pallet_index(20)] pub type MessageQueue = pallet_message_queue; } #[cfg(feature = "runtime-benchmarks")] mod benches { frame_benchmarking::define_benchmarks!( [frame_system, SystemBench::<Runtime>] [pallet_balances, Balances] [pallet_session, SessionBench::<Runtime>] [pallet_timestamp, Timestamp] [pallet_message_queue, MessageQueue] [pallet_sudo, Sudo] [pallet_collator_selection, CollatorSelection] [cumulus_pallet_parachain_system, ParachainSystem] [cumulus_pallet_xcmp_queue, XcmpQueue] ); } type EventRecord = frame_system::EventRecord< <Runtime as frame_system::Config>::RuntimeEvent, <Runtime as frame_system::Config>::Hash, >; // Prints debug output of the `revive` pallet to stdout if the node is // started with `-lruntime::revive=trace` or `-lruntime::contracts=debug`. const CONTRACTS_DEBUG_OUTPUT: pallet_contracts::DebugInfo = pallet_contracts::DebugInfo::UnsafeDebug; const CONTRACTS_EVENTS: pallet_contracts::CollectEvents = pallet_contracts::CollectEvents::UnsafeCollect; const REVIVE_DEBUG_OUTPUT: pallet_revive::DebugInfo = pallet_revive::DebugInfo::UnsafeDebug; const REVIVE_EVENTS: pallet_revive::CollectEvents = pallet_revive::CollectEvents::UnsafeCollect; impl_runtime_apis! { impl sp_consensus_aura::AuraApi<Block, AuraId> for Runtime { fn slot_duration() -> sp_consensus_aura::SlotDuration { sp_consensus_aura::SlotDuration::from_millis(SLOT_DURATION) } fn authorities() -> Vec<AuraId> { pallet_aura::Authorities::<Runtime>::get().into_inner() } } impl cumulus_primitives_aura::AuraUnincludedSegmentApi<Block> for Runtime { fn can_build_upon( included_hash: <Block as BlockT>::Hash, slot: cumulus_primitives_aura::Slot ) -> bool { ConsensusHook::can_build_upon(included_hash, slot) } } impl sp_api::Core<Block> for Runtime { fn version() -> RuntimeVersion { VERSION } fn execute_block(block: Block) { Executive::execute_block(block) } fn initialize_block(header: &<Block as BlockT>::Header) -> sp_runtime::ExtrinsicInclusionMode { Executive::initialize_block(header) } } impl sp_api::Metadata<Block> for Runtime { fn metadata() -> OpaqueMetadata { OpaqueMetadata::new(Runtime::metadata().into()) } fn metadata_at_version(version: u32) -> Option<OpaqueMetadata> { Runtime::metadata_at_version(version) } fn metadata_versions() -> Vec<u32> { Runtime::metadata_versions() } } impl sp_block_builder::BlockBuilder<Block> for Runtime { fn apply_extrinsic(extrinsic: <Block as BlockT>::Extrinsic) -> ApplyExtrinsicResult { Executive::apply_extrinsic(extrinsic) } fn finalize_block() -> <Block as BlockT>::Header { Executive::finalize_block() } fn inherent_extrinsics(data: sp_inherents::InherentData) -> Vec<<Block as BlockT>::Extrinsic> { data.create_extrinsics() } fn check_inherents( block: Block, data: sp_inherents::InherentData, ) -> sp_inherents::CheckInherentsResult { data.check_extrinsics(&block) } } impl sp_transaction_pool::runtime_api::TaggedTransactionQueue<Block> for Runtime { fn validate_transaction( source: TransactionSource, tx: <Block as BlockT>::Extrinsic, block_hash: <Block as BlockT>::Hash, ) -> TransactionValidity { Executive::validate_transaction(source, tx, block_hash) } } impl sp_offchain::OffchainWorkerApi<Block> for Runtime { fn offchain_worker(header: &<Block as BlockT>::Header) { Executive::offchain_worker(header) } } impl sp_session::SessionKeys<Block> for Runtime { fn generate_session_keys(seed: Option<Vec<u8>>) -> Vec<u8> { SessionKeys::generate(seed) } fn decode_session_keys( encoded: Vec<u8>, ) -> Option<Vec<(Vec<u8>, KeyTypeId)>> { SessionKeys::decode_into_raw_public_keys(&encoded) } } impl frame_system_rpc_runtime_api::AccountNonceApi<Block, AccountId, Nonce> for Runtime { fn account_nonce(account: AccountId) -> Nonce { System::account_nonce(account) } } impl pallet_transaction_payment_rpc_runtime_api::TransactionPaymentApi<Block, Balance> for Runtime { fn query_info( uxt: <Block as BlockT>::Extrinsic, len: u32, ) -> pallet_transaction_payment_rpc_runtime_api::RuntimeDispatchInfo<Balance> { TransactionPayment::query_info(uxt, len) } fn query_fee_details( uxt: <Block as BlockT>::Extrinsic, len: u32, ) -> pallet_transaction_payment::FeeDetails<Balance> { TransactionPayment::query_fee_details(uxt, len) } fn query_weight_to_fee(weight: Weight) -> Balance { TransactionPayment::weight_to_fee(weight) } fn query_length_to_fee(length: u32) -> Balance { TransactionPayment::length_to_fee(length) } } impl pallet_transaction_payment_rpc_runtime_api::TransactionPaymentCallApi<Block, Balance, RuntimeCall> for Runtime { fn query_call_info( call: RuntimeCall, len: u32, ) -> pallet_transaction_payment::RuntimeDispatchInfo<Balance> { TransactionPayment::query_call_info(call, len) } fn query_call_fee_details( call: RuntimeCall, len: u32, ) -> pallet_transaction_payment::FeeDetails<Balance> { TransactionPayment::query_call_fee_details(call, len) } fn query_weight_to_fee(weight: Weight) -> Balance { TransactionPayment::weight_to_fee(weight) } fn query_length_to_fee(length: u32) -> Balance { TransactionPayment::length_to_fee(length) } } impl cumulus_primitives_core::CollectCollationInfo<Block> for Runtime { fn collect_collation_info(header: &<Block as BlockT>::Header) -> cumulus_primitives_core::CollationInfo { ParachainSystem::collect_collation_info(header) } } #[cfg(feature = "try-runtime")] impl frame_try_runtime::TryRuntime<Block> for Runtime { fn on_runtime_upgrade(checks: frame_try_runtime::UpgradeCheckSelect) -> (Weight, Weight) { let weight = Executive::try_runtime_upgrade(checks).unwrap(); (weight, RuntimeBlockWeights::get().max_block) } fn execute_block( block: Block, state_root_check: bool, signature_check: bool, select: frame_try_runtime::TryStateSelect, ) -> Weight { // NOTE: intentional unwrap: we don't want to propagate the error backwards, and want to // have a backtrace here. Executive::try_execute_block(block, state_root_check, signature_check, select).unwrap() } } #[cfg(feature = "runtime-benchmarks")] impl frame_benchmarking::Benchmark<Block> for Runtime { fn benchmark_metadata(extra: bool) -> ( Vec<frame_benchmarking::BenchmarkList>, Vec<frame_support::traits::StorageInfo>, ) { use frame_benchmarking::{Benchmarking, BenchmarkList}; use frame_support::traits::StorageInfoTrait; use frame_system_benchmarking::Pallet as SystemBench; use cumulus_pallet_session_benchmarking::Pallet as SessionBench; let mut list = Vec::<BenchmarkList>::new(); list_benchmarks!(list, extra); let storage_info = AllPalletsWithSystem::storage_info(); (list, storage_info) } fn dispatch_benchmark( config: frame_benchmarking::BenchmarkConfig ) -> Result<Vec<frame_benchmarking::BenchmarkBatch>, sp_runtime::RuntimeString> { use frame_benchmarking::{BenchmarkError, Benchmarking, BenchmarkBatch}; use frame_system_benchmarking::Pallet as SystemBench; impl frame_system_benchmarking::Config for Runtime { fn setup_set_code_requirements(code: &Vec<u8>) -> Result<(), BenchmarkError> { ParachainSystem::initialize_for_set_code_benchmark(code.len() as u32); Ok(()) } fn verify_set_code() { System::assert_last_event(cumulus_pallet_parachain_system::Event::<Runtime>::ValidationFunctionStored.into()); } } use cumulus_pallet_session_benchmarking::Pallet as SessionBench; impl cumulus_pallet_session_benchmarking::Config for Runtime {} use frame_support::traits::WhitelistedStorageKeys; let whitelist = AllPalletsWithSystem::whitelisted_storage_keys(); let mut batches = Vec::<BenchmarkBatch>::new(); let params = (&config, &whitelist); add_benchmarks!(params, batches); if batches.is_empty() { return Err("Benchmark not found for this pallet.".into()) } Ok(batches) } } impl pallet_revive::ReviveApi<Block, AccountId, Balance, BlockNumber, Hash, EventRecord> for Runtime { fn call( origin: AccountId, dest: AccountId, value: Balance, gas_limit: Option<Weight>, storage_deposit_limit: Option<Balance>, input_data: Vec<u8>, ) -> pallet_revive::ContractExecResult<Balance, EventRecord> { Revive::bare_call( RuntimeOrigin::signed(origin), dest, value, gas_limit.unwrap_or(RuntimeBlockWeights::get().max_block), storage_deposit_limit.unwrap_or(u128::MAX), input_data, REVIVE_DEBUG_OUTPUT, REVIVE_EVENTS, ) } fn instantiate( origin: AccountId, value: Balance, gas_limit: Option<Weight>, storage_deposit_limit: Option<Balance>, code: pallet_revive::Code<Hash>, data: Vec<u8>, salt: Vec<u8>, ) -> pallet_revive::ContractInstantiateResult<AccountId, Balance, EventRecord> { Revive::bare_instantiate( RuntimeOrigin::signed(origin), value, gas_limit.unwrap_or(RuntimeBlockWeights::get().max_block), storage_deposit_limit.unwrap_or(u128::MAX), code, data, salt, REVIVE_DEBUG_OUTPUT, REVIVE_EVENTS, ) } fn upload_code( origin: AccountId, code: Vec<u8>, storage_deposit_limit: Option<Balance>, ) -> pallet_revive::CodeUploadResult<Hash, Balance> { Revive::bare_upload_code( RuntimeOrigin::signed(origin), code, storage_deposit_limit.unwrap_or(u128::MAX), ) } fn get_storage( address: AccountId, key: Vec<u8>, ) -> pallet_revive::GetStorageResult { Revive::get_storage( address, key ) } } impl pallet_contracts::ContractsApi<Block, AccountId, Balance, BlockNumber, Hash, EventRecord> for Runtime { fn call( origin: AccountId, dest: AccountId, value: Balance, gas_limit: Option<Weight>, storage_deposit_limit: Option<Balance>, input_data: Vec<u8>, ) -> pallet_contracts::ContractExecResult<Balance, EventRecord> { let gas_limit = gas_limit.unwrap_or(RuntimeBlockWeights::get().max_block); Contracts::bare_call( origin, dest, value, gas_limit, storage_deposit_limit, input_data, CONTRACTS_DEBUG_OUTPUT, CONTRACTS_EVENTS, pallet_contracts::Determinism::Enforced, ) } fn instantiate( origin: AccountId, value: Balance, gas_limit: Option<Weight>, storage_deposit_limit: Option<Balance>, code: pallet_contracts::Code<Hash>, data: Vec<u8>, salt: Vec<u8>, ) -> pallet_contracts::ContractInstantiateResult<AccountId, Balance, EventRecord> { let gas_limit = gas_limit.unwrap_or(RuntimeBlockWeights::get().max_block); Contracts::bare_instantiate( origin, value, gas_limit, storage_deposit_limit, code, data, salt, CONTRACTS_DEBUG_OUTPUT, CONTRACTS_EVENTS, ) } fn upload_code( origin: AccountId, code: Vec<u8>, storage_deposit_limit: Option<Balance>, determinism: pallet_contracts::Determinism, ) -> pallet_contracts::CodeUploadResult<Hash, Balance> { Contracts::bare_upload_code(origin, code, storage_deposit_limit, determinism) } fn get_storage( address: AccountId, key: Vec<u8>, ) -> pallet_contracts::GetStorageResult { Contracts::get_storage(address, key) } } impl sp_genesis_builder::GenesisBuilder<Block> for Runtime { fn build_state(config: Vec<u8>) -> sp_genesis_builder::Result { build_state::<RuntimeGenesisConfig>(config) } fn get_preset(id: &Option<sp_genesis_builder::PresetId>) -> Option<Vec<u8>> { get_preset::<RuntimeGenesisConfig>(id, |_| None) } fn preset_names() -> Vec<sp_genesis_builder::PresetId> { Default::default() } } } cumulus_pallet_parachain_system::register_validate_block! { Runtime = Runtime, BlockExecutor = cumulus_pallet_aura_ext::BlockExecutor::<Runtime, Executive>, }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/parachain-runtime/src/contracts_config.rs
parachain-runtime/src/contracts_config.rs
use crate::{ Balance, Balances, BalancesCall, Perbill, RandomnessCollectiveFlip, Runtime, RuntimeCall, RuntimeEvent, RuntimeHoldReason, Timestamp, }; use frame_support::{ parameter_types, traits::{ConstBool, ConstU32}, }; use frame_system::EnsureSigned; pub enum AllowBalancesCall {} impl frame_support::traits::Contains<RuntimeCall> for AllowBalancesCall { fn contains(call: &RuntimeCall) -> bool { matches!(call, RuntimeCall::Balances(BalancesCall::transfer_allow_death { .. })) } } // Unit = the base number of indivisible units for balances const UNIT: Balance = 1_000_000_000_000; const MILLIUNIT: Balance = 1_000_000_000; const fn deposit(items: u32, bytes: u32) -> Balance { (items as Balance * UNIT + (bytes as Balance) * (5 * MILLIUNIT / 100)) / 10 } fn schedule<T: pallet_contracts::Config>() -> pallet_contracts::Schedule<T> { pallet_contracts::Schedule { limits: pallet_contracts::Limits { runtime_memory: 1024 * 1024 * 1024, validator_runtime_memory: 1024 * 1024 * 1024 * 2, ..Default::default() }, ..Default::default() } } parameter_types! { pub const DepositPerItem: Balance = deposit(1, 0); pub const DepositPerByte: Balance = deposit(0, 1); pub Schedule: pallet_contracts::Schedule<Runtime> = schedule::<Runtime>(); pub const DefaultDepositLimit: Balance = deposit(1024, 1024 * 1024); pub const CodeHashLockupDepositPercent: Perbill = Perbill::from_percent(0); pub const MaxDelegateDependencies: u32 = 32; } impl pallet_contracts::Config for Runtime { type Time = Timestamp; type Randomness = RandomnessCollectiveFlip; type Currency = Balances; type RuntimeEvent = RuntimeEvent; type RuntimeCall = RuntimeCall; /// The safest default is to allow no calls at all. /// /// Runtimes should whitelist dispatchables that are allowed to be called from contracts /// and make sure they are stable. Dispatchables exposed to contracts are not allowed to /// change because that would break already deployed contracts. The `RuntimeCall` structure /// itself is not allowed to change the indices of existing pallets, too. type CallFilter = AllowBalancesCall; type DepositPerItem = DepositPerItem; type DepositPerByte = DepositPerByte; type CallStack = [pallet_contracts::Frame<Self>; 23]; type WeightPrice = pallet_transaction_payment::Pallet<Self>; type WeightInfo = pallet_contracts::weights::SubstrateWeight<Self>; type ChainExtension = (); type Schedule = Schedule; type AddressGenerator = pallet_contracts::DefaultAddressGenerator; // This node is geared towards development and testing of contracts. // We decided to increase the default allowed contract size for this // reason (the default is `128 * 1024`). // // Our reasoning is that the error code `CodeTooLarge` is thrown // if a too-large contract is uploaded. We noticed that it poses // less friction during development when the requirement here is // just more lax. type MaxCodeLen = ConstU32<{ 256 * 1024 }>; type DefaultDepositLimit = DefaultDepositLimit; type MaxStorageKeyLen = ConstU32<128>; type MaxTransientStorageSize = ConstU32<{ 1024 * 1024 }>; type MaxDebugBufferLen = ConstU32<{ 2 * 1024 * 1024 }>; type UnsafeUnstableInterface = ConstBool<true>; type CodeHashLockupDepositPercent = CodeHashLockupDepositPercent; type MaxDelegateDependencies = MaxDelegateDependencies; type RuntimeHoldReason = RuntimeHoldReason; type Environment = (); type Debug = (); type ApiVersion = (); type Migrations = (); #[cfg(feature = "parachain")] type Xcm = pallet_xcm::Pallet<Self>; #[cfg(not(feature = "parachain"))] type Xcm = (); type UploadOrigin = EnsureSigned<Self::AccountId>; type InstantiateOrigin = EnsureSigned<Self::AccountId>; }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/parachain-runtime/src/assets_config.rs
parachain-runtime/src/assets_config.rs
use crate::{AccountId, Balance, Balances, Runtime, RuntimeEvent}; use frame_support::{ parameter_types, traits::{AsEnsureOriginWithArg, ConstU128, ConstU32}, }; use frame_system::EnsureSigned; pub const MILLICENTS: Balance = 1_000_000_000; pub const CENTS: Balance = 1_000 * MILLICENTS; // assume this is worth about a cent. pub const DOLLARS: Balance = 100 * CENTS; parameter_types! { pub const AssetDeposit: Balance = 100 * DOLLARS; pub const ApprovalDeposit: Balance = 1 * DOLLARS; pub const StringLimit: u32 = 50; pub const MetadataDepositBase: Balance = 10 * DOLLARS; pub const MetadataDepositPerByte: Balance = 1 * DOLLARS; } impl pallet_assets::Config for Runtime { type RuntimeEvent = RuntimeEvent; type Balance = u128; type AssetId = u32; type AssetIdParameter = codec::Compact<u32>; type CreateOrigin = AsEnsureOriginWithArg<EnsureSigned<AccountId>>; type Currency = Balances; type ForceOrigin = frame_system::EnsureRoot<AccountId>; type AssetDeposit = AssetDeposit; type AssetAccountDeposit = ConstU128<DOLLARS>; type MetadataDepositBase = MetadataDepositBase; type MetadataDepositPerByte = MetadataDepositPerByte; type ApprovalDeposit = ApprovalDeposit; type StringLimit = StringLimit; type Freezer = (); type Extra = (); type WeightInfo = pallet_assets::weights::SubstrateWeight<Runtime>; type RemoveItemsLimit = ConstU32<1000>; type CallbackHandle = (); #[cfg(feature = "runtime-benchmarks")] type BenchmarkHelper = (); }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/parachain-runtime/src/revive_config.rs
parachain-runtime/src/revive_config.rs
use crate::{ Balance, Balances, BalancesCall, Perbill, Runtime, RuntimeCall, RuntimeEvent, RuntimeHoldReason, Timestamp, }; use frame_support::{ parameter_types, traits::{ConstBool, ConstU32}, }; use frame_system::EnsureSigned; pub enum AllowBalancesCall {} impl frame_support::traits::Contains<RuntimeCall> for AllowBalancesCall { fn contains(call: &RuntimeCall) -> bool { matches!(call, RuntimeCall::Balances(BalancesCall::transfer_allow_death { .. })) } } // Unit = the base number of indivisible units for balances const UNIT: Balance = 1_000_000_000_000; const MILLIUNIT: Balance = 1_000_000_000; const fn deposit(items: u32, bytes: u32) -> Balance { (items as Balance * UNIT + (bytes as Balance) * (5 * MILLIUNIT / 100)) / 10 } parameter_types! { pub const DepositPerItem: Balance = deposit(1, 0); pub const DepositPerByte: Balance = deposit(0, 1); pub const DefaultDepositLimit: Balance = deposit(1024, 1024 * 1024); pub const CodeHashLockupDepositPercent: Perbill = Perbill::from_percent(0); pub const MaxDelegateDependencies: u32 = 32; } impl pallet_revive::Config for Runtime { type Time = Timestamp; type Currency = Balances; type RuntimeEvent = RuntimeEvent; type RuntimeCall = RuntimeCall; /// The safest default is to allow no calls at all. /// /// Runtimes should whitelist dispatchables that are allowed to be called from contracts /// and make sure they are stable. Dispatchables exposed to contracts are not allowed to /// change because that would break already deployed contracts. The `RuntimeCall` structure /// itself is not allowed to change the indices of existing pallets, too. type CallFilter = AllowBalancesCall; type DepositPerItem = DepositPerItem; type DepositPerByte = DepositPerByte; type WeightPrice = pallet_transaction_payment::Pallet<Self>; type WeightInfo = pallet_revive::weights::SubstrateWeight<Self>; type ChainExtension = (); type AddressGenerator = pallet_revive::DefaultAddressGenerator; type MaxCodeLen = ConstU32<{ 123 * 1024 }>; type RuntimeMemory = ConstU32<{ 128 * 1024 * 1024 }>; type PVFMemory = ConstU32<{ 512 * 1024 * 1024 }>; type UnsafeUnstableInterface = ConstBool<true>; type CodeHashLockupDepositPercent = CodeHashLockupDepositPercent; type RuntimeHoldReason = RuntimeHoldReason; type Debug = (); type Migrations = (); #[cfg(feature = "parachain")] type Xcm = pallet_xcm::Pallet<Self>; #[cfg(not(feature = "parachain"))] type Xcm = (); type UploadOrigin = EnsureSigned<Self::AccountId>; type InstantiateOrigin = EnsureSigned<Self::AccountId>; }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/parachain-runtime/src/xcm_config.rs
parachain-runtime/src/xcm_config.rs
use super::{ AccountId, AllPalletsWithSystem, Balances, ParachainInfo, ParachainSystem, PolkadotXcm, Runtime, RuntimeCall, RuntimeEvent, RuntimeOrigin, WeightToFee, XcmpQueue, }; use frame_support::{ parameter_types, traits::{ConstU32, Contains, Everything, Nothing}, weights::Weight, }; use frame_system::EnsureRoot; use pallet_xcm::XcmPassthrough; use polkadot_runtime_common::impls::ToAuthor; use xcm::latest::prelude::*; #[allow(deprecated)] use xcm_builder::CurrencyAdapter; use xcm_builder::{ AccountId32Aliases, AllowExplicitUnpaidExecutionFrom, AllowTopLevelPaidExecutionFrom, DenyReserveTransferToRelayChain, DenyThenTry, EnsureXcmOrigin, FixedWeightBounds, FrameTransactionalProcessor, IsConcrete, NativeAsset, ParentIsPreset, RelayChainAsNative, SiblingParachainAsNative, SignedAccountId32AsNative, SignedToAccountId32, SovereignSignedViaLocation, TakeWeightCredit, TrailingSetTopicAsId, UsingComponents, WithComputedOrigin, WithUniqueTopic, }; use xcm_executor::XcmExecutor; parameter_types! { pub const TokenLocation: Location = Here.into_location(); pub const RelayLocation: Location = Location::parent(); pub const RelayNetwork: Option<NetworkId> = None; pub RelayChainOrigin: RuntimeOrigin = cumulus_pallet_xcm::Origin::Relay.into(); pub UniversalLocation: InteriorLocation = Parachain(ParachainInfo::parachain_id().into()).into(); } /// Type for specifying how a `Location` can be converted into an `AccountId`. This is used /// when determining ownership of accounts for asset transacting and when attempting to use XCM /// `Transact` in order to determine the dispatch Origin. pub type LocationToAccountId = ( // Straight up local `AccountId32` origins just alias directly to `AccountId`. AccountId32Aliases<RelayNetwork, AccountId>, // The parent (Relay-chain) origin converts to the parent `AccountId`. ParentIsPreset<AccountId>, ); /// Means for transacting assets on this chain. #[allow(deprecated)] pub type LocalAssetTransactor = CurrencyAdapter< // Use this currency: Balances, // Use this currency when it is a fungible asset matching the given location or name: IsConcrete<RelayLocation>, // Do a simple punn to convert an AccountId32 Location into a native chain account ID: LocationToAccountId, // Our chain's account ID type (we can't get away without mentioning it explicitly): AccountId, // We don't track any teleports. (), >; #[allow(deprecated)] pub type LocalBalancesTransactor = CurrencyAdapter<Balances, IsConcrete<TokenLocation>, LocationToAccountId, AccountId, ()>; pub type AssetTransactors = (LocalBalancesTransactor, LocalAssetTransactor); /// This is the type we use to convert an (incoming) XCM origin into a local `Origin` instance, /// ready for dispatching a transaction with Xcm's `Transact`. There is an `OriginKind` which can /// biases the kind of local `Origin` it will become. pub type XcmOriginToTransactDispatchOrigin = ( // Sovereign account converter; this attempts to derive an `AccountId` from the origin location // using `LocationToAccountId` and then turn that into the usual `Signed` origin. Useful for // foreign chains who want to have a local sovereign account on this chain which they control. SovereignSignedViaLocation<LocationToAccountId, RuntimeOrigin>, // Native converter for Relay-chain (Parent) location; will convert to a `Relay` origin when // recognized. RelayChainAsNative<RelayChainOrigin, RuntimeOrigin>, // Native converter for sibling Parachains; will convert to a `SiblingPara` origin when // recognized. SiblingParachainAsNative<cumulus_pallet_xcm::Origin, RuntimeOrigin>, // Native signed account converter; this just converts an `AccountId32` origin into a normal // `RuntimeOrigin::Signed` origin of the same 32-byte value. SignedAccountId32AsNative<RelayNetwork, RuntimeOrigin>, // Xcm origins can be represented natively under the Xcm pallet's Xcm origin. XcmPassthrough<RuntimeOrigin>, ); parameter_types! { // One XCM operation is 1_000_000_000 weight - almost certainly a conservative estimate. pub UnitWeightCost: Weight = Weight::from_parts(1_000_000_000, 64 * 1024); pub const MaxInstructions: u32 = 100; pub const MaxAssetsIntoHolding: u32 = 64; } pub struct ParentOrParentsPlurality; impl Contains<Location> for ParentOrParentsPlurality { fn contains(location: &Location) -> bool { matches!(location.unpack(), (1, []) | (1, [Plurality { .. }])) } } pub type Barrier = TrailingSetTopicAsId< DenyThenTry< DenyReserveTransferToRelayChain, ( TakeWeightCredit, WithComputedOrigin< ( AllowTopLevelPaidExecutionFrom<Everything>, AllowExplicitUnpaidExecutionFrom<ParentOrParentsPlurality>, // ^^^ Parent and its exec plurality get free execution ), UniversalLocation, ConstU32<8>, >, ), >, >; pub struct XcmConfig; impl xcm_executor::Config for XcmConfig { type RuntimeCall = RuntimeCall; type XcmSender = XcmRouter; type AssetTransactor = AssetTransactors; type OriginConverter = XcmOriginToTransactDispatchOrigin; type IsReserve = NativeAsset; type IsTeleporter = (); // Teleporting is disabled. type UniversalLocation = UniversalLocation; type Barrier = Barrier; type Weigher = FixedWeightBounds<UnitWeightCost, RuntimeCall, MaxInstructions>; type Trader = UsingComponents<WeightToFee, RelayLocation, AccountId, Balances, ToAuthor<Runtime>>; type ResponseHandler = PolkadotXcm; type AssetTrap = PolkadotXcm; type AssetClaims = PolkadotXcm; type SubscriptionService = PolkadotXcm; type PalletInstancesInfo = AllPalletsWithSystem; type MaxAssetsIntoHolding = MaxAssetsIntoHolding; type AssetLocker = (); type AssetExchanger = (); type FeeManager = (); type MessageExporter = (); type UniversalAliases = Nothing; type CallDispatcher = RuntimeCall; type SafeCallFilter = Everything; type Aliasers = Nothing; type TransactionalProcessor = FrameTransactionalProcessor; type HrmpNewChannelOpenRequestHandler = (); type HrmpChannelAcceptedHandler = (); type HrmpChannelClosingHandler = (); type XcmRecorder = PolkadotXcm; } /// No local origins on this chain are allowed to dispatch XCM sends/executions. pub type LocalOriginToLocation = SignedToAccountId32<RuntimeOrigin, AccountId, RelayNetwork>; /// The means for routing XCM messages which are not for local execution into the right message /// queues. pub type XcmRouter = WithUniqueTopic<( // Two routers - use UMP to communicate with the relay chain: cumulus_primitives_utility::ParentAsUmp<ParachainSystem, (), ()>, // ..and XCMP to communicate with the sibling chains. XcmpQueue, )>; #[cfg(feature = "runtime-benchmarks")] parameter_types! { pub ReachableDest: Option<Location> = Some(Parent.into()); } impl pallet_xcm::Config for Runtime { type RuntimeEvent = RuntimeEvent; type SendXcmOrigin = EnsureXcmOrigin<RuntimeOrigin, LocalOriginToLocation>; type XcmRouter = XcmRouter; type ExecuteXcmOrigin = EnsureXcmOrigin<RuntimeOrigin, LocalOriginToLocation>; type XcmExecuteFilter = Everything; type XcmExecutor = XcmExecutor<XcmConfig>; type XcmTeleportFilter = Everything; type XcmReserveTransferFilter = Nothing; type Weigher = FixedWeightBounds<UnitWeightCost, RuntimeCall, MaxInstructions>; type UniversalLocation = UniversalLocation; type RuntimeOrigin = RuntimeOrigin; type RuntimeCall = RuntimeCall; const VERSION_DISCOVERY_QUEUE_SIZE: u32 = 100; // ^ Override for AdvertisedXcmVersion default type AdvertisedXcmVersion = pallet_xcm::CurrentXcmVersion; type Currency = Balances; type CurrencyMatcher = (); type TrustedLockers = (); type SovereignAccountOf = LocationToAccountId; type MaxLockers = ConstU32<8>; type WeightInfo = pallet_xcm::TestWeightInfo; type AdminOrigin = EnsureRoot<AccountId>; type MaxRemoteLockConsumers = ConstU32<0>; type RemoteLockConsumerIdentifier = (); } impl cumulus_pallet_xcm::Config for Runtime { type RuntimeEvent = RuntimeEvent; type XcmExecutor = XcmExecutor<XcmConfig>; }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/parachain-runtime/src/weights/paritydb_weights.rs
parachain-runtime/src/weights/paritydb_weights.rs
// This file is part of Substrate. // Copyright (C) 2022 Parity Technologies (UK) Ltd. // SPDX-License-Identifier: Apache-2.0 // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. pub mod constants { use frame_support::{ parameter_types, weights::{constants, RuntimeDbWeight}, }; parameter_types! { /// `ParityDB` can be enabled with a feature flag, but is still experimental. These weights /// are available for brave runtime engineers who may want to try this out as default. pub const ParityDbWeight: RuntimeDbWeight = RuntimeDbWeight { read: 8_000 * constants::WEIGHT_REF_TIME_PER_NANOS, write: 50_000 * constants::WEIGHT_REF_TIME_PER_NANOS, }; } #[cfg(test)] mod test_db_weights { use super::constants::ParityDbWeight as W; use frame_support::weights::constants; /// Checks that all weights exist and have sane values. // NOTE: If this test fails but you are sure that the generated values are fine, // you can delete it. #[test] fn sane() { // At least 1 µs. assert!( W::get().reads(1).ref_time() >= constants::WEIGHT_REF_TIME_PER_MICROS, "Read weight should be at least 1 µs." ); assert!( W::get().writes(1).ref_time() >= constants::WEIGHT_REF_TIME_PER_MICROS, "Write weight should be at least 1 µs." ); // At most 1 ms. assert!( W::get().reads(1).ref_time() <= constants::WEIGHT_REF_TIME_PER_MILLIS, "Read weight should be at most 1 ms." ); assert!( W::get().writes(1).ref_time() <= constants::WEIGHT_REF_TIME_PER_MILLIS, "Write weight should be at most 1 ms." ); } } }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/parachain-runtime/src/weights/block_weights.rs
parachain-runtime/src/weights/block_weights.rs
// This file is part of Substrate. // Copyright (C) 2022 Parity Technologies (UK) Ltd. // SPDX-License-Identifier: Apache-2.0 // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. pub mod constants { use frame_support::{ parameter_types, weights::{constants, Weight}, }; parameter_types! { /// Importing a block with 0 Extrinsics. pub const BlockExecutionWeight: Weight = Weight::from_parts(constants::WEIGHT_REF_TIME_PER_NANOS.saturating_mul(5_000_000), 0); } #[cfg(test)] mod test_weights { use frame_support::weights::constants; /// Checks that the weight exists and is sane. // NOTE: If this test fails but you are sure that the generated values are fine, // you can delete it. #[test] fn sane() { let w = super::constants::BlockExecutionWeight::get(); // At least 100 µs. assert!( w.ref_time() >= 100u64 * constants::WEIGHT_REF_TIME_PER_MICROS, "Weight should be at least 100 µs." ); // At most 50 ms. assert!( w.ref_time() <= 50u64 * constants::WEIGHT_REF_TIME_PER_MILLIS, "Weight should be at most 50 ms." ); } } }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/parachain-runtime/src/weights/mod.rs
parachain-runtime/src/weights/mod.rs
// This file is part of Substrate. // Copyright (C) 2022 Parity Technologies (UK) Ltd. // SPDX-License-Identifier: Apache-2.0 // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //! Expose the auto generated weight files. pub mod block_weights; pub mod extrinsic_weights; pub mod paritydb_weights; pub mod rocksdb_weights; pub use block_weights::constants::BlockExecutionWeight; pub use extrinsic_weights::constants::ExtrinsicBaseWeight; pub use rocksdb_weights::constants::RocksDbWeight;
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/parachain-runtime/src/weights/rocksdb_weights.rs
parachain-runtime/src/weights/rocksdb_weights.rs
// This file is part of Substrate. // Copyright (C) 2022 Parity Technologies (UK) Ltd. // SPDX-License-Identifier: Apache-2.0 // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. pub mod constants { use frame_support::{ parameter_types, weights::{constants, RuntimeDbWeight}, }; parameter_types! { /// By default, Substrate uses `RocksDB`, so this will be the weight used throughout /// the runtime. pub const RocksDbWeight: RuntimeDbWeight = RuntimeDbWeight { read: 25_000 * constants::WEIGHT_REF_TIME_PER_NANOS, write: 100_000 * constants::WEIGHT_REF_TIME_PER_NANOS, }; } #[cfg(test)] mod test_db_weights { use super::constants::RocksDbWeight as W; use frame_support::weights::constants; /// Checks that all weights exist and have sane values. // NOTE: If this test fails but you are sure that the generated values are fine, // you can delete it. #[test] fn sane() { // At least 1 µs. assert!( W::get().reads(1).ref_time() >= constants::WEIGHT_REF_TIME_PER_MICROS, "Read weight should be at least 1 µs." ); assert!( W::get().writes(1).ref_time() >= constants::WEIGHT_REF_TIME_PER_MICROS, "Write weight should be at least 1 µs." ); // At most 1 ms. assert!( W::get().reads(1).ref_time() <= constants::WEIGHT_REF_TIME_PER_MILLIS, "Read weight should be at most 1 ms." ); assert!( W::get().writes(1).ref_time() <= constants::WEIGHT_REF_TIME_PER_MILLIS, "Write weight should be at most 1 ms." ); } } }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
paritytech/substrate-contracts-node
https://github.com/paritytech/substrate-contracts-node/blob/f209befc88cb54ff50b7483c13d19e62213d0c60/parachain-runtime/src/weights/extrinsic_weights.rs
parachain-runtime/src/weights/extrinsic_weights.rs
// This file is part of Substrate. // Copyright (C) 2022 Parity Technologies (UK) Ltd. // SPDX-License-Identifier: Apache-2.0 // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. pub mod constants { use frame_support::{ parameter_types, weights::{constants, Weight}, }; parameter_types! { /// Executing a NO-OP `System::remarks` Extrinsic. pub const ExtrinsicBaseWeight: Weight = Weight::from_parts(constants::WEIGHT_REF_TIME_PER_NANOS.saturating_mul(125_000), 0); } #[cfg(test)] mod test_weights { use frame_support::weights::constants; /// Checks that the weight exists and is sane. // NOTE: If this test fails but you are sure that the generated values are fine, // you can delete it. #[test] fn sane() { let w = super::constants::ExtrinsicBaseWeight::get(); // At least 10 µs. assert!( w.ref_time() >= 10u64 * constants::WEIGHT_REF_TIME_PER_MICROS, "Weight should be at least 10 µs." ); // At most 1 ms. assert!( w.ref_time() <= constants::WEIGHT_REF_TIME_PER_MILLIS, "Weight should be at most 1 ms." ); } } }
rust
Unlicense
f209befc88cb54ff50b7483c13d19e62213d0c60
2026-01-04T20:19:32.556120Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/lib.rs
src/lib.rs
pub mod app; pub mod mcp;
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/main.rs
src/main.rs
use clap::{Parser, Subcommand}; use rusqlite::Connection; use std::path::PathBuf; use ghost::app::{commands, config, error::Result, logging, storage}; #[derive(Parser, Debug)] #[command(name = "ghost")] #[command(about = "A simple background process manager")] #[command( long_about = "A simple background process manager.\n\nRun without arguments to start the interactive TUI mode." )] #[command(version)] struct Cli { #[command(subcommand)] command: Option<Commands>, } #[derive(Subcommand, Debug)] enum Commands { /// Run one or more commands in the background /// /// Single command: ghost run sleep 10 /// Multiple commands: ghost run "sleep 10" "echo hello" Run { /// Commands to run. For multiple commands, quote each command. /// Example: ghost run "sleep 10" "echo hello" #[arg(required = true)] commands: Vec<String>, /// Working directory for the command(s) #[arg(short, long)] cwd: Option<PathBuf>, /// Environment variables (KEY=VALUE format) #[arg(short, long)] env: Vec<String>, }, /// List all background processes List { /// Filter by status (running, exited, killed) #[arg(short, long)] status: Option<String>, }, /// Show logs for a process Log { /// Task ID to show logs for task_id: String, /// Follow log output (like tail -f) #[arg(short, long)] follow: bool, }, /// Stop a background process Stop { /// Task ID to stop task_id: String, /// Force kill the process (SIGKILL instead of SIGTERM) #[arg(short, long)] force: bool, }, /// Check status of a background process Status { /// Task ID to check task_id: String, }, /// Clean up old finished tasks Cleanup { /// Delete tasks older than this many days (default: 30) #[arg(short, long, default_value = "30")] days: u64, /// Filter by status (exited, killed, all). Default: exited,killed #[arg(short, long)] status: Option<String>, /// Show what would be deleted without actually deleting #[arg(short = 'n', long)] dry_run: bool, /// Delete all finished tasks regardless of age #[arg(short, long)] all: bool, }, /// Run MCP server for ghost operations Mcp, } #[tokio::main] async fn main() { let cli = Cli::parse(); let result = match cli.command { Some(cmd) => { // Initialize database connection once for all commands (except TUI) match storage::init_database() { Ok(conn) => match cmd { Commands::Run { commands, cwd, env } => run_commands(&conn, commands, cwd, env), Commands::List { status } => commands::list(&conn, status, true).map(|_| ()), Commands::Log { task_id, follow } => { commands::log(&conn, &task_id, follow, true) .await .map(|_| ()) } Commands::Stop { task_id, force } => { commands::stop(&conn, &task_id, force, true) } Commands::Status { task_id } => { commands::status(&conn, &task_id, true).map(|_| ()) } Commands::Cleanup { days, status, dry_run, all, } => commands::cleanup(&conn, days, status, dry_run, all), Commands::Mcp => { // Initialize file logger for MCP server let log_dir = config::get_log_dir(); let _guard = logging::init_file_logger(&log_dir); ghost::mcp::run_stdio_server(conn).await.map_err(|e| { ghost::app::error::GhostError::Config { message: e.to_string(), } }) } }, Err(e) => Err(e), } } None => commands::tui().await, }; if let Err(e) = result { eprintln!("Error: {e}"); std::process::exit(1); } } /// Run one or more commands based on the input format /// /// Determines whether to use single-command mode (backward compatible) /// or multi-command mode based on whether the first argument contains spaces. fn run_commands( conn: &Connection, args: Vec<String>, cwd: Option<PathBuf>, env: Vec<String>, ) -> Result<()> { if args.is_empty() { return Err(ghost::app::error::GhostError::InvalidArgument { message: "No command specified".to_string(), }); } // Determine if this is multi-command format: // - If first arg contains space -> multi-command format (each arg is a full command) // - Otherwise -> single-command format (all args form one command) let is_multi_command = args.first().map(|s| s.contains(' ')).unwrap_or(false); if is_multi_command { // Multi-command mode: each argument is a complete command string // Note: Error messages are printed by spawn_multi, so we don't need to handle failures here let _ = commands::spawn_multi(conn, args, cwd, env, true); Ok(()) } else { // Single-command mode: all arguments form one command (backward compatible) commands::spawn(conn, args, cwd, env, true).map(|_| ()) } }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/config.rs
src/app/config.rs
use std::path::PathBuf; /// Configuration for Ghost application #[derive(Debug, Clone)] pub struct Config { pub data_dir: PathBuf, pub log_dir: PathBuf, pub db_path: PathBuf, } impl Default for Config { fn default() -> Self { let data_dir = get_data_dir(); let log_dir = data_dir.join("logs"); let db_path = data_dir.join("tasks.db"); Config { data_dir, log_dir, db_path, } } } impl Config { /// Create a new config with custom data directory pub fn with_data_dir(data_dir: PathBuf) -> Self { let log_dir = data_dir.join("logs"); let db_path = data_dir.join("tasks.db"); Config { data_dir, log_dir, db_path, } } /// Ensure all required directories exist pub fn ensure_directories(&self) -> std::io::Result<()> { std::fs::create_dir_all(&self.data_dir)?; std::fs::create_dir_all(&self.log_dir)?; Ok(()) } /// Get the database path for this config pub fn get_db_path(&self) -> PathBuf { self.db_path.clone() } } /// Get the default data directory for Ghost pub fn get_data_dir() -> PathBuf { // Check if GHOST_DATA_DIR environment variable is set if let Ok(data_dir) = std::env::var("GHOST_DATA_DIR") { PathBuf::from(data_dir) } else { dirs::data_dir() .unwrap_or_else(|| PathBuf::from(".")) .join("ghost") } } /// Get the default log directory pub fn get_log_dir() -> PathBuf { get_data_dir().join("logs") } /// Get the default database path pub fn get_db_path() -> PathBuf { get_data_dir().join("tasks.db") } /// Environment variable parsing utilities pub mod env { use crate::app::error::{GhostError, Result}; /// Parse environment variables from KEY=VALUE format pub fn parse_env_vars(env_strings: &[String]) -> Result<Vec<(String, String)>> { let mut env_vars = Vec::new(); for env_str in env_strings { if let Some((key, value)) = env_str.split_once('=') { env_vars.push((key.to_string(), value.to_string())); } else { return Err(GhostError::InvalidArgument { message: format!( "Invalid environment variable format: {env_str}. Use KEY=VALUE" ), }); } } Ok(env_vars) } } #[cfg(test)] mod tests { use super::*; use tempfile::tempdir; #[test] fn test_config_default() { let config = Config::default(); assert!(config.data_dir.ends_with("ghost")); assert!(config.log_dir.ends_with("logs")); assert!(config.db_path.ends_with("tasks.db")); } #[test] fn test_config_with_custom_dir() { let temp_dir = tempdir().unwrap(); let config = Config::with_data_dir(temp_dir.path().to_path_buf()); assert_eq!(config.data_dir, temp_dir.path()); assert_eq!(config.log_dir, temp_dir.path().join("logs")); assert_eq!(config.db_path, temp_dir.path().join("tasks.db")); } #[test] fn test_ensure_directories() { let temp_dir = tempdir().unwrap(); let config = Config::with_data_dir(temp_dir.path().join("ghost")); config.ensure_directories().unwrap(); assert!(config.data_dir.exists()); assert!(config.log_dir.exists()); } #[test] fn test_parse_env_vars_valid() { let env_strings = vec!["KEY1=value1".to_string(), "KEY2=value2".to_string()]; let result = env::parse_env_vars(&env_strings).unwrap(); assert_eq!(result.len(), 2); assert_eq!(result[0], ("KEY1".to_string(), "value1".to_string())); assert_eq!(result[1], ("KEY2".to_string(), "value2".to_string())); } #[test] fn test_parse_env_vars_invalid() { let env_strings = vec!["INVALID_FORMAT".to_string()]; let result = env::parse_env_vars(&env_strings); assert!(result.is_err()); } }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/port_detector.rs
src/app/port_detector.rs
use crate::app::error::{GhostError, Result}; use std::process::Command; use std::sync::OnceLock; #[derive(Debug, Clone)] pub struct ListeningPort { pub protocol: String, pub local_addr: String, pub state: String, } // Cache the lsof availability check result static LSOF_AVAILABLE: OnceLock<bool> = OnceLock::new(); /// Check if lsof command is available on the system (cached) pub fn is_lsof_available() -> bool { *LSOF_AVAILABLE.get_or_init(|| { Command::new("which") .arg("lsof") .output() .map(|output| output.status.success()) .unwrap_or(false) }) } /// Check if lsof command is available on the system pub fn check_lsof_availability() -> Result<()> { if !is_lsof_available() { return Err(GhostError::CommandNotFound { command: "lsof".to_string(), }); } Ok(()) } /// Parse lsof machine-readable format (-F flag) output fn parse_lsof_machine_format(output: &str) -> Vec<ListeningPort> { let mut ports = Vec::new(); let mut current_protocol = String::new(); let mut current_addr = String::new(); let mut current_state = String::new(); let mut in_network_fd = false; for line in output.lines() { if line.is_empty() { continue; } let tag = &line[0..1]; let value = &line[1..]; match tag { "f" => { // File descriptor - reset state for new FD in_network_fd = false; current_protocol.clear(); current_addr.clear(); current_state.clear(); } "t" => { // Type - check if it's a network connection if value.starts_with("IPv4") || value.starts_with("IPv6") { in_network_fd = true; } } "n" => { // Name - contains address information if in_network_fd { current_addr = value.to_string(); } } "P" => { // Protocol - TCP or UDP if in_network_fd { current_protocol = value.to_lowercase(); } } "T" => { // TCP/TPI info - contains state like LISTEN if in_network_fd && value.starts_with("ST=LISTEN") { current_state = "LISTEN".to_string(); // We have all the information, add the port if !current_protocol.is_empty() && !current_addr.is_empty() { ports.push(ListeningPort { protocol: current_protocol.clone(), local_addr: current_addr.clone(), state: current_state.clone(), }); } } } _ => {} } } ports } /// Detect listening ports for a given process ID pub fn detect_listening_ports(pid: u32) -> Result<Vec<ListeningPort>> { // Check if lsof is available check_lsof_availability()?; #[cfg(any(target_os = "macos", target_os = "linux"))] return detect_ports_using_lsof(pid); #[cfg(not(any(target_os = "macos", target_os = "linux")))] return Ok(Vec::new()); } /// Common implementation for macOS and Linux using lsof #[cfg(any(target_os = "macos", target_os = "linux"))] fn detect_ports_using_lsof(pid: u32) -> Result<Vec<ListeningPort>> { let output = Command::new("lsof") .args(["-nP", "-i", "-a", "-p", &pid.to_string(), "-F"]) .output() .map_err(|e| GhostError::ProcessOperation { message: format!("Failed to execute lsof: {e}"), })?; if !output.status.success() { // Process might not have any network connections return Ok(Vec::new()); } let stdout = String::from_utf8_lossy(&output.stdout); Ok(parse_lsof_machine_format(&stdout)) } #[cfg(test)] mod tests { use super::*; #[test] fn test_detect_listening_ports() { // This test requires a running process with known ports // In a real test, we would start a test server let ports = detect_listening_ports(std::process::id()); assert!(ports.is_ok()); } #[test] fn test_check_lsof_availability() { // This test checks the lsof availability check function // The actual result depends on whether lsof is installed on the system let result = check_lsof_availability(); // We can't guarantee it's installed, so we just check the function works assert!(result.is_ok() || result.is_err()); // Test the cached version let is_available = is_lsof_available(); // The cached result should match the check result assert_eq!(is_available, result.is_ok()); } #[test] fn test_parse_lsof_machine_format() { // Test parsing of lsof -F format output let sample_output = "p1234\nfcwd\ntCWD\nn/home/user\n\ f6\ntREG\na r\ni123456\nn/usr/bin/app\n\ f10\ntIPv4\nPTCP\nn*:8080\nTST=LISTEN\n"; let ports = parse_lsof_machine_format(sample_output); assert_eq!(ports.len(), 1); assert_eq!(ports[0].protocol, "tcp"); assert_eq!(ports[0].local_addr, "*:8080"); assert_eq!(ports[0].state, "LISTEN"); } #[test] fn test_parse_malformed_lsof_output() { // Test with completely empty output let empty_output = ""; let ports = parse_lsof_machine_format(empty_output); assert_eq!(ports.len(), 0); // Test with malformed lines (missing tag characters) let malformed_output = "1234\nf10\nt\nn*:8080\n"; let ports = parse_lsof_machine_format(malformed_output); assert_eq!(ports.len(), 0); // Test with truncated output (missing required fields) let truncated_output = "p1234\nf10\ntIPv4\n"; let ports = parse_lsof_machine_format(truncated_output); assert_eq!(ports.len(), 0); // Test with wrong tag order let wrong_order_output = "TST=LISTEN\nPTCP\ntIPv4\nn*:8080\nf10\n"; let ports = parse_lsof_machine_format(wrong_order_output); assert_eq!(ports.len(), 0); // Test with corrupted protocol field let corrupted_protocol = "f10\ntIPv4\nP\nn*:8080\nTST=LISTEN\n"; let ports = parse_lsof_machine_format(corrupted_protocol); assert_eq!(ports.len(), 0); // Test with corrupted address field let corrupted_address = "f10\ntIPv4\nPTCP\nn\nTST=LISTEN\n"; let ports = parse_lsof_machine_format(corrupted_address); assert_eq!(ports.len(), 0); // Test with non-LISTEN state let non_listen_output = "f10\ntIPv4\nPTCP\nn*:8080\nTST=ESTABLISHED\n"; let ports = parse_lsof_machine_format(non_listen_output); assert_eq!(ports.len(), 0); // Test with mixed valid and invalid entries let mixed_output = "p1234\n\ f10\ntIPv4\nPTCP\nn*:8080\nTST=LISTEN\n\ f11\ntIPv4\n\nn*:9000\nTST=LISTEN\n\ f12\ntIPv4\nPUDP\nn*:7000\nTST=LISTEN\n"; let ports = parse_lsof_machine_format(mixed_output); assert_eq!(ports.len(), 2); // Should only get the valid TCP and UDP entries assert!( ports .iter() .any(|p| p.local_addr == "*:8080" && p.protocol == "tcp") ); assert!( ports .iter() .any(|p| p.local_addr == "*:7000" && p.protocol == "udp") ); } #[test] fn test_parse_partial_lsof_output() { // Test with incomplete network file descriptor information let partial_fd_output = "f10\ntIPv4\nPTCP\n"; // Missing address and state let ports = parse_lsof_machine_format(partial_fd_output); assert_eq!(ports.len(), 0); // Test with missing protocol let missing_protocol_output = "f10\ntIPv4\nn*:8080\nTST=LISTEN\n"; let ports = parse_lsof_machine_format(missing_protocol_output); assert_eq!(ports.len(), 0); // Test with incomplete state information let incomplete_state_output = "f10\ntIPv4\nPTCP\nn*:8080\nT\n"; let ports = parse_lsof_machine_format(incomplete_state_output); assert_eq!(ports.len(), 0); // Test with IPv6 addresses let ipv6_output = "f10\ntIPv6\nPTCP\nn[::1]:8080\nTST=LISTEN\n"; let ports = parse_lsof_machine_format(ipv6_output); assert_eq!(ports.len(), 1); assert_eq!(ports[0].local_addr, "[::1]:8080"); assert_eq!(ports[0].protocol, "tcp"); // Test with multiple protocols let multi_protocol_output = "p1234\n\ f10\ntIPv4\nPTCP\nn*:8080\nTST=LISTEN\n\ f11\ntIPv4\nPUDP\nn*:9000\nTST=LISTEN\n\ f12\ntIPv6\nPTCP\nn[::1]:7000\nTST=LISTEN\n"; let ports = parse_lsof_machine_format(multi_protocol_output); assert_eq!(ports.len(), 3); let tcp_v4 = ports.iter().find(|p| p.local_addr == "*:8080").unwrap(); assert_eq!(tcp_v4.protocol, "tcp"); let udp_v4 = ports.iter().find(|p| p.local_addr == "*:9000").unwrap(); assert_eq!(udp_v4.protocol, "udp"); let tcp_v6 = ports.iter().find(|p| p.local_addr == "[::1]:7000").unwrap(); assert_eq!(tcp_v6.protocol, "tcp"); } #[test] fn test_detect_ports_error_scenarios() { // Test with invalid PID (should return empty result, not error) let result = detect_listening_ports(999999); assert!(result.is_ok()); assert_eq!(result.unwrap().len(), 0); // Test with PID 0 (system process) let result = detect_listening_ports(0); assert!(result.is_ok()); // Should not crash, may or may not have ports // Test with current process PID (should work) let current_pid = std::process::id(); let result = detect_listening_ports(current_pid); assert!(result.is_ok()); // Should not error even if no ports } #[cfg(unix)] #[test] fn test_lsof_permission_scenarios() { // Test that we handle lsof permission errors gracefully // This tests the error handling path when lsof exits with non-zero status // Try to get ports for a process that likely doesn't exist or we can't access let result = detect_listening_ports(1); // init process, often not accessible // The function should not panic and should return a valid result assert!(result.is_ok()); // The result might be empty (no ports) or contain ports if we have permission } #[test] fn test_edge_case_addresses() { // Test with various address formats let edge_case_output = "p1234\n\ f10\ntIPv4\nPTCP\nn127.0.0.1:8080\nTST=LISTEN\n\ f11\ntIPv4\nPTCP\nn0.0.0.0:9000\nTST=LISTEN\n\ f12\ntIPv6\nPTCP\nn[::]:7000\nTST=LISTEN\n\ f13\ntIPv4\nPUDP\nn*:5353\nTST=LISTEN\n"; let ports = parse_lsof_machine_format(edge_case_output); assert_eq!(ports.len(), 4); // Verify all addresses are preserved correctly let addresses: Vec<&str> = ports.iter().map(|p| p.local_addr.as_str()).collect(); assert!(addresses.contains(&"127.0.0.1:8080")); assert!(addresses.contains(&"0.0.0.0:9000")); assert!(addresses.contains(&"[::]:7000")); assert!(addresses.contains(&"*:5353")); } #[test] fn test_protocol_case_handling() { // Test that protocol names are properly normalized to lowercase let mixed_case_output = "f10\ntIPv4\nPTCP\nn*:8080\nTST=LISTEN\n\ f11\ntIPv4\nPUDP\nn*:9000\nTST=LISTEN\n"; let ports = parse_lsof_machine_format(mixed_case_output); assert_eq!(ports.len(), 2); for port in &ports { // All protocols should be lowercase assert!(port.protocol.chars().all(|c| c.is_lowercase())); } } }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/process.rs
src/app/process.rs
use nix::sys::signal::{self, Signal}; use nix::unistd::Pid; use nix::unistd::setsid; use std::fs::File; use std::os::unix::process::CommandExt as _; use std::path::PathBuf; use std::process::{Child, Command, Stdio}; use uuid::Uuid; #[derive(Debug, Clone)] pub struct ProcessInfo { pub id: String, pub pid: u32, pub pgid: i32, pub command: Vec<String>, pub log_path: PathBuf, pub env: Vec<(String, String)>, } use crate::app::error::{GhostError, Result}; /// Spawn a background process with logging /// Returns both ProcessInfo and Child handle to allow proper cleanup pub fn spawn_background_process( command: Vec<String>, cwd: Option<PathBuf>, log_dir: Option<PathBuf>, ) -> Result<(ProcessInfo, Child)> { spawn_background_process_with_env(command, cwd, log_dir, Vec::new()) } /// Spawn a background process with logging and custom environment variables /// Returns both ProcessInfo and Child handle to allow proper cleanup pub fn spawn_background_process_with_env( command: Vec<String>, cwd: Option<PathBuf>, log_dir: Option<PathBuf>, custom_env: Vec<(String, String)>, ) -> Result<(ProcessInfo, Child)> { // Generate task ID and prepare paths let task_id = Uuid::new_v4().to_string(); let log_dir = log_dir.unwrap_or_else(crate::app::config::get_log_dir); // Create log directory if it doesn't exist std::fs::create_dir_all(&log_dir)?; let log_path = log_dir.join(format!("{task_id}.log")); // Create log file let log_file = File::create(&log_path).map_err(|e| GhostError::LogFileCreation { path: log_path.to_string_lossy().to_string(), source: e, })?; // Setup command let mut cmd = Command::new(&command[0]); cmd.args(&command[1..]) .stdin(Stdio::null()) .stdout(Stdio::from(log_file.try_clone()?)) .stderr(Stdio::from(log_file)); // Set current working directory if specified if let Some(ref cwd) = cwd { cmd.current_dir(cwd); } // Collect all environment variables (inherited + custom) let mut all_env: Vec<(String, String)> = std::env::vars().collect(); // Add custom environment variables for (key, value) in &custom_env { cmd.env(key, value); // Update or add to all_env if let Some(pos) = all_env.iter().position(|(k, _)| k == key) { all_env[pos] = (key.clone(), value.clone()); } else { all_env.push((key.clone(), value.clone())); } } unsafe { cmd.pre_exec(|| { setsid()?; Ok(()) }); } // Spawn the process let child = cmd.spawn().map_err(|e| GhostError::ProcessSpawn { message: format!("Failed to spawn process: {e}"), })?; let pid = child.id(); // The process group ID should be the same as PID after setsid() let pgid = pid as i32; let info = ProcessInfo { id: task_id, pid, pgid, command, log_path, env: all_env, }; Ok((info, child)) } /// Check if a process is still running pub fn exists(pid: u32) -> bool { { // Send signal 0 to check if process exists // We need to check errno to distinguish between "no permission" and "no process" match signal::kill(Pid::from_raw(pid as i32), None) { Ok(_) => true, Err(nix::errno::Errno::ESRCH) => false, // No such process Err(_) => true, // Other errors (e.g., EPERM) mean process exists } } } /// Kill a process pub fn kill(pid: u32, force: bool) -> Result<()> { { let signal = if force { Signal::SIGKILL } else { Signal::SIGTERM }; signal::kill(Pid::from_raw(pid as i32), signal)?; Ok(()) } } /// Kill a process group pub fn kill_group(pgid: i32, force: bool) -> Result<()> { let signal = if force { Signal::SIGKILL } else { Signal::SIGTERM }; // Negative PID means process group signal::kill(Pid::from_raw(-pgid), signal)?; Ok(()) } #[cfg(test)] mod tests { use super::*; use std::thread; use std::time::Duration; #[test] fn test_spawn_background_process() { let temp_dir = tempfile::tempdir().unwrap(); let log_dir = temp_dir.path().to_path_buf(); // Spawn a simple sleep command let command = vec!["sleep".to_string(), "2".to_string()]; let result = spawn_background_process(command.clone(), None, Some(log_dir.clone())); assert!(result.is_ok()); let (process_info, mut child) = result.unwrap(); // Check process info assert!(!process_info.id.is_empty()); assert!(process_info.pid > 0); assert_eq!(process_info.command, command); assert!(!process_info.env.is_empty()); // Should have inherited environment // Check log file exists assert!(process_info.log_path.exists()); // Check process is running assert!(exists(process_info.pid)); // Wait a bit and check again thread::sleep(Duration::from_millis(100)); assert!(exists(process_info.pid)); // Kill the process let _ = kill(process_info.pid, true); // Clean up zombie by waiting let _ = child.wait(); } #[test] fn test_spawn_with_output() { let temp_dir = tempfile::tempdir().unwrap(); let log_dir = temp_dir.path().to_path_buf(); // Spawn echo command let command = vec![ "sh".to_string(), "-c".to_string(), "echo 'Hello, Ghost!' && echo 'Error message' >&2".to_string(), ]; let result = spawn_background_process(command, None, Some(log_dir)); assert!(result.is_ok()); let (process_info, mut child) = result.unwrap(); // Wait for process to complete let _ = child.wait(); // Check log content let log_content = std::fs::read_to_string(&process_info.log_path).unwrap(); assert!(log_content.contains("Hello, Ghost!")); assert!(log_content.contains("Error message")); } #[test] fn test_spawn_with_cwd() { let temp_dir = tempfile::tempdir().unwrap(); let log_dir = temp_dir.path().to_path_buf(); let test_cwd = temp_dir.path().join("test_cwd"); // Create test directory and file std::fs::create_dir_all(&test_cwd).unwrap(); let test_file = test_cwd.join("test_file.txt"); std::fs::write(&test_file, "test content").unwrap(); // Spawn command that reads the file in the specified cwd let command = vec!["cat".to_string(), "test_file.txt".to_string()]; let result = spawn_background_process(command, Some(test_cwd), Some(log_dir)); assert!(result.is_ok()); let (process_info, mut child) = result.unwrap(); // Wait for process to complete let _ = child.wait(); // Check log content to verify cwd was used let log_content = std::fs::read_to_string(&process_info.log_path).unwrap(); assert!(log_content.contains("test content")); } #[test] fn test_kill_process_force() { // Test that SIGKILL works even when SIGTERM is trapped let temp_dir = tempfile::tempdir().unwrap(); let log_dir = temp_dir.path().to_path_buf(); // Create a script that ignores SIGTERM let script_content = "#!/bin/sh\ntrap '' TERM\nwhile true; do sleep 1; done"; let script_path = temp_dir.path().join("ignore_term.sh"); std::fs::write(&script_path, script_content).unwrap(); use std::os::unix::fs::PermissionsExt; let mut perms = std::fs::metadata(&script_path).unwrap().permissions(); perms.set_mode(0o755); std::fs::set_permissions(&script_path, perms).unwrap(); let command = vec![script_path.to_string_lossy().to_string()]; let result = spawn_background_process(command, None, Some(log_dir)); assert!(result.is_ok()); let (process_info, mut child) = result.unwrap(); let pid = process_info.pid; // Verify process is running assert!(exists(pid)); // Try to kill with SIGTERM (should not work due to trap) let _ = kill(pid, false); thread::sleep(Duration::from_millis(200)); // Force kill with SIGKILL - this should always work let kill_result = kill(pid, true); assert!(kill_result.is_ok()); // Wait to reap the zombie let _ = child.wait(); assert!(!exists(pid)); } #[test] fn test_spawn_with_custom_env() { let temp_dir = tempfile::tempdir().unwrap(); let log_dir = temp_dir.path().to_path_buf(); // Spawn command with custom environment variable let command = vec![ "sh".to_string(), "-c".to_string(), "echo $TEST_CUSTOM_VAR".to_string(), ]; let custom_env = vec![("TEST_CUSTOM_VAR".to_string(), "Hello Ghost!".to_string())]; let result = spawn_background_process_with_env( command.clone(), None, Some(log_dir), custom_env.clone(), ); assert!(result.is_ok()); let (process_info, mut child) = result.unwrap(); // Check that custom env var is in the process info let has_custom_var = process_info .env .iter() .any(|(k, v)| k == "TEST_CUSTOM_VAR" && v == "Hello Ghost!"); assert!(has_custom_var); // Wait for process to complete let _ = child.wait(); // Check log content to verify env var was used let log_content = std::fs::read_to_string(&process_info.log_path).unwrap(); assert!(log_content.contains("Hello Ghost!")); } }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/storage.rs
src/app/storage.rs
pub mod cleanup; pub mod database; pub mod task; pub mod task_repository; pub mod task_status; // Re-export for backward compatibility pub use cleanup::{cleanup_old_tasks, cleanup_tasks_by_criteria, get_cleanup_candidates}; pub use database::{init_database, init_database_with_config}; pub use task::Task; pub use task_repository::{ delete_task, get_task, get_tasks, get_tasks_with_process_check, insert_task, row_to_task, update_task_status, update_task_status_by_process_check, }; pub use task_status::TaskStatus;
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/error.rs
src/app/error.rs
#[derive(Debug, thiserror::Error)] pub enum GhostError { // Process-related errors #[error("Process spawn failed: {message}")] ProcessSpawn { message: String }, #[error("Process operation failed: {message}")] ProcessOperation { message: String }, #[error("Log file creation failed: {path} - {source}")] LogFileCreation { path: String, #[source] source: std::io::Error, }, // Storage-related errors #[error("Database error: {source}")] Database { #[from] source: rusqlite::Error, }, #[error("Data serialization error: {source}")] Serialization { #[from] source: serde_json::Error, }, // File system errors #[error("File operation failed: {source}")] Io { #[from] source: std::io::Error, }, // Task management errors #[error("Task not found: {task_id}")] TaskNotFound { task_id: String }, #[error("Task operation failed: {task_id} - {message}")] TaskOperation { task_id: String, message: String }, // Configuration errors #[error("Configuration error: {message}")] Config { message: String }, // Input validation errors #[error("Invalid argument: {message}")] InvalidArgument { message: String }, // Command not found errors #[error("Command not found: {command}")] CommandNotFound { command: String }, // System-level errors #[error("Unix system error: {source}")] Unix { #[from] source: nix::Error, }, // File watching errors (for log following) #[error("File watching error: {message}")] FileWatch { message: String }, } pub type Result<T> = std::result::Result<T, GhostError>;
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/display.rs
src/app/display.rs
use crate::app::storage::Task; /// Display a list of tasks in a formatted table pub fn print_task_list(tasks: &[Task]) { if tasks.is_empty() { println!("No tasks found."); return; } print_table_header(); for task in tasks { let command_display = format_command_truncated(&task.command, 30); let started = format_timestamp(task.started_at, "%Y-%m-%d %H:%M"); let cwd_display = task.cwd.as_deref().unwrap_or("-"); println!( "{:<36} {:<8} {:<10} {:<20} {:<30} {}", &task.id, task.pid, task.status.as_str(), started, command_display, cwd_display ); } } /// Print the table header for task list fn print_table_header() { println!( "{:<36} {:<8} {:<10} {:<20} {:<30} Directory", "Task ID", "PID", "Status", "Started", "Command" ); println!("{}", "-".repeat(134)); } /// Display detailed information about a single task pub fn print_task_details(task: &Task) { let task_id = &task.id; println!("Task: {task_id}"); let pid = task.pid; println!("PID: {pid}"); let status = &task.status; println!("Status: {status}"); let command = format_command_full(&task.command); println!("Command: {command}"); if let Some(ref cwd) = task.cwd { println!("Working directory: {cwd}"); } println!( "Started: {}", format_timestamp(task.started_at, "%Y-%m-%d %H:%M:%S") ); if let Some(finished_at) = task.finished_at { println!( "Finished: {}", format_timestamp(finished_at, "%Y-%m-%d %H:%M:%S") ); } if let Some(exit_code) = task.exit_code { println!("Exit code: {exit_code}"); } let log_path = &task.log_path; println!("Log file: {log_path}"); } /// Display information about a started process pub fn print_process_started(task_id: &str, pid: u32, log_path: &std::path::Path) { println!("Started background process:"); println!(" Task ID: {task_id}"); println!(" PID: {pid}"); println!(" Log file: {}", log_path.display()); } /// Display log follow header pub fn print_log_follow_header(task_id: &str, log_path: &str) { println!("Following logs for task {task_id} (Ctrl+C to stop):"); println!("Log file: {log_path}"); println!("{}", "-".repeat(40)); } // Helper functions for formatting /// Format a command JSON string for display with truncation fn format_command_truncated(command_json: &str, max_length: usize) -> String { let command_str = format_command_full(command_json); truncate_string(&command_str, max_length) } /// Format a command JSON string for full display fn format_command_full(command_json: &str) -> String { let command: Vec<String> = serde_json::from_str(command_json).unwrap_or_default(); command.join(" ") } /// Format a timestamp to a human-readable string fn format_timestamp(timestamp: i64, format_str: &str) -> String { chrono::DateTime::from_timestamp(timestamp, 0) .map(|dt| dt.format(format_str).to_string()) .unwrap_or_else(|| "Unknown".to_string()) } /// Truncate a string to the specified length with ellipsis fn truncate_string(s: &str, max_length: usize) -> String { if s.len() > max_length { if max_length >= 3 { format!("{truncated}...", truncated = &s[..max_length - 3]) } else { s[..max_length].to_string() } } else { s.to_string() } }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/commands.rs
src/app/commands.rs
use std::path::PathBuf; use crate::app::{config, display, error, error::Result, helpers, process, storage}; use rusqlite::Connection; /// Run a command in the background pub fn spawn( conn: &Connection, command: Vec<String>, cwd: Option<PathBuf>, env: Vec<String>, show_output: bool, ) -> Result<process::ProcessInfo> { if command.is_empty() { return Err(error::GhostError::InvalidArgument { message: "No command specified".to_string(), }); } let env_vars = config::env::parse_env_vars(&env)?; let (process_info, _) = spawn_and_register_process(command, cwd, env_vars, conn)?; if show_output { display::print_process_started(&process_info.id, process_info.pid, &process_info.log_path); } Ok(process_info) } /// Spawn process and register it in the database pub fn spawn_and_register_process( command: Vec<String>, cwd: Option<PathBuf>, env_vars: Vec<(String, String)>, conn: &Connection, ) -> Result<(process::ProcessInfo, std::process::Child)> { // If no cwd is specified, use the current directory let effective_cwd = match cwd { Some(path) => Some(path), None => std::env::current_dir().ok(), }; let (process_info, child) = process::spawn_background_process_with_env( command.clone(), effective_cwd.clone(), None, env_vars, )?; // Save to database with the actual environment variables from the process let env = if process_info.env.is_empty() { None } else { Some(process_info.env.as_slice()) }; storage::insert_task( conn, &process_info.id, process_info.pid, Some(process_info.pgid), &command, env, effective_cwd.as_deref(), &process_info.log_path, )?; Ok((process_info, child)) } /// List all background processes pub fn list( conn: &Connection, status_filter: Option<String>, show_output: bool, ) -> Result<Vec<storage::task::Task>> { let tasks = storage::get_tasks_with_process_check(conn, status_filter.as_deref())?; if show_output { display::print_task_list(&tasks); } Ok(tasks) } /// Show logs for a process pub async fn log( conn: &Connection, task_id: &str, follow: bool, show_output: bool, ) -> Result<String> { let task = storage::get_task(conn, task_id)?; let log_path = PathBuf::from(&task.log_path); let content = std::fs::read_to_string(&log_path).map_err(|e| error::GhostError::InvalidArgument { message: format!("Failed to read log file: {e}"), })?; if show_output { if follow { display::print_log_follow_header(task_id, &task.log_path); helpers::follow_log_file(&log_path).await?; } else { print!("{content}"); } } Ok(content) } /// Stop a background process pub fn stop(conn: &Connection, task_id: &str, force: bool, show_output: bool) -> Result<()> { let task = storage::get_task(conn, task_id)?; helpers::validate_task_running(&task)?; // Kill the process group if available, otherwise kill individual process if let Some(pgid) = task.pgid { process::kill_group(pgid, force)?; } else { process::kill(task.pid, force)?; } // Update status in database let status = if force { storage::TaskStatus::Killed } else { storage::TaskStatus::Exited }; storage::update_task_status(conn, task_id, status, None)?; if show_output { let pid = task.pid; println!("Process {task_id} ({pid}) has been {status}"); } Ok(()) } /// Check status of a background process pub fn status(conn: &Connection, task_id: &str, show_output: bool) -> Result<storage::task::Task> { // This will update the status if the process is no longer running let task = storage::update_task_status_by_process_check(conn, task_id)?; if show_output { display::print_task_details(&task); } Ok(task) } /// Clean up old finished tasks pub fn cleanup( conn: &Connection, days: u64, status: Option<String>, dry_run: bool, all: bool, ) -> Result<()> { // Parse status filter let status_filter = parse_status_filter(status.as_deref())?; // Determine days filter - None if --all is specified let days_filter = if all { None } else { Some(days) }; if dry_run { // Show what would be deleted let candidates = storage::get_cleanup_candidates(conn, days_filter, &status_filter)?; if candidates.is_empty() { println!("No tasks found matching cleanup criteria."); return Ok(()); } println!( "The following {} task(s) would be deleted:", candidates.len() ); display::print_task_list(&candidates); if all { println!( "\nNote: --all flag specified, all finished tasks would be deleted regardless of age." ); } else { println!("\nNote: Only tasks older than {days} days would be deleted."); } } else { // Actually delete tasks let deleted_count = storage::cleanup_tasks_by_criteria(conn, days_filter, &status_filter)?; if deleted_count == 0 { println!("No tasks found matching cleanup criteria."); } else { println!("Successfully deleted {deleted_count} task(s)."); if all { println!("Deleted all finished tasks regardless of age."); } else { println!( "Deleted tasks older than {} days with status: {}.", days, format_status_list(&status_filter) ); } } } Ok(()) } /// Parse status filter string into TaskStatus enum list fn parse_status_filter(status: Option<&str>) -> Result<Vec<storage::TaskStatus>> { match status { Some("all") => { // All statuses except running (don't delete running tasks) Ok(vec![ storage::TaskStatus::Exited, storage::TaskStatus::Killed, storage::TaskStatus::Unknown, ]) } Some(status_str) => { let statuses: Result<Vec<_>> = status_str .split(',') .map(|s| s.trim()) .map(|s| match s { "exited" => Ok(storage::TaskStatus::Exited), "killed" => Ok(storage::TaskStatus::Killed), "unknown" => Ok(storage::TaskStatus::Unknown), "running" => Err(error::GhostError::InvalidArgument { message: "Cannot cleanup running tasks".to_string(), }), _ => Err(error::GhostError::InvalidArgument { message: format!( "Invalid status: {s}. Valid options: exited, killed, unknown, all" ), }), }) .collect(); statuses } None => { // Default: exited and killed only Ok(vec![ storage::TaskStatus::Exited, storage::TaskStatus::Killed, ]) } } } /// Format status list for display fn format_status_list(statuses: &[storage::TaskStatus]) -> String { statuses .iter() .map(|s| s.as_str()) .collect::<Vec<_>>() .join(", ") } /// Spawn result for a single command in multi-command execution #[derive(Debug)] pub struct SpawnResult { /// The original command string pub command_str: String, /// Result of spawning the process pub result: Result<process::ProcessInfo>, } /// Run multiple commands in parallel /// /// Each command string is parsed and spawned as an independent process. /// Returns results for all commands, even if some fail. pub fn spawn_multi( conn: &Connection, command_strs: Vec<String>, cwd: Option<PathBuf>, env: Vec<String>, show_output: bool, ) -> Vec<SpawnResult> { let env_vars = match config::env::parse_env_vars(&env) { Ok(vars) => vars, Err(e) => { // If env parsing fails, return error for all commands let error_msg = e.to_string(); return command_strs .into_iter() .map(|cmd| SpawnResult { command_str: cmd, result: Err(error::GhostError::InvalidArgument { message: error_msg.clone(), }), }) .collect(); } }; command_strs .into_iter() .map(|command_str| { let result = spawn_single_command(&command_str, cwd.clone(), env_vars.clone(), conn); if show_output { match &result { Ok(info) => { display::print_process_started(&info.id, info.pid, &info.log_path); } Err(e) => { eprintln!("Failed to spawn '{command_str}': {e}"); } } } SpawnResult { command_str, result, } }) .collect() } /// Spawn a single command from a command string fn spawn_single_command( command_str: &str, cwd: Option<PathBuf>, env_vars: Vec<(String, String)>, conn: &Connection, ) -> Result<process::ProcessInfo> { // Parse the command string into command and arguments let command = helpers::parse_command(command_str)?; // Spawn and register the process let (process_info, _) = spawn_and_register_process(command, cwd, env_vars, conn)?; Ok(process_info) } /// Start TUI mode pub async fn tui() -> Result<()> { use crossterm::{ event::{DisableMouseCapture, EnableMouseCapture, Event, EventStream}, execute, terminal::{EnterAlternateScreen, LeaveAlternateScreen, disable_raw_mode, enable_raw_mode}, }; use futures::StreamExt; use ratatui::{Terminal, backend::CrosstermBackend}; use std::io; use tokio::time::{Duration, interval}; use crate::app::tui::app::TuiApp; // Setup terminal enable_raw_mode()?; let mut stdout = io::stdout(); execute!(stdout, EnterAlternateScreen, EnableMouseCapture)?; let backend = CrosstermBackend::new(stdout); let mut terminal = Terminal::new(backend)?; // Create app let mut app = TuiApp::new()?; app.refresh_tasks()?; // Setup refresh interval and event stream let mut refresh_interval = interval(Duration::from_secs(1)); let mut event_stream = EventStream::new(); let result = loop { // Draw the UI terminal.draw(|f| app.render(f))?; // Handle input and refresh tokio::select! { // Handle keyboard events from async stream Some(event_result) = event_stream.next() => { match event_result { Ok(Event::Key(key)) => { if let Err(e) = app.handle_key(key) { break Err(e); } if app.should_quit() { break Ok(()); } } Err(e) => { break Err(error::GhostError::Io { source: e }); } _ => {} // Ignore other events (Mouse, Resize, etc.) } } // Refresh tasks periodically _ = refresh_interval.tick() => { if let Err(e) = app.refresh_tasks() { break Err(e); } } } }; // Restore terminal disable_raw_mode()?; execute!( terminal.backend_mut(), LeaveAlternateScreen, DisableMouseCapture )?; terminal.show_cursor()?; result } #[cfg(test)] mod tests { use super::*; fn setup_test_db() -> Connection { let conn = Connection::open_in_memory().unwrap(); storage::database::init_schema(&conn).unwrap(); conn } #[test] fn test_spawn_multi_two_commands() { let conn = setup_test_db(); let commands = vec!["sleep 1".to_string(), "echo hello".to_string()]; let results = spawn_multi(&conn, commands, None, vec![], false); assert_eq!(results.len(), 2); assert!(results[0].result.is_ok()); assert!(results[1].result.is_ok()); // Verify both tasks are in database let tasks = storage::get_tasks_with_process_check(&conn, None).unwrap(); assert_eq!(tasks.len(), 2); // Clean up: kill spawned processes for result in &results { if let Ok(info) = &result.result { let _ = process::kill(info.pid, true); } } } #[test] fn test_spawn_multi_one_fails() { let conn = setup_test_db(); // First command is valid, second is empty (will fail to parse) let commands = vec!["sleep 1".to_string(), "".to_string()]; let results = spawn_multi(&conn, commands, None, vec![], false); assert_eq!(results.len(), 2); assert!(results[0].result.is_ok()); assert!(results[1].result.is_err()); // Empty command fails // Clean up for result in &results { if let Ok(info) = &result.result { let _ = process::kill(info.pid, true); } } } #[test] fn test_spawn_multi_empty_command_fails() { let conn = setup_test_db(); let commands = vec!["".to_string()]; let results = spawn_multi(&conn, commands, None, vec![], false); assert_eq!(results.len(), 1); assert!(results[0].result.is_err()); } #[test] fn test_spawn_multi_preserves_command_str() { let conn = setup_test_db(); let commands = vec!["sleep 1".to_string(), "echo 'hello world'".to_string()]; let results = spawn_multi(&conn, commands.clone(), None, vec![], false); assert_eq!(results[0].command_str, commands[0]); assert_eq!(results[1].command_str, commands[1]); // Clean up for result in &results { if let Ok(info) = &result.result { let _ = process::kill(info.pid, true); } } } }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/mod.rs
src/app/mod.rs
pub mod commands; pub mod config; pub mod display; pub mod error; pub mod helpers; pub mod logging; pub mod port_detector; pub mod process; pub mod process_state; pub mod storage; pub mod tui;
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/logging.rs
src/app/logging.rs
use std::path::Path; use tracing_appender::non_blocking::WorkerGuard; use tracing_rolling_file::RollingFileAppenderBase; use tracing_subscriber::fmt::format::FmtSpan; /// Initialize file logger with rolling file appender /// /// Creates a rolling file appender that: /// - Writes to `ghost.log` in the specified directory /// - Rotates when file size exceeds 5MB /// - Keeps up to 5 historical log files /// /// # Arguments /// * `log_dir` - Directory where log files will be stored /// /// # Returns /// * `Some(WorkerGuard)` - Guard that must be held to ensure logs are flushed /// * `None` - If logger initialization failed pub fn init_file_logger(log_dir: &Path) -> Option<WorkerGuard> { // Create parent directory if it doesn't exist if let Err(e) = std::fs::create_dir_all(log_dir) { eprintln!("Warning: Failed to create log directory: {e}"); return None; } let log_file_path = log_dir.join("ghost.log"); // Create rolling file appender with 5MB size limit let file_appender = RollingFileAppenderBase::builder() .filename(log_file_path.to_string_lossy().to_string()) .max_filecount(5) .condition_max_file_size(5 * 1024 * 1024) // 5MB .build() .ok()?; // Get non-blocking writer let (non_blocking, guard) = file_appender.get_non_blocking_appender(); // Initialize tracing subscriber tracing_subscriber::fmt() .with_writer(non_blocking) .with_ansi(false) .with_span_events(FmtSpan::NONE) .with_target(false) .init(); Some(guard) } #[cfg(test)] mod tests { use tempfile::tempdir; use tracing_rolling_file::RollingFileAppenderBase; #[test] fn test_rolling_file_appender_creates_log_file() { let temp_dir = tempdir().unwrap(); let log_dir = temp_dir.path(); let log_file_path = log_dir.join("test.log"); // Create rolling file appender let file_appender = RollingFileAppenderBase::builder() .filename(log_file_path.to_string_lossy().to_string()) .max_filecount(5) .condition_max_file_size(5 * 1024 * 1024) .build(); assert!(file_appender.is_ok(), "File appender should be created"); // Write something to trigger file creation use std::io::Write; let mut appender = file_appender.unwrap(); writeln!(appender, "test log message").unwrap(); // Log file should be created assert!(log_file_path.exists(), "test.log should be created"); } #[test] fn test_create_log_directory() { let temp_dir = tempdir().unwrap(); let log_dir = temp_dir.path().join("nested").join("logs"); // Directory should not exist yet assert!(!log_dir.exists()); // Create directory std::fs::create_dir_all(&log_dir).unwrap(); // Directory should be created assert!(log_dir.exists(), "Log directory should be created"); } }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/process_state.rs
src/app/process_state.rs
use crate::app::{ process, storage::{Task, TaskStatus}, }; /// Check and update the status of a single task based on process existence pub fn update_task_status_if_needed(task: &mut Task) -> bool { if task.status == TaskStatus::Running && !process::exists(task.pid) { task.status = TaskStatus::Exited; task.finished_at = Some(crate::app::helpers::now_timestamp()); true // Status was updated } else { false // Status was not updated } } /// Determine task status based on process state pub fn determine_task_status(pid: u32) -> TaskStatus { if process::exists(pid) { TaskStatus::Running } else { TaskStatus::Exited } } #[cfg(test)] mod tests { use super::*; use crate::app::storage::Task; #[test] fn test_update_task_status_if_needed_running_nonexistent() { let mut task = Task { id: "test".to_string(), pid: 99999, // Non-existent PID pgid: None, command: "[]".to_string(), env: None, cwd: None, status: TaskStatus::Running, exit_code: None, started_at: 0, finished_at: None, log_path: "/tmp/test.log".to_string(), }; let updated = update_task_status_if_needed(&mut task); assert!(updated); assert_eq!(task.status, TaskStatus::Exited); assert!(task.finished_at.is_some()); } #[test] fn test_update_task_status_if_needed_already_exited() { let mut task = Task { id: "test".to_string(), pid: 1, // Likely existing PID pgid: None, command: "[]".to_string(), env: None, cwd: None, status: TaskStatus::Exited, exit_code: None, started_at: 0, finished_at: None, log_path: "/tmp/test.log".to_string(), }; let updated = update_task_status_if_needed(&mut task); assert!(!updated); assert_eq!(task.status, TaskStatus::Exited); } }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/storage/task_status.rs
src/app/storage/task_status.rs
use serde::{Deserialize, Serialize}; #[derive(Debug, Clone, Copy, PartialEq, Eq, Serialize, Deserialize)] pub enum TaskStatus { Running, Exited, Killed, Unknown, } impl std::fmt::Display for TaskStatus { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { write!(f, "{}", self.as_str()) } } impl TaskStatus { /// Convert TaskStatus to string for database storage pub fn as_str(&self) -> &'static str { match self { TaskStatus::Running => "running", TaskStatus::Exited => "exited", TaskStatus::Killed => "killed", TaskStatus::Unknown => "unknown", } } /// Parse TaskStatus from string (for database retrieval) #[allow(clippy::should_implement_trait)] pub fn from_str(s: &str) -> TaskStatus { s.parse().unwrap_or(TaskStatus::Unknown) } } impl std::str::FromStr for TaskStatus { type Err = String; fn from_str(s: &str) -> std::result::Result<Self, Self::Err> { match s { "running" => Ok(TaskStatus::Running), "exited" => Ok(TaskStatus::Exited), "killed" => Ok(TaskStatus::Killed), "unknown" => Ok(TaskStatus::Unknown), _ => Err(format!("Unknown task status: {s}")), } } }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/storage/task_repository.rs
src/app/storage/task_repository.rs
use std::path::Path; use rusqlite::{Connection, Result as SqliteResult, Row}; use super::task::Task; use super::task_status::TaskStatus; use crate::app::error::Result; use crate::app::process_state; /// Insert a new task into the database #[allow(clippy::too_many_arguments)] pub fn insert_task( conn: &Connection, id: &str, pid: u32, pgid: Option<i32>, command: &[String], env: Option<&[(String, String)]>, cwd: Option<&Path>, log_path: &Path, ) -> Result<()> { let command_json = serde_json::to_string(command)?; let env_json = env.map(serde_json::to_string).transpose()?; let cwd_str = cwd.map(|p| p.to_string_lossy().to_string()); let started_at = crate::app::helpers::now_timestamp(); conn.execute( r#" INSERT INTO tasks ( id, pid, pgid, command, env, cwd, status, started_at, log_path ) VALUES (?1, ?2, ?3, ?4, ?5, ?6, 'running', ?7, ?8) "#, ( id, pid as i64, pgid.map(|p| p as i64), command_json, env_json, cwd_str, started_at, log_path.to_string_lossy(), ), )?; Ok(()) } /// Get a task by ID pub fn get_task(conn: &Connection, task_id: &str) -> Result<Task> { let mut stmt = conn.prepare( "SELECT id, pid, pgid, command, env, cwd, status, exit_code, started_at, finished_at, log_path FROM tasks WHERE id = ?1" )?; let task = stmt .query_row([task_id], row_to_task) .map_err(|e| match e { rusqlite::Error::QueryReturnedNoRows => crate::app::error::GhostError::TaskNotFound { task_id: task_id.to_string(), }, _ => e.into(), })?; Ok(task) } /// Get all tasks, optionally filtered by status pub fn get_tasks(conn: &Connection, status_filter: Option<&str>) -> Result<Vec<Task>> { let base_sql = "SELECT id, pid, pgid, command, env, cwd, status, exit_code, started_at, finished_at, log_path FROM tasks"; let order_clause = " ORDER BY started_at DESC"; let sql = match status_filter { Some(_) => format!("{base_sql} WHERE status = ?1{order_clause}"), None => format!("{base_sql}{order_clause}"), }; let mut stmt = conn.prepare(&sql)?; let task_iter = match status_filter { Some(status) => stmt.query_map([status], row_to_task)?, None => stmt.query_map([], row_to_task)?, }; let mut tasks = Vec::new(); for task in task_iter { tasks.push(task?); } Ok(tasks) } /// Get all tasks with process status checking pub fn get_tasks_with_process_check( conn: &Connection, status_filter: Option<&str>, ) -> Result<Vec<Task>> { let mut tasks = get_tasks(conn, status_filter)?; // Update status for running tasks for task in &mut tasks { if task.status == TaskStatus::Running && let Ok(updated_task) = update_task_status_by_process_check(conn, &task.id) { *task = updated_task; } } Ok(tasks) } /// Update task status pub fn update_task_status( conn: &Connection, task_id: &str, new_status: TaskStatus, exit_code: Option<i32>, ) -> Result<()> { let finished_at = if matches!(new_status, TaskStatus::Running) { None } else { Some(crate::app::helpers::now_timestamp()) }; conn.execute( "UPDATE tasks SET status = ?1, exit_code = ?2, finished_at = ?3 WHERE id = ?4", (new_status.as_str(), exit_code, finished_at, task_id), )?; Ok(()) } /// Update task status by checking if the process is still running pub fn update_task_status_by_process_check(conn: &Connection, task_id: &str) -> Result<Task> { let task = get_task(conn, task_id)?; if task.status == TaskStatus::Running { let new_status = process_state::determine_task_status(task.pid); update_task_status(conn, task_id, new_status, None)?; // Return updated task get_task(conn, task_id) } else { Ok(task) } } /// Delete a task by ID pub fn delete_task(conn: &Connection, task_id: &str) -> Result<()> { let rows_affected = conn.execute("DELETE FROM tasks WHERE id = ?1", [task_id])?; if rows_affected == 0 { return Err(crate::app::error::GhostError::TaskNotFound { task_id: task_id.to_string(), }); } Ok(()) } /// Helper function to convert a row to a Task pub fn row_to_task(row: &Row) -> SqliteResult<Task> { Ok(Task { id: row.get(0)?, pid: row.get::<_, i64>(1)? as u32, pgid: row.get::<_, Option<i64>>(2)?.map(|p| p as i32), command: row.get(3)?, env: row.get(4)?, cwd: row.get(5)?, status: TaskStatus::from_str(&row.get::<_, String>(6)?), exit_code: row.get::<_, Option<i64>>(7)?.map(|c| c as i32), started_at: row.get(8)?, finished_at: row.get(9)?, log_path: row.get(10)?, }) }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/storage/database.rs
src/app/storage/database.rs
use crate::app::error::Result; use rusqlite::Connection; /// Initialize schema on an existing connection (for testing) pub(crate) fn init_schema(conn: &Connection) -> Result<()> { // Create tasks table conn.execute( r#" CREATE TABLE IF NOT EXISTS tasks ( id TEXT PRIMARY KEY, pid INTEGER NOT NULL, pgid INTEGER, command TEXT NOT NULL, env TEXT, cwd TEXT, status TEXT NOT NULL DEFAULT 'running', exit_code INTEGER, started_at INTEGER NOT NULL, finished_at INTEGER, log_path TEXT NOT NULL ) "#, [], )?; // Create indexes for performance conn.execute( "CREATE INDEX IF NOT EXISTS idx_tasks_status ON tasks(status)", [], )?; conn.execute("CREATE INDEX IF NOT EXISTS idx_tasks_pid ON tasks(pid)", [])?; conn.execute( "CREATE INDEX IF NOT EXISTS idx_tasks_started_at ON tasks(started_at)", [], )?; Ok(()) } /// Initialize the database and create tables if they don't exist pub fn init_database() -> Result<Connection> { init_database_with_config(None) } /// Initialize the database with a specific config pub fn init_database_with_config(config: Option<crate::app::config::Config>) -> Result<Connection> { let config = config.unwrap_or_default(); config.ensure_directories()?; let db_path = config.get_db_path(); let conn = Connection::open(db_path)?; // Enable WAL mode for better concurrency conn.pragma_update(None, "journal_mode", "WAL")?; conn.pragma_update(None, "synchronous", "NORMAL")?; init_schema(&conn)?; Ok(conn) }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/storage/cleanup.rs
src/app/storage/cleanup.rs
use rusqlite::Connection; use super::task::Task; use super::task_repository::{row_to_task, update_task_status_by_process_check}; use super::task_status::TaskStatus; use crate::app::error::Result; /// Clean up old tasks (legacy function) pub fn cleanup_old_tasks(conn: &Connection, days: u64) -> Result<usize> { let cutoff_time = crate::app::helpers::now_timestamp() - (days * 24 * 60 * 60) as i64; let rows_affected = conn.execute( "DELETE FROM tasks WHERE status IN ('exited', 'killed') AND finished_at IS NOT NULL AND finished_at < ?1", [cutoff_time], )?; Ok(rows_affected) } /// Get tasks that would be cleaned up (for dry-run) pub fn get_cleanup_candidates( conn: &Connection, days: Option<u64>, status_filter: &[TaskStatus], ) -> Result<Vec<Task>> { // First, update status for all running tasks let running_sql = "SELECT id FROM tasks WHERE status = 'running'"; let mut running_stmt = conn.prepare(running_sql)?; let running_ids: Vec<String> = running_stmt .query_map([], |row| row.get(0))? .collect::<std::result::Result<Vec<_>, _>>()?; // Update status for each running task for task_id in running_ids { update_task_status_by_process_check(conn, &task_id)?; } // Now get cleanup candidates with filters applied let mut sql = "SELECT id, pid, pgid, command, env, cwd, status, exit_code, started_at, finished_at, log_path FROM tasks".to_string(); let mut params: Vec<Box<dyn rusqlite::ToSql + '_>> = Vec::new(); let mut conditions = Vec::new(); // Add status filter if !status_filter.is_empty() { let status_placeholders = status_filter .iter() .map(|_| "?") .collect::<Vec<_>>() .join(","); conditions.push(format!("status IN ({status_placeholders})")); for status in status_filter { params.push(Box::new(status.as_str())); } } // Add time filter if specified if let Some(days) = days { let cutoff_time = crate::app::helpers::now_timestamp() - (days * 24 * 60 * 60) as i64; conditions.push("finished_at IS NOT NULL AND finished_at < ?".to_string()); params.push(Box::new(cutoff_time)); } // Build WHERE clause if we have conditions if !conditions.is_empty() { sql.push_str(" WHERE "); sql.push_str(&conditions.join(" AND ")); } sql.push_str(" ORDER BY finished_at DESC"); let mut stmt = conn.prepare(&sql)?; let param_refs: Vec<&dyn rusqlite::ToSql> = params.iter().map(|p| p.as_ref()).collect(); let task_iter = stmt.query_map(&param_refs[..], row_to_task)?; let mut tasks = Vec::new(); for task in task_iter { tasks.push(task?); } Ok(tasks) } /// Clean up tasks with more granular control pub fn cleanup_tasks_by_criteria( conn: &Connection, days: Option<u64>, status_filter: &[TaskStatus], ) -> Result<usize> { // First, get the tasks that will be deleted (to access log files) let tasks_to_delete = get_cleanup_candidates(conn, days, status_filter)?; if tasks_to_delete.is_empty() { return Ok(0); } // Delete log files first for task in &tasks_to_delete { if std::path::Path::new(&task.log_path).exists() && let Err(e) = std::fs::remove_file(&task.log_path) { eprintln!( "Warning: Failed to delete log file {}: {}", task.log_path, e ); } } // Then delete from database using task IDs let task_ids: Vec<_> = tasks_to_delete.iter().map(|t| &t.id).collect(); let placeholders = task_ids.iter().map(|_| "?").collect::<Vec<_>>().join(","); let sql = format!("DELETE FROM tasks WHERE id IN ({placeholders})"); let param_refs: Vec<&dyn rusqlite::ToSql> = task_ids .iter() .map(|id| *id as &dyn rusqlite::ToSql) .collect(); let rows_affected = conn.execute(&sql, &param_refs[..])?; Ok(rows_affected) }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/storage/task.rs
src/app/storage/task.rs
use super::task_status::TaskStatus; #[derive(Debug, Clone, serde::Serialize, serde::Deserialize)] pub struct Task { pub id: String, pub pid: u32, pub pgid: Option<i32>, pub command: String, // JSON serialized Vec<String> pub env: Option<String>, // JSON serialized environment variables pub cwd: Option<String>, pub status: TaskStatus, pub exit_code: Option<i32>, pub started_at: i64, // Unix timestamp pub finished_at: Option<i64>, pub log_path: String, }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/helpers/file_watcher.rs
src/app/helpers/file_watcher.rs
use std::path::PathBuf; use std::time::Duration; use tokio::fs::File; use tokio::io::{AsyncBufReadExt, AsyncSeekExt, BufReader}; use tokio::sync::mpsc; use crate::app::{error, error::Result}; /// Follow a log file and print new lines as they appear (tail -f behavior) pub async fn follow_log_file(file_path: &PathBuf) -> Result<()> { use notify::{Config, PollWatcher, RecursiveMode, Watcher}; use std::io::SeekFrom; if !tokio::fs::try_exists(file_path).await? { return Err(error::GhostError::InvalidArgument { message: format!("File not found: {path}", path = file_path.display()), }); } // Read and print existing content first let mut file = File::open(file_path).await?; let mut reader = BufReader::new(&mut file); let mut line = String::new(); while reader.read_line(&mut line).await? > 0 { print!("{line}"); line.clear(); } // Get current file position let mut last_position = file.stream_position().await?; // Set up file system watcher let (tx, mut rx) = mpsc::channel(100); let mut watcher = PollWatcher::new( move |res: notify::Result<notify::Event>| { if let Ok(event) = res { let _ = tx.blocking_send(event); } }, Config::default().with_poll_interval(Duration::from_millis(200)), ) .map_err(|e| error::GhostError::FileWatch { message: format!("Failed to create file watcher: {e}"), })?; // Watch the file for changes watcher .watch(file_path, RecursiveMode::NonRecursive) .map_err(|e| error::GhostError::FileWatch { message: format!("Failed to watch file: {e}"), })?; // Main event loop loop { tokio::select! { Some(event) = rx.recv() => { // File was modified, read new content if event.kind.is_modify() { let metadata = tokio::fs::metadata(file_path).await?; let current_size = metadata.len(); if current_size > last_position { // File has grown, read new lines let mut file = File::open(file_path).await?; file.seek(SeekFrom::Start(last_position)).await?; let mut reader = BufReader::new(file); let mut line = String::new(); while reader.read_line(&mut line).await? > 0 { print!("{line}"); use std::io::Write; std::io::stdout().flush().unwrap_or(()); line.clear(); } last_position = current_size; } } } _ = tokio::signal::ctrl_c() => { // Ctrl+C was pressed, break the loop println!("\nLog following stopped."); break; } } } Ok(()) }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/helpers/task_validation.rs
src/app/helpers/task_validation.rs
use crate::app::{error, error::Result, storage}; /// Validate that a task is in running state pub fn validate_task_running(task: &storage::Task) -> Result<()> { if task.status != storage::TaskStatus::Running { return Err(error::GhostError::TaskOperation { task_id: task.id.clone(), message: format!( "Task is not running (status: {status})", status = task.status ), }); } Ok(()) }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/helpers/command_parser.rs
src/app/helpers/command_parser.rs
//! Command string parser for multi-command execution //! //! Parses command strings like "sleep 10" into ["sleep", "10"] use crate::app::error::{GhostError, Result}; /// Parse a command string into command name and arguments /// /// # Examples /// - "sleep 10" -> ["sleep", "10"] /// - "echo 'hello world'" -> ["echo", "hello world"] /// - "echo \"hello world\"" -> ["echo", "hello world"] pub fn parse_command(command_str: &str) -> Result<Vec<String>> { let trimmed = command_str.trim(); if trimmed.is_empty() { return Err(GhostError::InvalidArgument { message: "Empty command string".to_string(), }); } let mut result = Vec::new(); let mut current = String::new(); let mut in_single_quote = false; let mut in_double_quote = false; for c in trimmed.chars() { match c { '\'' if !in_double_quote => { in_single_quote = !in_single_quote; } '"' if !in_single_quote => { in_double_quote = !in_double_quote; } ' ' | '\t' if !in_single_quote && !in_double_quote => { if !current.is_empty() { result.push(current.clone()); current.clear(); } } _ => { current.push(c); } } } if !current.is_empty() { result.push(current); } if in_single_quote || in_double_quote { return Err(GhostError::InvalidArgument { message: "Unclosed quote in command string".to_string(), }); } if result.is_empty() { return Err(GhostError::InvalidArgument { message: "Empty command string".to_string(), }); } Ok(result) } #[cfg(test)] mod tests { use super::*; #[test] fn test_parse_simple_command() { let result = parse_command("sleep 10").unwrap(); assert_eq!(result, vec!["sleep", "10"]); } #[test] fn test_parse_single_command_no_args() { let result = parse_command("ls").unwrap(); assert_eq!(result, vec!["ls"]); } #[test] fn test_parse_command_with_single_quotes() { let result = parse_command("echo 'hello world'").unwrap(); assert_eq!(result, vec!["echo", "hello world"]); } #[test] fn test_parse_command_with_double_quotes() { let result = parse_command("echo \"hello world\"").unwrap(); assert_eq!(result, vec!["echo", "hello world"]); } #[test] fn test_parse_command_with_multiple_args() { let result = parse_command("grep -r 'pattern' /path/to/dir").unwrap(); assert_eq!(result, vec!["grep", "-r", "pattern", "/path/to/dir"]); } #[test] fn test_parse_empty_string_returns_error() { let result = parse_command(""); assert!(result.is_err()); } #[test] fn test_parse_whitespace_only_returns_error() { let result = parse_command(" "); assert!(result.is_err()); } #[test] fn test_parse_unclosed_double_quote_returns_error() { let result = parse_command("echo \"hello world"); assert!(result.is_err()); } #[test] fn test_parse_unclosed_single_quote_returns_error() { let result = parse_command("echo 'hello world"); assert!(result.is_err()); } }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/helpers/time.rs
src/app/helpers/time.rs
use std::time::{SystemTime, UNIX_EPOCH}; /// Get current Unix timestamp in seconds pub fn now_timestamp() -> i64 { SystemTime::now() .duration_since(UNIX_EPOCH) .unwrap() .as_secs() as i64 }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/helpers/mod.rs
src/app/helpers/mod.rs
pub mod command_parser; pub mod file_watcher; pub mod task_validation; pub mod time; // Re-export for backward compatibility pub use command_parser::parse_command; pub use file_watcher::follow_log_file; pub use task_validation::validate_task_running; pub use time::now_timestamp;
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/tui/app.rs
src/app/tui/app.rs
use crossterm::event::{KeyCode, KeyEvent, KeyModifiers}; use ratatui::{Frame, layout::Rect}; use rusqlite::Connection; use std::collections::HashMap; use std::fs; use std::process::Child; use std::time::SystemTime; use tui_scrollview::ScrollViewState; use super::log_viewer_scrollview::LogViewerScrollWidget; use super::table_state_scroll::TableScroll; use super::{TaskFilter, ViewMode}; use crate::app::config::Config; use crate::app::error::Result; use crate::app::storage; use crate::app::storage::task::Task; use crate::app::storage::task_repository; /// Cache for log file content struct LogCache { content: Vec<String>, last_modified: SystemTime, file_size: u64, } enum UpdateStrategy { FullReload, Incremental(u64), // previous file size UseCache, } pub struct TuiApp { pub tasks: Vec<Task>, pub table_scroll: TableScroll, pub filter: TaskFilter, pub should_quit: bool, pub view_mode: ViewMode, pub log_scroll_offset: usize, pub log_lines_count: usize, pub log_scroll_state: ScrollViewState, pub selected_task_id: Option<String>, pub env_scroll_state: ScrollViewState, pub last_render_area: Rect, pub auto_scroll_enabled: bool, conn: Connection, log_cache: HashMap<String, LogCache>, child_processes: HashMap<String, Child>, } impl TuiApp { pub fn new() -> Result<Self> { let conn = storage::init_database()?; Ok(Self { tasks: Vec::new(), table_scroll: TableScroll::new(), filter: TaskFilter::All, should_quit: false, view_mode: ViewMode::TaskList, log_scroll_offset: 0, log_lines_count: 0, log_scroll_state: ScrollViewState::default(), selected_task_id: None, env_scroll_state: ScrollViewState::default(), last_render_area: Rect::default(), auto_scroll_enabled: true, // Auto-scroll enabled by default conn, log_cache: HashMap::new(), child_processes: HashMap::new(), }) } /// Create a new TuiApp with a specific config (for testing) pub fn new_with_config(config: Config) -> Result<Self> { let conn = storage::init_database_with_config(Some(config))?; Ok(Self { tasks: Vec::new(), table_scroll: TableScroll::new(), filter: TaskFilter::All, should_quit: false, view_mode: ViewMode::TaskList, log_scroll_offset: 0, log_lines_count: 0, log_scroll_state: ScrollViewState::default(), selected_task_id: None, env_scroll_state: ScrollViewState::default(), last_render_area: Rect::default(), auto_scroll_enabled: true, // Auto-scroll enabled by default conn, log_cache: HashMap::new(), child_processes: HashMap::new(), }) } /// Load tasks from database pub fn refresh_tasks(&mut self) -> Result<()> { // Clean up finished child processes first self.cleanup_finished_processes(); // Filter status for database query let status_filter = match self.filter { TaskFilter::All => None, TaskFilter::Running => Some("running"), TaskFilter::Exited => Some("exited"), TaskFilter::Killed => Some("killed"), }; self.tasks = task_repository::get_tasks_with_process_check(&self.conn, status_filter)?; // Update table scroll with new item count self.table_scroll.set_total_items(self.tasks.len()); Ok(()) } /// Handle keyboard input pub fn handle_key(&mut self, key: KeyEvent) -> Result<()> { match self.view_mode { ViewMode::TaskList => self.handle_task_list_key(key), ViewMode::LogView => self.handle_log_view_key(key), ViewMode::ProcessDetails => self.handle_process_details_key(key), } } fn handle_task_list_key(&mut self, key: KeyEvent) -> Result<()> { match key.code { KeyCode::Char('q') => { self.should_quit = true; } KeyCode::Char('j') => { self.table_scroll.next(); } KeyCode::Char('k') if !key.modifiers.contains(KeyModifiers::CONTROL) => { self.table_scroll.previous(); } KeyCode::Char('g') if key.modifiers.contains(KeyModifiers::NONE) => { self.table_scroll.first(); } KeyCode::Char('G') => { self.table_scroll.last(); } KeyCode::Char('l') => { if !self.tasks.is_empty() { self.view_mode = ViewMode::LogView; self.initialize_log_view(); } } KeyCode::Char('c') if key.modifiers.contains(KeyModifiers::CONTROL) => { self.should_quit = true; } KeyCode::Char('s') => { if !self.tasks.is_empty() { self.stop_task(false); } } KeyCode::Char('k') if key.modifiers.contains(KeyModifiers::CONTROL) => { if !self.tasks.is_empty() { self.stop_task(true); } } KeyCode::Tab => { self.cycle_filter(); self.refresh_tasks()?; } KeyCode::Enter => { if !self.tasks.is_empty() { let selected_task = &self.tasks[self.selected_index()]; self.selected_task_id = Some(selected_task.id.clone()); self.view_mode = ViewMode::ProcessDetails; self.env_scroll_state = ScrollViewState::default(); } } KeyCode::Char('d') if key.modifiers.contains(KeyModifiers::CONTROL) => { let page_size = self.calculate_table_page_size(); self.table_scroll.page_down(page_size); } KeyCode::Char('u') if key.modifiers.contains(KeyModifiers::CONTROL) => { let page_size = self.calculate_table_page_size(); self.table_scroll.page_up(page_size); } KeyCode::Char('r') => { self.rerun_selected_command()?; } _ => {} } Ok(()) } fn handle_log_view_key(&mut self, key: KeyEvent) -> Result<()> { match key.code { KeyCode::Esc => { self.view_mode = ViewMode::TaskList; self.log_scroll_state.scroll_to_top(); } KeyCode::Char('q') => { self.should_quit = true; } KeyCode::Char('j') => { self.auto_scroll_enabled = false; // Disable auto-scroll on manual navigation self.log_scroll_state.scroll_down(); } KeyCode::Char('k') => { self.auto_scroll_enabled = false; // Disable auto-scroll on manual navigation self.log_scroll_state.scroll_up(); } KeyCode::Char('h') => { self.auto_scroll_enabled = false; // Disable auto-scroll on manual navigation self.log_scroll_state.scroll_left(); } KeyCode::Char('l') => { self.auto_scroll_enabled = false; // Disable auto-scroll on manual navigation self.log_scroll_state.scroll_right(); } KeyCode::Char('g') if key.modifiers.contains(KeyModifiers::NONE) => { self.auto_scroll_enabled = false; // Disable auto-scroll on manual navigation self.log_scroll_state.scroll_to_top(); } KeyCode::Char('G') => { self.auto_scroll_enabled = false; // Disable auto-scroll on manual navigation self.log_scroll_state.scroll_to_bottom(); } KeyCode::Char('c') if key.modifiers.contains(KeyModifiers::CONTROL) => { self.should_quit = true; } KeyCode::Char('d') if key.modifiers.contains(KeyModifiers::CONTROL) => { self.auto_scroll_enabled = false; // Disable auto-scroll on manual navigation self.log_scroll_state.scroll_page_down(); } KeyCode::Char('u') if key.modifiers.contains(KeyModifiers::CONTROL) => { self.auto_scroll_enabled = false; // Disable auto-scroll on manual navigation self.log_scroll_state.scroll_page_up(); } KeyCode::Char('f') if key.modifiers.contains(KeyModifiers::CONTROL) => { // Toggle auto-scroll mode with Ctrl+F self.auto_scroll_enabled = !self.auto_scroll_enabled; // If enabling auto-scroll, immediately scroll to bottom if self.auto_scroll_enabled { self.log_scroll_state.scroll_to_bottom(); } } _ => {} } Ok(()) } fn handle_process_details_key(&mut self, key: KeyEvent) -> Result<()> { match key.code { KeyCode::Esc => { self.view_mode = ViewMode::TaskList; self.env_scroll_state = ScrollViewState::default(); } KeyCode::Char('j') => { self.env_scroll_state.scroll_down(); } KeyCode::Char('k') => { self.env_scroll_state.scroll_up(); } KeyCode::Char('q') => { self.should_quit = true; } KeyCode::Char('c') if key.modifiers.contains(KeyModifiers::CONTROL) => { self.should_quit = true; } KeyCode::Char('d') if key.modifiers.contains(KeyModifiers::CONTROL) => { self.env_scroll_state.scroll_page_down(); } KeyCode::Char('u') if key.modifiers.contains(KeyModifiers::CONTROL) => { self.env_scroll_state.scroll_page_up(); } _ => {} } Ok(()) } fn initialize_log_view(&mut self) { if let Some(selected) = self.table_scroll.selected() && selected < self.tasks.len() { let selected_task = &self.tasks[selected]; let log_path = &selected_task.log_path; // Check cache first if let Some(cache) = self.log_cache.get(log_path) { self.log_lines_count = cache.content.len(); } else { // If not in cache, we'll load it on first render self.log_lines_count = 0; } // Reset scroll state to start from the top self.log_scroll_state.scroll_to_top(); } } /// Render the TUI pub fn render(&mut self, frame: &mut Frame) { let area = frame.area(); self.last_render_area = area; match self.view_mode { ViewMode::TaskList => self.render_task_list(frame, area), ViewMode::LogView => self.render_log_view(frame, area), ViewMode::ProcessDetails => self.render_process_details(frame, area), } } /// Render task list widget fn render_task_list(&mut self, frame: &mut Frame, area: Rect) { use super::task_list::TaskListWidget; let widget = TaskListWidget::new(&self.tasks, &self.filter, &mut self.table_scroll); frame.render_widget(widget, area); } /// Render log view widget fn render_log_view(&mut self, frame: &mut Frame, area: Rect) { if let Some(selected) = self.table_scroll.selected() && selected < self.tasks.len() { let selected_task = &self.tasks[selected]; let log_path = &selected_task.log_path; // Check if we need to reload or incrementally update the file let update_strategy = if let Ok(metadata) = fs::metadata(log_path) { let modified = metadata.modified().unwrap_or(SystemTime::UNIX_EPOCH); let file_size = metadata.len(); if let Some(cache) = self.log_cache.get(log_path) { if modified > cache.last_modified { if file_size > cache.file_size { // File grew, use incremental update UpdateStrategy::Incremental(cache.file_size) } else { // File changed in other ways, full reload UpdateStrategy::FullReload } } else { // No changes UpdateStrategy::UseCache } } else { // No cache exists, need to load UpdateStrategy::FullReload } } else { // File doesn't exist or can't read metadata UpdateStrategy::UseCache }; // Use scrollview widget let mut scrollview_widget = match update_strategy { UpdateStrategy::FullReload => LogViewerScrollWidget::new(selected_task), UpdateStrategy::Incremental(previous_size) => { let cache = self.log_cache.get(log_path).unwrap(); LogViewerScrollWidget::load_incremental_content( selected_task, cache.content.clone(), previous_size, ) } UpdateStrategy::UseCache => { let cache = self.log_cache.get(log_path).unwrap(); LogViewerScrollWidget::with_cached_content(selected_task, cache.content.clone()) } }; // Set auto-scroll state from app if self.auto_scroll_enabled { scrollview_widget.enable_auto_scroll(); } else { scrollview_widget.disable_auto_scroll(); } // Update cache if needed if matches!( update_strategy, UpdateStrategy::FullReload | UpdateStrategy::Incremental(_) ) && let Ok(metadata) = fs::metadata(log_path) { let modified = metadata.modified().unwrap_or(SystemTime::UNIX_EPOCH); self.log_cache.insert( log_path.clone(), LogCache { content: scrollview_widget.get_lines().to_vec(), last_modified: modified, file_size: metadata.len(), }, ); } // Handle auto-scroll update let new_line_count = scrollview_widget.get_lines_count(); self.handle_auto_scroll_update(&update_strategy, new_line_count); // Render with scrollview state frame.render_stateful_widget(scrollview_widget, area, &mut self.log_scroll_state); } } pub fn should_quit(&self) -> bool { self.should_quit } /// Stop the selected task fn stop_task(&mut self, force: bool) { if self.selected_index() < self.tasks.len() { let task = &self.tasks[self.selected_index()]; let task_id = &task.id; // Send signal to stop the task (commands::stop handles process group killing) // Use show_output=false to suppress console output in TUI let _ = crate::app::commands::stop(&self.conn, task_id, force, false); // Refresh task list to update status let _ = self.refresh_tasks(); } } /// Cycle through task filters fn cycle_filter(&mut self) { self.filter = match self.filter { TaskFilter::All => TaskFilter::Running, TaskFilter::Running => TaskFilter::Exited, TaskFilter::Exited => TaskFilter::Killed, TaskFilter::Killed => TaskFilter::All, }; // Reset selection when changing filter self.table_scroll = TableScroll::new(); } /// Render process details view fn render_process_details(&mut self, frame: &mut Frame, area: Rect) { use super::process_details::ProcessDetailsWidget; // Find the selected task if let Some(task_id) = &self.selected_task_id { if let Some(task) = self.tasks.iter().find(|t| t.id == *task_id) { let widget = ProcessDetailsWidget::new(task); widget.render(frame, area, &mut self.env_scroll_state); } else { // Task not found, go back to task list self.view_mode = ViewMode::TaskList; self.selected_task_id = None; } } else { // No task selected, go back to task list self.view_mode = ViewMode::TaskList; } } // Accessor methods for tests compatibility pub fn selected_index(&self) -> usize { self.table_scroll.selected().unwrap_or(0) } pub fn set_selected_index(&mut self, index: usize) { if index < self.tasks.len() { self.table_scroll.select(Some(index)); } } pub fn table_scroll_offset(&self) -> usize { // Calculate visible offset based on selection let selected = self.table_scroll.selected().unwrap_or(0); selected.saturating_sub(2) // Keep some context above } fn calculate_table_page_size(&self) -> usize { // Calculate the visible height of the table based on last render area // Account for borders (2), header (1), footer separator (1), and footer (1) let overhead = 5; self.last_render_area.height.saturating_sub(overhead) as usize } /// Rerun the selected task's command fn rerun_selected_command(&mut self) -> Result<()> { if !self.tasks.is_empty() { let selected_task = &self.tasks[self.selected_index()]; // Parse the JSON command string let command: Vec<String> = serde_json::from_str(&selected_task.command).map_err(|e| { crate::app::error::GhostError::InvalidArgument { message: format!("Failed to parse command JSON: {e}"), } })?; // Parse environment variables if they exist let env_vars = if let Some(env_json) = &selected_task.env { let env_pairs: Vec<(String, String)> = serde_json::from_str(env_json).map_err(|e| { crate::app::error::GhostError::InvalidArgument { message: format!("Failed to parse environment JSON: {e}"), } })?; env_pairs } else { vec![] }; // Parse working directory let cwd = selected_task.cwd.as_ref().map(std::path::PathBuf::from); // Spawn the command let (process_info, child) = crate::app::commands::spawn_and_register_process( command.clone(), cwd, env_vars, &self.conn, )?; // Store the child process for proper cleanup self.child_processes.insert(process_info.id.clone(), child); // Refresh the task list to show the new process self.refresh_tasks()?; } Ok(()) } /// Handle auto-scroll update based on new content fn handle_auto_scroll_update( &mut self, update_strategy: &UpdateStrategy, new_line_count: usize, ) { let had_new_content = matches!( update_strategy, UpdateStrategy::FullReload | UpdateStrategy::Incremental(_) ) && new_line_count > self.log_lines_count; self.log_lines_count = new_line_count; // If auto-scroll is enabled and we have new content, scroll to bottom if self.auto_scroll_enabled && had_new_content { self.log_scroll_state.scroll_to_bottom(); } } /// Clean up finished child processes to prevent zombie processes fn cleanup_finished_processes(&mut self) { self.child_processes.retain(|_, child| { match child.try_wait() { Ok(Some(_)) => false, // Process finished, remove it Ok(None) => true, // Process still running, keep it Err(_) => false, } }); } }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/tui/process_details.rs
src/app/tui/process_details.rs
use ratatui::{ Frame, layout::{Alignment, Constraint, Direction, Layout, Rect, Size}, style::{Color, Modifier, Style}, symbols, text::{Line, Span}, widgets::{Block, Borders, Paragraph, Wrap}, }; use tui_scrollview::{ScrollView, ScrollViewState, ScrollbarVisibility}; use crate::app::port_detector::detect_listening_ports; use crate::app::storage::task::Task; use crate::app::storage::task_status::TaskStatus; use chrono::{TimeZone, Utc}; pub struct ProcessDetailsWidget<'a> { task: &'a Task, } impl<'a> ProcessDetailsWidget<'a> { pub fn new(task: &'a Task) -> Self { Self { task } } fn format_command(&self) -> String { // Parse JSON command if let Ok(command_vec) = serde_json::from_str::<Vec<String>>(&self.task.command) { command_vec.join(" ") } else { // Fallback if parsing fails self.task.command.clone() } } pub fn render(self, frame: &mut Frame, area: Rect, env_scroll_state: &mut ScrollViewState) { let chunks = Layout::default() .direction(Direction::Vertical) .constraints([ Constraint::Length(8), // Basic info section (6 lines + 2 borders) Constraint::Length(5), // Listening ports section Constraint::Min(5), // Environment variables section Constraint::Length(2), // Footer ]) .split(area); // Render basic info section self.render_basic_info(frame, chunks[0]); // Render listening ports section self.render_listening_ports(frame, chunks[1]); // Render environment variables section self.render_environment_variables(frame, chunks[2], env_scroll_state); // Render footer self.render_footer(frame, chunks[3]); } fn render_basic_info(&self, frame: &mut Frame, area: Rect) { let block = Block::default() .title(" Process Details ") .borders(Borders::ALL) .border_style(Style::default().fg(Color::Cyan)); // Calculate runtime let runtime = { let started = Utc.timestamp_opt(self.task.started_at, 0).single().unwrap(); let elapsed = if self.task.status == TaskStatus::Running { let now = Utc::now(); now.signed_duration_since(started) } else if let Some(finished_at) = self.task.finished_at { let ended = Utc.timestamp_opt(finished_at, 0).single().unwrap(); ended.signed_duration_since(started) } else { chrono::Duration::zero() }; let hours = elapsed.num_hours(); let minutes = elapsed.num_minutes() % 60; let seconds = elapsed.num_seconds() % 60; if hours > 0 { format!("{hours}h {minutes}m {seconds}s") } else if minutes > 0 { format!("{minutes}m {seconds}s") } else { format!("{seconds}s") } }; // Format status with color let status_style = match self.task.status { TaskStatus::Running => Style::default().fg(Color::Green), TaskStatus::Exited => Style::default().fg(Color::Yellow), TaskStatus::Killed => Style::default().fg(Color::Red), TaskStatus::Unknown => Style::default().fg(Color::Gray), }; let status_text = format!("{} ({})", self.task.status.as_str(), runtime); // Build info lines let info_lines = vec![ Line::from(vec![ Span::styled("Task ID: ", Style::default().add_modifier(Modifier::BOLD)), Span::raw(&self.task.id), ]), Line::from(vec![ Span::styled("Command: ", Style::default().add_modifier(Modifier::BOLD)), Span::raw(self.format_command()), ]), Line::from(vec![ Span::styled("Status: ", Style::default().add_modifier(Modifier::BOLD)), Span::styled(status_text, status_style), ]), Line::from(vec![ Span::styled("PID: ", Style::default().add_modifier(Modifier::BOLD)), Span::raw(self.task.pid.to_string()), Span::raw(" | "), Span::styled("PGID: ", Style::default().add_modifier(Modifier::BOLD)), Span::raw(self.task.pgid.map_or("N/A".to_string(), |p| p.to_string())), ]), Line::from(vec![ Span::styled("Directory: ", Style::default().add_modifier(Modifier::BOLD)), Span::raw(self.task.cwd.as_deref().unwrap_or("N/A")), ]), Line::from(vec![ Span::styled("Log File: ", Style::default().add_modifier(Modifier::BOLD)), Span::raw(&self.task.log_path), ]), ]; let paragraph = Paragraph::new(info_lines) .block(block) .wrap(Wrap { trim: true }); frame.render_widget(paragraph, area); } fn render_listening_ports(&self, frame: &mut Frame, area: Rect) { let block = Block::default() .title(" Listening Ports ") .borders(Borders::ALL) .border_style(Style::default().fg(Color::Cyan)); let port_lines = if self.task.status == TaskStatus::Running { // Get actual listening ports for running processes match detect_listening_ports(self.task.pid) { Ok(ports) => { if ports.is_empty() { vec![Line::from(Span::styled( "Not listening on any ports", Style::default().fg(Color::DarkGray), ))] } else { ports .into_iter() .map(|port| { Line::from(vec![ Span::styled( format!("{:<6}", port.protocol), Style::default().fg(Color::Blue), ), Span::raw(format!("{:<20}", port.local_addr)), Span::styled( port.state.clone(), Style::default().fg(Color::Green), ), ]) }) .collect() } } Err(e) => { // Check if it's a command not found error if let crate::app::error::GhostError::CommandNotFound { command } = e { vec![ Line::from(Span::styled( format!("{command} command not found"), Style::default().fg(Color::Yellow), )), Line::from(Span::styled( format!("Please install {command} to enable port detection"), Style::default().fg(Color::DarkGray), )), ] } else { vec![Line::from(Span::styled( format!("Failed to detect ports: {e:?}"), Style::default().fg(Color::Red), ))] } } } } else { vec![Line::from(Span::styled( "Process not running", Style::default().fg(Color::DarkGray), ))] }; let paragraph = Paragraph::new(port_lines) .block(block) .wrap(Wrap { trim: true }); frame.render_widget(paragraph, area); } fn render_environment_variables( &self, frame: &mut Frame, area: Rect, scroll_state: &mut ScrollViewState, ) { // Split the area into content and footer separator let chunks = Layout::default() .direction(Direction::Vertical) .constraints([Constraint::Min(1), Constraint::Length(1)]) .split(area); let block = Block::default() .title(" Environment Variables ") .borders(Borders::LEFT | Borders::RIGHT | Borders::TOP) .border_style(Style::default().fg(Color::Cyan)); let inner = block.inner(chunks[0]); // Prepare environment variables content let env_lines: Vec<String> = if let Some(env_json) = &self.task.env { // Parse JSON environment variables if let Ok(env_map) = serde_json::from_str::<Vec<(String, String)>>(env_json) { env_map .iter() .map(|(key, value)| format!("{key}={value}")) .collect() } else { vec!["Failed to parse environment variables".to_string()] } } else { vec!["No environment variables set".to_string()] }; // Render the block first frame.render_widget(block, chunks[0]); // Calculate content size - use inner width to avoid horizontal scrolling let content_height = env_lines.len() as u16; let content_size = Size::new(inner.width, content_height); // Create scrollview widget with proper size and hide horizontal scrollbar let mut scroll_view = ScrollView::new(content_size) .horizontal_scrollbar_visibility(ScrollbarVisibility::Never) .vertical_scrollbar_visibility(ScrollbarVisibility::Never); // Render environment variables with wrapping let env_text = env_lines.join("\n"); let env_paragraph = Paragraph::new(env_text) .style(Style::default()) .wrap(Wrap { trim: false }); // Use the inner width for rendering to enable text wrapping scroll_view.render_widget(env_paragraph, Rect::new(0, 0, inner.width, content_height)); // Render the scrollview frame.render_stateful_widget(scroll_view, inner, scroll_state); // Draw the separator line between environment variables and footer // Using direct buffer manipulation like LogViewerScrollWidget let buf = frame.buffer_mut(); if chunks[1].y > 0 && chunks[0].width > 0 { // Left connection: ├ buf[(chunks[0].x, chunks[1].y)] .set_symbol(symbols::line::VERTICAL_RIGHT) .set_style(Style::default().fg(Color::Cyan)); // Horizontal line for x in chunks[0].x + 1..chunks[0].x + chunks[0].width - 1 { buf[(x, chunks[1].y)] .set_symbol(symbols::line::HORIZONTAL) .set_style(Style::default().fg(Color::Cyan)); } // Right connection: ┤ buf[(chunks[0].x + chunks[0].width - 1, chunks[1].y)] .set_symbol(symbols::line::VERTICAL_LEFT) .set_style(Style::default().fg(Color::Cyan)); } } fn render_footer(&self, frame: &mut Frame, area: Rect) { // Render keybinds let keybinds = vec![ Span::styled("[q]", Style::default().fg(Color::Yellow)), Span::raw(" Quit "), Span::styled("[Esc]", Style::default().fg(Color::Yellow)), Span::raw(" Back to list "), Span::styled("[j/k]", Style::default().fg(Color::Yellow)), Span::raw(" Scroll "), Span::styled("[C-d/C-u]", Style::default().fg(Color::Yellow)), Span::raw(" Page"), ]; let keybind_paragraph = Paragraph::new(Line::from(keybinds)) .style(Style::default()) .alignment(Alignment::Center) .block( Block::default() .borders(Borders::LEFT | Borders::RIGHT | Borders::BOTTOM) .border_style(Style::default().fg(Color::Cyan)), ); frame.render_widget(keybind_paragraph, area); } }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/tui/task_list.rs
src/app/tui/task_list.rs
use ratatui::{ Frame, layout::{Constraint, Rect}, style::{Color, Style}, widgets::{Block, Borders, Cell, Row, StatefulWidget, Table, TableState, Widget}, }; // Layout constants const ID_COLUMN_WIDTH: u16 = 38; // Full UUID (36 chars) + 2 for padding const PID_COLUMN_WIDTH: u16 = 8; const STATUS_COLUMN_WIDTH: u16 = 9; const STARTED_COLUMN_WIDTH: u16 = 16; const COMMAND_COLUMN_MIN_WIDTH: u16 = 20; const DIRECTORY_COLUMN_MIN_WIDTH: u16 = 20; // Column constraints for the table const COLUMN_CONSTRAINTS: [Constraint; 6] = [ Constraint::Length(ID_COLUMN_WIDTH), Constraint::Length(PID_COLUMN_WIDTH), Constraint::Length(STATUS_COLUMN_WIDTH), Constraint::Length(STARTED_COLUMN_WIDTH), Constraint::Min(COMMAND_COLUMN_MIN_WIDTH), Constraint::Min(DIRECTORY_COLUMN_MIN_WIDTH), ]; use super::{App, TaskFilter, table_state_scroll::TableScroll}; use crate::app::storage::task::Task; use crate::app::storage::task_status::TaskStatus; impl App { pub fn render_task_list(&mut self, frame: &mut Frame, area: Rect) { let task_list_widget = TaskListWidget::new(&self.tasks, &self.filter, &mut self.table_scroll); frame.render_widget(task_list_widget, area); } } pub struct TaskListWidget<'a> { tasks: &'a [Task], filter: &'a TaskFilter, table_scroll: &'a mut TableScroll, } impl<'a> TaskListWidget<'a> { pub fn new( tasks: &'a [Task], filter: &'a TaskFilter, table_scroll: &'a mut TableScroll, ) -> Self { Self { tasks, filter, table_scroll, } } fn filter_name(&self) -> &'static str { match self.filter { TaskFilter::All => "All", TaskFilter::Running => "Running", TaskFilter::Exited => "Exited", TaskFilter::Killed => "Killed", } } fn status_style(&self, status: &TaskStatus) -> Style { match status { TaskStatus::Running => Style::default().fg(Color::Green), TaskStatus::Exited => Style::default().fg(Color::Blue), TaskStatus::Killed => Style::default().fg(Color::Red), TaskStatus::Unknown => Style::default().fg(Color::Gray), } } fn parse_command(&self, command_json: &str) -> String { match serde_json::from_str::<Vec<String>>(command_json) { Ok(cmd_vec) => cmd_vec.join(" "), Err(_) => command_json.to_string(), } } fn format_timestamp(&self, timestamp: i64) -> String { use chrono::{DateTime, Utc}; let dt = DateTime::<Utc>::from_timestamp(timestamp, 0).unwrap(); dt.format("%Y-%m-%d %H:%M").to_string() } fn create_header_row(&self) -> Row<'_> { Row::new(vec![ Cell::from(" ID"), Cell::from(" PID"), Cell::from(" Status"), Cell::from(" Started"), Cell::from(" Command"), Cell::from(" Directory"), ]) .style(Style::default()) } } impl<'a> Widget for TaskListWidget<'a> { fn render(self, area: Rect, buf: &mut ratatui::buffer::Buffer) { let filter_name = self.filter_name(); let title = format!( " Ghost v{} [Filter: {filter_name}] ", env!("CARGO_PKG_VERSION") ); // Create main block let block = Block::default() .borders(Borders::ALL) .title(title) .border_style(Style::default().fg(Color::Green)); // Get inner area for content let inner_area = block.inner(area); // Render the block border first ratatui::widgets::Widget::render(block, area, buf); // Calculate areas dynamically based on available space // For 12-line terminal: total=12, border=2, inner=10, content=7, separator=1, footer=1, remaining=1 // For the specific test case: height=12, inner=10, we want content=5 to match expected output let content_height = if inner_area.height == 10 { 5 // Specific for 12-line terminal test - gets us 6 content lines with header } else { inner_area.height.saturating_sub(2) }; // Render table content self.render_table_content( Rect { x: inner_area.x, y: inner_area.y, width: inner_area.width, height: content_height, }, buf, ); // Only render footer if there's enough space if inner_area.height >= 2 { // Render footer separator (right before the footer text) let footer_text_y = inner_area.y + inner_area.height - 1; let separator_y = footer_text_y - 1; if separator_y >= inner_area.y { self.render_footer_separator(inner_area.x, separator_y, inner_area.width, buf); } // Render footer text at the last line of inner area self.render_footer_text(inner_area.x, footer_text_y, inner_area.width, buf); } } } impl<'a> TaskListWidget<'a> { fn render_table_content(&self, area: Rect, buf: &mut ratatui::buffer::Buffer) { if self.tasks.is_empty() { let rows: Vec<Row<'_>> = vec![]; let table = Table::new(rows, COLUMN_CONSTRAINTS).header(self.create_header_row()); ratatui::widgets::Widget::render(table, area, buf); } else { // Table with tasks let rows: Vec<Row> = self .tasks .iter() .map(|task| { let status_style = self.status_style(&task.status); let task_id = &task.id; let pid = task.pid; let status = task.status.as_str(); let timestamp = self.format_timestamp(task.started_at); let command = self.parse_command(&task.command); let directory = task.cwd.as_deref().unwrap_or("-"); Row::new(vec![ Cell::from(format!(" {task_id}")), // Show full ID Cell::from(format!(" {pid}")), Cell::from(format!(" {status}")).style(status_style), Cell::from(format!(" {timestamp}")), Cell::from(format!(" {command}")), Cell::from(format!(" {directory}")), ]) }) .collect(); let table = Table::new(rows, COLUMN_CONSTRAINTS) .header(self.create_header_row()) .row_highlight_style(Style::default().bg(Color::DarkGray)); // Use a temporary table state and apply the selection let mut table_state = TableState::default(); table_state.select(self.table_scroll.selected()); StatefulWidget::render(table, area, buf, &mut table_state); } } fn render_footer_separator( &self, x: u16, y: u16, width: u16, buf: &mut ratatui::buffer::Buffer, ) { // Draw the separator line: ├─────...─────┤ // Need to overwrite the left and right border characters buf[(x - 1, y)].set_symbol("├"); for i in 0..width { buf[(x + i, y)] .set_symbol("─") .set_style(Style::default().fg(Color::Green)); } buf[(x + width, y)].set_symbol("┤"); } fn render_footer_text(&self, x: u16, y: u16, width: u16, buf: &mut ratatui::buffer::Buffer) { let keybinds_text = " j/k:Move l:Log r:Rerun s/C-k:Stop q:Quit g/G:Top/Bot C-d/u:Page"; // Draw the text for (i, ch) in keybinds_text.chars().enumerate() { let pos_x = x + i as u16; if pos_x < x + width { buf[(pos_x, y)].set_symbol(&ch.to_string()); } } // Fill remaining space with spaces up to the border let text_len = keybinds_text.chars().count() as u16; for i in text_len..width { buf[(x + i, y)].set_symbol(" "); } } }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/tui/table_state_scroll.rs
src/app/tui/table_state_scroll.rs
use ratatui::widgets::TableState; /// Wrapper for managing table scrolling with TableState #[derive(Debug, Default)] pub struct TableScroll { state: TableState, total_items: usize, } impl TableScroll { pub fn new() -> Self { let mut state = TableState::default(); state.select(Some(0)); Self { state, total_items: 0, } } pub fn with_items(total_items: usize) -> Self { let mut state = TableState::default(); if total_items > 0 { state.select(Some(0)); } Self { state, total_items } } pub fn state_mut(&mut self) -> &mut TableState { &mut self.state } pub fn selected(&self) -> Option<usize> { self.state.selected() } pub fn select(&mut self, index: Option<usize>) { self.state.select(index); } pub fn next(&mut self) { if self.total_items == 0 { return; } let i = match self.state.selected() { Some(i) => { if i >= self.total_items - 1 { i // Stay at the last item } else { i + 1 } } None => 0, }; self.state.select(Some(i)); } pub fn previous(&mut self) { if self.total_items == 0 { return; } let i = match self.state.selected() { Some(i) => { if i == 0 { 0 // Stay at the first item } else { i - 1 } } None => 0, }; self.state.select(Some(i)); } pub fn first(&mut self) { if self.total_items > 0 { self.state.select(Some(0)); } } pub fn last(&mut self) { if self.total_items > 0 { self.state.select(Some(self.total_items - 1)); } } pub fn set_total_items(&mut self, total: usize) { self.total_items = total; // Adjust selection if current selection is out of bounds if let Some(selected) = self.state.selected() { if selected >= total && total > 0 { self.state.select(Some(total - 1)); } else if total == 0 { self.state.select(None); } } else if total > 0 { self.state.select(Some(0)); } } pub fn page_down(&mut self, page_size: usize) { if self.total_items == 0 || page_size == 0 { return; } let current = self.state.selected().unwrap_or(0); let new_index = (current + page_size).min(self.total_items - 1); self.state.select(Some(new_index)); } pub fn page_up(&mut self, page_size: usize) { if self.total_items == 0 || page_size == 0 { return; } let current = self.state.selected().unwrap_or(0); let new_index = current.saturating_sub(page_size); self.state.select(Some(new_index)); } } #[cfg(test)] mod tests { use super::*; #[test] fn test_new_table_scroll() { let scroll = TableScroll::new(); assert_eq!(scroll.selected(), Some(0)); assert_eq!(scroll.total_items, 0); } #[test] fn test_with_items() { let scroll = TableScroll::with_items(5); assert_eq!(scroll.selected(), Some(0)); assert_eq!(scroll.total_items, 5); } #[test] fn test_with_zero_items() { let scroll = TableScroll::with_items(0); assert_eq!(scroll.selected(), None); assert_eq!(scroll.total_items, 0); } #[test] fn test_next_navigation() { let mut scroll = TableScroll::with_items(3); // Start at 0 assert_eq!(scroll.selected(), Some(0)); // Move to 1 scroll.next(); assert_eq!(scroll.selected(), Some(1)); // Move to 2 scroll.next(); assert_eq!(scroll.selected(), Some(2)); // Stay at 2 (no wrap) scroll.next(); assert_eq!(scroll.selected(), Some(2)); // Confirm it stays at 2 scroll.next(); assert_eq!(scroll.selected(), Some(2)); } #[test] fn test_previous_navigation() { let mut scroll = TableScroll::with_items(3); // Start at 0 assert_eq!(scroll.selected(), Some(0)); // Stay at 0 (no wrap) scroll.previous(); assert_eq!(scroll.selected(), Some(0)); // Move to 2 scroll.select(Some(2)); assert_eq!(scroll.selected(), Some(2)); // Move to 1 scroll.previous(); assert_eq!(scroll.selected(), Some(1)); // Move to 0 scroll.previous(); assert_eq!(scroll.selected(), Some(0)); // Stay at 0 scroll.previous(); assert_eq!(scroll.selected(), Some(0)); } #[test] fn test_first_last() { let mut scroll = TableScroll::with_items(5); // Move to middle scroll.select(Some(2)); assert_eq!(scroll.selected(), Some(2)); // Go to first scroll.first(); assert_eq!(scroll.selected(), Some(0)); // Go to last scroll.last(); assert_eq!(scroll.selected(), Some(4)); } #[test] fn test_empty_table_navigation() { let mut scroll = TableScroll::with_items(0); // All navigation should do nothing scroll.next(); assert_eq!(scroll.selected(), None); scroll.previous(); assert_eq!(scroll.selected(), None); scroll.first(); assert_eq!(scroll.selected(), None); scroll.last(); assert_eq!(scroll.selected(), None); } #[test] fn test_set_total_items() { let mut scroll = TableScroll::with_items(5); scroll.select(Some(3)); // Reduce items - selection should adjust scroll.set_total_items(2); assert_eq!(scroll.selected(), Some(1)); assert_eq!(scroll.total_items, 2); // Set to zero - selection should be None scroll.set_total_items(0); assert_eq!(scroll.selected(), None); assert_eq!(scroll.total_items, 0); // Increase from zero - selection should be Some(0) scroll.set_total_items(3); assert_eq!(scroll.selected(), Some(0)); assert_eq!(scroll.total_items, 3); } #[test] fn test_single_item_navigation() { let mut scroll = TableScroll::with_items(1); // Should stay at 0 assert_eq!(scroll.selected(), Some(0)); // Next should stay at 0 scroll.next(); assert_eq!(scroll.selected(), Some(0)); // Previous should stay at 0 scroll.previous(); assert_eq!(scroll.selected(), Some(0)); } #[test] fn test_page_navigation() { let mut scroll = TableScroll::with_items(20); // Start at 0 assert_eq!(scroll.selected(), Some(0)); // Page down by 5 scroll.page_down(5); assert_eq!(scroll.selected(), Some(5)); // Page down by 10 more scroll.page_down(10); assert_eq!(scroll.selected(), Some(15)); // Page down to the end scroll.page_down(10); assert_eq!(scroll.selected(), Some(19)); // Should stop at last item // Page up by 5 scroll.page_up(5); assert_eq!(scroll.selected(), Some(14)); // Page up by 20 (should go to start) scroll.page_up(20); assert_eq!(scroll.selected(), Some(0)); } #[test] fn test_page_navigation_empty() { let mut scroll = TableScroll::with_items(0); // Should do nothing on empty table scroll.page_down(5); assert_eq!(scroll.selected(), None); scroll.page_up(5); assert_eq!(scroll.selected(), None); } }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/tui/log_viewer_scrollview.rs
src/app/tui/log_viewer_scrollview.rs
use ratatui::{ buffer::Buffer, layout::{Constraint, Layout, Rect, Size}, style::{Color, Style}, symbols, text::{Line, Span}, widgets::{Block, Borders, Paragraph, StatefulWidget, Widget}, }; use serde_json; use tui_scrollview::{ScrollView, ScrollViewState}; use crate::app::storage::task::Task; const MAX_LINES_IN_MEMORY: usize = 10_000; /// A log viewer widget using tui-scrollview for efficient scrolling #[derive(Clone)] pub struct LogViewerScrollWidget { lines: Vec<String>, task_id: String, command: String, auto_scroll_enabled: bool, } impl LogViewerScrollWidget { /// Create a new log viewer with loaded content pub fn new(task: &Task) -> Self { let lines = Self::load_log_file(&task.log_path); Self { lines, task_id: task.id.clone(), command: Self::parse_command(&task.command), auto_scroll_enabled: true, // Auto-scroll is enabled by default } } /// Create with cached content pub fn with_cached_content(task: &Task, cached_lines: Vec<String>) -> Self { Self { lines: cached_lines, task_id: task.id.clone(), command: Self::parse_command(&task.command), auto_scroll_enabled: true, // Auto-scroll is enabled by default } } /// Load incremental content from a file pub fn load_incremental_content( task: &Task, mut existing_lines: Vec<String>, previous_size: u64, ) -> Self { // Try to read only the new content if let Ok(mut file) = std::fs::File::open(&task.log_path) { use std::io::{Read, Seek, SeekFrom}; // Seek to the previous end position if file.seek(SeekFrom::Start(previous_size)).is_ok() { let mut new_content = String::new(); if file.read_to_string(&mut new_content).is_ok() { // Append new lines to existing lines for line in new_content.lines() { existing_lines.push(line.to_string()); } // Apply memory limit if needed if existing_lines.len() > MAX_LINES_IN_MEMORY { let skip_count = existing_lines.len() - MAX_LINES_IN_MEMORY; existing_lines = existing_lines.into_iter().skip(skip_count).collect(); } } } } Self { lines: existing_lines, task_id: task.id.clone(), command: Self::parse_command(&task.command), auto_scroll_enabled: true, // Auto-scroll is enabled by default } } /// Load log file with memory limit fn load_log_file(path: &str) -> Vec<String> { match std::fs::read_to_string(path) { Ok(content) => { let lines: Vec<String> = content.lines().map(String::from).collect(); if lines.len() > MAX_LINES_IN_MEMORY { // Take last MAX_LINES_IN_MEMORY lines let skip_count = lines.len() - MAX_LINES_IN_MEMORY; lines.into_iter().skip(skip_count).collect() } else { lines } } Err(_) => vec!["Error: Could not read log file".to_string()], } } /// Get the lines for external use (caching) pub fn get_lines(&self) -> &[String] { &self.lines } /// Get the total line count pub fn get_lines_count(&self) -> usize { self.lines.len() } /// Check if auto-scroll is enabled pub fn is_auto_scroll_enabled(&self) -> bool { self.auto_scroll_enabled } /// Toggle auto-scroll mode pub fn toggle_auto_scroll(&mut self) { self.auto_scroll_enabled = !self.auto_scroll_enabled; } /// Disable auto-scroll (called when user manually scrolls) pub fn disable_auto_scroll(&mut self) { self.auto_scroll_enabled = false; } /// Enable auto-scroll pub fn enable_auto_scroll(&mut self) { self.auto_scroll_enabled = true; } /// Parse command from JSON format to readable string fn parse_command(command_json: &str) -> String { // Try to parse the JSON array format like ["npm","run","dev"] if let Ok(parsed) = serde_json::from_str::<Vec<String>>(command_json) { parsed.join(" ") } else { // If parsing fails, return the original string command_json.to_string() } } /// Create footer widget fn create_footer(&'_ self) -> Paragraph<'_> { let auto_scroll_status = if self.auto_scroll_enabled { "ON" } else { "OFF" }; let keybinds = format!( " j/k:Scroll h/l:H-Scroll g/G:Top/Bot C-d/u:Page C-f:Auto({auto_scroll_status}) Esc:Back q " ); Paragraph::new(keybinds).block( Block::default() .borders(Borders::ALL) .border_style(Style::default().fg(Color::LightMagenta)), ) } } impl StatefulWidget for LogViewerScrollWidget { type State = ScrollViewState; fn render(self, area: Rect, buf: &mut Buffer, state: &mut Self::State) { // Layout: content + footer (3) let chunks = Layout::vertical([Constraint::Min(5), Constraint::Length(3)]).split(area); // Render footer self.create_footer().render(chunks[1], buf); // Calculate line number width based on total lines let line_count = self.lines.len(); let line_number_width = line_count.to_string().len().max(1) + 1; // +1 for space let line_number_area_width = line_number_width as u16; // Calculate content size (lines count, max line width) let content_width = self .lines .iter() .map(|line| line.len() + line_number_width) .max() .unwrap_or(80) as u16; let content_height = self.lines.len() as u16; let content_size = Size::new(content_width, content_height); // Create a block for the content area with borders and title let title = format!(" {} - {} ", self.task_id, self.command); let content_block = Block::default() .borders(Borders::TOP | Borders::LEFT | Borders::RIGHT) .border_style(Style::default().fg(Color::LightMagenta)) .title(title); // Get the inner area for the scroll view let content_inner = content_block.inner(chunks[0]); // Render the content block content_block.render(chunks[0], buf); // Draw the separator line between content and footer // The separator is at the top of the footer block (chunks[1].y) if chunks[1].y > 0 { // Left connection: ├ buf[(chunks[0].x, chunks[1].y)].set_symbol(symbols::line::VERTICAL_RIGHT); // Horizontal line for x in chunks[0].x + 1..chunks[0].x + chunks[0].width - 1 { buf[(x, chunks[1].y)].set_symbol(symbols::line::HORIZONTAL); } // Right connection: ┤ buf[(chunks[0].x + chunks[0].width - 1, chunks[1].y)] .set_symbol(symbols::line::VERTICAL_LEFT); } // Create scroll view with content size and hide scrollbars let mut scroll_view = ScrollView::new(content_size) .scrollbars_visibility(tui_scrollview::ScrollbarVisibility::Never); // Create line numbers paragraph with dynamic width let line_numbers: Vec<Line> = self .lines .iter() .enumerate() .map(|(idx, _)| { let line_number = format!("{:>width$} ", idx + 1, width = line_number_width - 1); Line::from(Span::styled( line_number, Style::default().fg(Color::DarkGray), )) }) .collect(); let line_numbers_paragraph = Paragraph::new(line_numbers); // Create content paragraph let content_lines: Vec<Line> = self .lines .iter() .map(|line| Line::from(line.as_str())) .collect(); let content_paragraph = Paragraph::new(content_lines); // Render line numbers and content inside scroll view scroll_view.render_widget( line_numbers_paragraph, Rect::new(0, 0, line_number_area_width, content_height), ); scroll_view.render_widget( content_paragraph, Rect::new( line_number_area_width, 0, content_width.saturating_sub(line_number_area_width), content_height, ), ); // Render the scroll view in the inner content area scroll_view.render(content_inner, buf, state); } } #[cfg(test)] mod tests { use super::*; use ratatui::Terminal; use ratatui::backend::TestBackend; use std::io::Write; use tempfile::NamedTempFile; fn create_test_task(log_path: String) -> Task { Task { id: "test_task_12345678".to_string(), pid: 12345, pgid: Some(12345), command: r#"["echo","test"]"#.to_string(), env: None, cwd: None, status: crate::app::storage::task_status::TaskStatus::Running, exit_code: None, started_at: 1704109200, finished_at: None, log_path, } } #[test] fn test_basic_rendering() { let mut temp_file = NamedTempFile::new().unwrap(); writeln!(temp_file, "Line 1").unwrap(); writeln!(temp_file, "Line 2").unwrap(); writeln!(temp_file, "Line 3").unwrap(); temp_file.flush().unwrap(); let task = create_test_task(temp_file.path().to_string_lossy().to_string()); let widget = LogViewerScrollWidget::new(&task); let backend = TestBackend::new(80, 20); let mut terminal = Terminal::new(backend).unwrap(); let mut scroll_state = ScrollViewState::default(); terminal .draw(|f| { widget.render(f.area(), f.buffer_mut(), &mut scroll_state); }) .unwrap(); let buffer = terminal.backend().buffer(); let content = buffer_to_string(buffer); // Check title contains full task ID assert!(content.contains("test_task_12345678")); assert!(content.contains("echo test")); // Check footer assert!(content.contains("j/k:Scroll")); assert!(content.contains("h/l:H-Scroll")); assert!(content.contains("g/G:Top/Bot")); // Check content with line numbers (dynamic width) assert!(content.contains("1 Line 1")); assert!(content.contains("2 Line 2")); assert!(content.contains("3 Line 3")); } #[test] fn test_memory_limit() { let mut temp_file = NamedTempFile::new().unwrap(); // Write more than MAX_LINES_IN_MEMORY lines for i in 0..15000 { writeln!(temp_file, "Line {i}").unwrap(); } temp_file.flush().unwrap(); let task = create_test_task(temp_file.path().to_string_lossy().to_string()); let widget = LogViewerScrollWidget::new(&task); // Should only load MAX_LINES_IN_MEMORY lines assert_eq!(widget.get_lines_count(), MAX_LINES_IN_MEMORY); // Should contain the last lines let lines = widget.get_lines(); assert!(lines[0].contains("Line 5000")); // 15000 - 10000 = 5000 assert!(lines[9999].contains("Line 14999")); } #[test] fn test_error_handling() { let task = create_test_task("/non/existent/file.log".to_string()); let widget = LogViewerScrollWidget::new(&task); // Should have error message assert_eq!(widget.get_lines_count(), 1); assert_eq!(widget.get_lines()[0], "Error: Could not read log file"); } #[test] fn test_cached_content() { let task = create_test_task("/dummy/path.log".to_string()); let cached_lines = vec![ "Cached line 1".to_string(), "Cached line 2".to_string(), "Cached line 3".to_string(), ]; let widget = LogViewerScrollWidget::with_cached_content(&task, cached_lines.clone()); assert_eq!(widget.get_lines_count(), 3); assert_eq!(widget.get_lines(), &cached_lines); } #[test] fn test_incremental_loading() { let mut temp_file = NamedTempFile::new().unwrap(); writeln!(temp_file, "Initial line 1").unwrap(); writeln!(temp_file, "Initial line 2").unwrap(); temp_file.flush().unwrap(); let initial_size = temp_file.as_file().metadata().unwrap().len(); let task = create_test_task(temp_file.path().to_string_lossy().to_string()); let existing_lines = vec!["Initial line 1".to_string(), "Initial line 2".to_string()]; // Add more lines writeln!(temp_file, "New line 3").unwrap(); writeln!(temp_file, "New line 4").unwrap(); temp_file.flush().unwrap(); // Load incrementally let widget = LogViewerScrollWidget::load_incremental_content(&task, existing_lines, initial_size); assert_eq!(widget.get_lines_count(), 4); let lines = widget.get_lines(); assert_eq!(lines[0], "Initial line 1"); assert_eq!(lines[1], "Initial line 2"); assert_eq!(lines[2], "New line 3"); assert_eq!(lines[3], "New line 4"); } fn buffer_to_string(buffer: &ratatui::buffer::Buffer) -> String { let mut result = String::new(); for y in 0..buffer.area.height { for x in 0..buffer.area.width { let cell = &buffer[(x, y)]; result.push_str(cell.symbol()); } if y < buffer.area.height - 1 { result.push('\n'); } } result } #[test] fn test_auto_scroll_disabled_by_default() { let mut temp_file = NamedTempFile::new().unwrap(); writeln!(temp_file, "Line 1").unwrap(); temp_file.flush().unwrap(); let task = create_test_task(temp_file.path().to_string_lossy().to_string()); let widget = LogViewerScrollWidget::new(&task); // Auto-scroll should be enabled by default assert!(widget.is_auto_scroll_enabled()); } #[test] fn test_auto_scroll_toggle() { let mut temp_file = NamedTempFile::new().unwrap(); writeln!(temp_file, "Line 1").unwrap(); temp_file.flush().unwrap(); let task = create_test_task(temp_file.path().to_string_lossy().to_string()); let mut widget = LogViewerScrollWidget::new(&task); // Initially enabled assert!(widget.is_auto_scroll_enabled()); // Toggle off widget.toggle_auto_scroll(); assert!(!widget.is_auto_scroll_enabled()); // Toggle back on widget.toggle_auto_scroll(); assert!(widget.is_auto_scroll_enabled()); } #[test] fn test_auto_scroll_disable() { let mut temp_file = NamedTempFile::new().unwrap(); writeln!(temp_file, "Line 1").unwrap(); temp_file.flush().unwrap(); let task = create_test_task(temp_file.path().to_string_lossy().to_string()); let mut widget = LogViewerScrollWidget::new(&task); // Initially enabled assert!(widget.is_auto_scroll_enabled()); // Disable auto-scroll widget.disable_auto_scroll(); assert!(!widget.is_auto_scroll_enabled()); } #[test] fn test_auto_scroll_enable() { let mut temp_file = NamedTempFile::new().unwrap(); writeln!(temp_file, "Line 1").unwrap(); temp_file.flush().unwrap(); let task = create_test_task(temp_file.path().to_string_lossy().to_string()); let mut widget = LogViewerScrollWidget::new(&task); // Disable first widget.disable_auto_scroll(); assert!(!widget.is_auto_scroll_enabled()); // Re-enable auto-scroll widget.enable_auto_scroll(); assert!(widget.is_auto_scroll_enabled()); } #[test] fn test_footer_displays_auto_scroll_status() { let mut temp_file = NamedTempFile::new().unwrap(); writeln!(temp_file, "Line 1").unwrap(); temp_file.flush().unwrap(); let task = create_test_task(temp_file.path().to_string_lossy().to_string()); let mut widget = LogViewerScrollWidget::new(&task); let backend = TestBackend::new(80, 20); let mut terminal = Terminal::new(backend).unwrap(); let mut scroll_state = ScrollViewState::default(); // Test with auto-scroll enabled (default) terminal .draw(|f| { widget .clone() .render(f.area(), f.buffer_mut(), &mut scroll_state); }) .unwrap(); let buffer = terminal.backend().buffer(); let content = buffer_to_string(buffer); // Should show Auto(ON) in footer (default state) assert!(content.contains("C-f:Auto(ON)")); // Test with auto-scroll disabled widget.disable_auto_scroll(); let backend = TestBackend::new(80, 20); let mut terminal = Terminal::new(backend).unwrap(); terminal .draw(|f| { widget .clone() .render(f.area(), f.buffer_mut(), &mut scroll_state); }) .unwrap(); let buffer = terminal.backend().buffer(); let content = buffer_to_string(buffer); // Should show Auto(OFF) in footer (after disabling) assert!(content.contains("C-f:Auto(OFF)")); } }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/app/tui/mod.rs
src/app/tui/mod.rs
pub mod app; pub mod log_viewer_scrollview; pub mod process_details; pub mod table_state_scroll; pub mod task_list; use self::table_state_scroll::TableScroll; use crate::app::storage::task::Task; pub struct App { pub tasks: Vec<Task>, pub selected_index: usize, pub filter: TaskFilter, pub table_scroll: TableScroll, } #[derive(Debug, Clone, PartialEq)] pub enum TaskFilter { All, Running, Exited, Killed, } #[derive(Debug, Clone, PartialEq)] pub enum ViewMode { TaskList, LogView, ProcessDetails, } impl Default for App { fn default() -> Self { Self::new() } } impl App { pub fn new() -> Self { Self { tasks: Vec::new(), selected_index: 0, filter: TaskFilter::All, table_scroll: TableScroll::new(), } } pub fn with_tasks(tasks: Vec<Task>) -> Self { let table_scroll = TableScroll::with_items(tasks.len()); Self { tasks, selected_index: 0, filter: TaskFilter::All, table_scroll, } } pub fn with_tasks_and_scroll(tasks: Vec<Task>, scroll_offset: usize) -> Self { let mut table_scroll = TableScroll::with_items(tasks.len()); if !tasks.is_empty() && scroll_offset < tasks.len() { table_scroll.select(Some(scroll_offset)); } Self { tasks, selected_index: 0, filter: TaskFilter::All, table_scroll, } } }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/src/mcp/mod.rs
src/mcp/mod.rs
use async_trait::async_trait; use rust_mcp_sdk::macros::{JsonSchema, mcp_tool}; use rust_mcp_sdk::mcp_server::{ServerHandler, server_runtime}; use rust_mcp_sdk::schema::schema_utils::CallToolError; use rust_mcp_sdk::schema::{ CallToolRequest, CallToolResult, Implementation, InitializeResult, LATEST_PROTOCOL_VERSION, ListToolsRequest, ListToolsResult, ServerCapabilities, ServerCapabilitiesTools, TextContent, }; use rust_mcp_sdk::{McpServer, tool_box}; use rust_mcp_transport::{StdioTransport, TransportOptions}; use serde::{Deserialize, Serialize}; use std::path::PathBuf; use std::sync::{Arc, Mutex}; use tracing::{error, info}; use crate::app::commands; use crate::app::storage::task_repository; use rusqlite::Connection; #[mcp_tool( name = "ghost_run", description = "Run one or more commands as background processes managed by ghost" )] #[derive(Debug, Deserialize, Serialize, JsonSchema, Clone)] pub struct RunTool { /// Commands to run. Each string is a complete command with arguments. /// Example: ["sleep 10", "echo hello"] pub commands: Vec<String>, /// Working directory (defaults to current directory) pub cwd: Option<String>, /// Environment variables (KEY=VALUE format) pub env: Option<Vec<String>>, } #[mcp_tool( name = "ghost_list", description = "List all processes managed by ghost" )] #[derive(Debug, Deserialize, Serialize, JsonSchema, Clone)] pub struct ListTool { /// Filter by status (running, stopped, failed) pub status: Option<String>, } #[mcp_tool(name = "ghost_stop", description = "Stop a running process by ID")] #[derive(Debug, Deserialize, Serialize, JsonSchema, Clone)] pub struct StopTool { /// Process ID to stop pub id: String, /// Force kill the process (SIGKILL instead of SIGTERM) pub force: Option<bool>, } #[mcp_tool(name = "ghost_log", description = "Get logs for a specific process")] #[derive(Debug, Deserialize, Serialize, JsonSchema, Clone)] pub struct LogTool { /// Process ID to get logs for pub id: String, } #[mcp_tool( name = "ghost_status", description = "Check status of a specific process" )] #[derive(Debug, Deserialize, Serialize, JsonSchema, Clone)] pub struct StatusTool { /// Process ID to check status for pub id: String, } tool_box!( GhostTools, [RunTool, ListTool, StopTool, LogTool, StatusTool] ); pub struct GhostServerHandler { conn: Arc<Mutex<Connection>>, } impl GhostServerHandler { pub fn new(conn: Connection) -> Self { Self { conn: Arc::new(Mutex::new(conn)), } } } #[async_trait] impl ServerHandler for GhostServerHandler { async fn handle_list_tools_request( &self, _request: ListToolsRequest, _runtime: Arc<dyn McpServer>, ) -> Result<ListToolsResult, rust_mcp_sdk::schema::RpcError> { Ok(ListToolsResult { tools: GhostTools::tools(), meta: None, next_cursor: None, }) } async fn handle_call_tool_request( &self, request: CallToolRequest, _runtime: Arc<dyn McpServer>, ) -> Result<CallToolResult, CallToolError> { let params = request.params.clone(); let tool = GhostTools::try_from(params)?; match tool { GhostTools::RunTool(t) => { if t.commands.is_empty() { return Err(CallToolError::from_message( "No commands specified".to_string(), )); } let cwd = t.cwd.map(PathBuf::from); let env = t.env.unwrap_or_default(); let conn = self.conn.lock().unwrap(); let results = commands::spawn_multi(&conn, t.commands, cwd, env, false); // Collect successful tasks and errors let mut tasks = Vec::new(); let mut errors = Vec::new(); for spawn_result in results { match spawn_result.result { Ok(info) => match task_repository::get_task(&conn, &info.id) { Ok(task) => tasks.push(task), Err(e) => errors.push(format!( "Failed to get task for '{}': {e}", spawn_result.command_str )), }, Err(e) => { errors.push(format!( "Failed to spawn '{}': {e}", spawn_result.command_str )); } } } // Build response with tasks and any errors let response = serde_json::json!({ "tasks": tasks, "errors": errors }); let result = serde_json::to_string_pretty(&response) .map_err(|e| CallToolError::from_message(format!("JSON error: {e}")))?; Ok(CallToolResult::text_content(vec![TextContent::new( result, None, None, )])) } GhostTools::ListTool(t) => { let conn = self.conn.lock().unwrap(); // Prepare status filter let tasks = commands::list(&conn, t.status, false).map_err(|e| { CallToolError::from_message(format!("Failed to list tasks: {e}")) })?; let result = serde_json::to_string_pretty(&tasks) .map_err(|e| CallToolError::from_message(format!("JSON error: {e}")))?; Ok(CallToolResult::text_content(vec![TextContent::new( result, None, None, )])) } GhostTools::StopTool(t) => { let conn = self.conn.lock().unwrap(); commands::stop(&conn, &t.id, t.force.unwrap_or(false), false) .map_err(|e| CallToolError::from_message(format!("Failed to stop: {e}")))?; Ok(CallToolResult::text_content(vec![TextContent::new( format!("Process {} stopped successfully", t.id), None, None, )])) } GhostTools::LogTool(t) => { let conn = self.conn.lock().unwrap(); let task = task_repository::get_task(&conn, &t.id) .map_err(|e| CallToolError::from_message(format!("Failed to get task: {e}")))?; let log_content = std::fs::read_to_string(&task.log_path) .map_err(|e| CallToolError::from_message(format!("Failed to read log: {e}")))?; Ok(CallToolResult::text_content(vec![TextContent::new( log_content, None, None, )])) } GhostTools::StatusTool(t) => { let conn = self.conn.lock().unwrap(); let task = commands::status(&conn, &t.id, false).map_err(|e| { CallToolError::from_message(format!("Failed to get status: {e}")) })?; let result = serde_json::to_string_pretty(&task) .map_err(|e| CallToolError::from_message(format!("JSON error: {e}")))?; Ok(CallToolResult::text_content(vec![TextContent::new( result, None, None, )])) } } } } pub async fn run_stdio_server(conn: Connection) -> Result<(), Box<dyn std::error::Error>> { info!("Ghost MCP server starting..."); let server_details = InitializeResult { server_info: Implementation { name: "ghost-mcp".into(), title: Some("Ghost MCP Server".into()), version: env!("CARGO_PKG_VERSION").into(), }, capabilities: ServerCapabilities { tools: Some(ServerCapabilitiesTools { list_changed: None }), ..Default::default() }, meta: None, instructions: Some( "Ghost MCP server for managing background processes. Use tools to run, list, stop, check status, cleanup old tasks, and view logs for processes.".into() ), protocol_version: LATEST_PROTOCOL_VERSION.into(), }; let transport = match StdioTransport::new(TransportOptions::default()) { Ok(t) => t, Err(e) => { error!("Failed to create transport: {e}"); return Err(e.into()); } }; let handler = GhostServerHandler::new(conn); let server = server_runtime::create_server(server_details, transport, handler); info!("Ghost MCP server initialized, waiting for connections..."); if let Err(e) = server.start().await { error!("MCP server error: {e}"); return Err(e.into()); } info!("Ghost MCP server shutting down"); Ok(()) }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/tests/tcp_server_helper.rs
tests/tcp_server_helper.rs
use std::net::TcpListener; use std::thread; use std::time::Duration; fn main() { // Get port from command line args, default to 0 (random port) let port = std::env::args() .nth(1) .and_then(|s| s.parse::<u16>().ok()) .unwrap_or(0); let listener = TcpListener::bind(format!("127.0.0.1:{port}")).expect("Failed to bind TCP listener"); let addr = listener.local_addr().unwrap(); println!("Listening on {addr}"); // Keep the server running for 30 seconds thread::sleep(Duration::from_secs(30)); }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/tests/mcp_server_tests.rs
tests/mcp_server_tests.rs
use std::sync::{Arc, Mutex}; use std::time::Duration; use async_trait::async_trait; use ghost::app::config::Config; use ghost::app::storage::{self, Task, TaskStatus}; use ghost::mcp::GhostServerHandler; use rusqlite::Connection; use rust_mcp_sdk::McpServer; use rust_mcp_sdk::auth::AuthInfo; use rust_mcp_sdk::mcp_server::ServerHandler; use rust_mcp_sdk::schema::schema_utils::{ClientMessage, MessageFromServer, ServerMessage}; use rust_mcp_sdk::schema::{ CallToolRequest, CallToolRequestParams, CallToolResult, ContentBlock, Implementation, InitializeRequestParams, InitializeResult, LATEST_PROTOCOL_VERSION, ServerCapabilities, }; use serde_json::{Value, json}; use tempfile::TempDir; use tokio::sync::RwLockReadGuard; struct McpTestContext { _temp_dir: TempDir, config: Config, original_data_dir: Option<String>, } impl McpTestContext { fn new() -> Self { let temp_dir = TempDir::new().expect("failed to create temp dir"); let config = Config::with_data_dir(temp_dir.path().to_path_buf()); config .ensure_directories() .expect("failed to init directories"); let original = std::env::var("GHOST_DATA_DIR").ok(); unsafe { std::env::set_var("GHOST_DATA_DIR", config.data_dir.clone()); } Self { _temp_dir: temp_dir, config, original_data_dir: original, } } fn connection(&self) -> Connection { storage::database::init_database_with_config(Some(self.config.clone())) .expect("failed to init database") } fn log_path(&self, filename: &str) -> std::path::PathBuf { self.config.log_dir.join(filename) } } impl Drop for McpTestContext { fn drop(&mut self) { if let Some(original) = &self.original_data_dir { unsafe { std::env::set_var("GHOST_DATA_DIR", original); } } else { unsafe { std::env::remove_var("GHOST_DATA_DIR"); } } } } #[derive(Debug)] struct DummyRuntime { server_info: InitializeResult, client_info: Mutex<Option<InitializeRequestParams>>, } impl Default for DummyRuntime { fn default() -> Self { Self { server_info: InitializeResult { server_info: Implementation { name: "ghost-test".into(), title: Some("Ghost Test Runtime".into()), version: "0.0.0".into(), }, capabilities: ServerCapabilities::default(), instructions: None, meta: None, protocol_version: LATEST_PROTOCOL_VERSION.into(), }, client_info: Mutex::new(None), } } } #[async_trait] impl McpServer for DummyRuntime { async fn start(self: Arc<Self>) -> rust_mcp_sdk::error::SdkResult<()> { Ok(()) } async fn set_client_details( &self, client_details: InitializeRequestParams, ) -> rust_mcp_sdk::error::SdkResult<()> { *self.client_info.lock().unwrap() = Some(client_details); Ok(()) } fn server_info(&self) -> &InitializeResult { &self.server_info } fn client_info(&self) -> Option<InitializeRequestParams> { self.client_info.lock().unwrap().clone() } async fn wait_for_initialization(&self) {} async fn send( &self, _message: MessageFromServer, _request_id: Option<rust_mcp_sdk::schema::RequestId>, _request_timeout: Option<Duration>, ) -> rust_mcp_sdk::error::SdkResult<Option<ClientMessage>> { Ok(None) } async fn send_batch( &self, _messages: Vec<ServerMessage>, _request_timeout: Option<Duration>, ) -> rust_mcp_sdk::error::SdkResult<Option<Vec<ClientMessage>>> { Ok(None) } async fn stderr_message(&self, _message: String) -> rust_mcp_sdk::error::SdkResult<()> { Ok(()) } async fn auth_info(&self) -> RwLockReadGuard<'_, Option<AuthInfo>> { unimplemented!() } async fn auth_info_cloned(&self) -> Option<AuthInfo> { unimplemented!() } async fn update_auth_info(&self, _auth_info: Option<AuthInfo>) { unimplemented!() } } fn make_call_request(name: &str, args: Value) -> CallToolRequest { let arguments = match args { Value::Object(map) => map, Value::Null => serde_json::Map::new(), other => panic!("tool arguments must be an object, got {other:?}"), }; CallToolRequest::new(CallToolRequestParams { name: name.to_string(), arguments: Some(arguments), }) } fn insert_task_with_log(ctx: &McpTestContext, conn: &Connection, id: &str, log_contents: &str) { let command = vec!["echo".to_string(), "ghost".to_string()]; let log_path = ctx.log_path(&format!("{id}.log")); std::fs::write(&log_path, log_contents).expect("failed to write log file"); storage::insert_task( conn, id, 12345, Some(12345), &command, None, None, &log_path, ) .expect("failed to insert task"); storage::update_task_status(conn, id, TaskStatus::Exited, Some(0)) .expect("failed to update task status"); } fn text_content(result: &CallToolResult) -> String { let block = result .content .first() .expect("tool result should contain content"); match block { ContentBlock::TextContent(text) => text.text.clone(), other => panic!("unexpected content block: {other:?}"), } } async fn call_tool(handler: &GhostServerHandler, name: &str, args: Value) -> CallToolResult { handler .handle_call_tool_request( make_call_request(name, args), Arc::new(DummyRuntime::default()), ) .await .unwrap_or_else(|_| panic!("{name} call should succeed")) } #[tokio::test] async fn ghost_list_returns_all_tasks() { let ctx = McpTestContext::new(); let conn = ctx.connection(); insert_task_with_log(&ctx, &conn, "task-alpha", "alpha log"); insert_task_with_log(&ctx, &conn, "task-beta", "beta log"); let handler = GhostServerHandler::new(conn); let response = call_tool(&handler, "ghost_list", json!({})).await; let payload = text_content(&response); let mut tasks: Vec<Task> = serde_json::from_str(&payload).expect("valid task list JSON"); tasks.sort_by(|a, b| a.id.cmp(&b.id)); let ids: Vec<_> = tasks.into_iter().map(|t| t.id).collect(); assert_eq!(ids, vec!["task-alpha", "task-beta"]); } #[tokio::test] async fn ghost_status_returns_requested_task() { let ctx = McpTestContext::new(); let conn = ctx.connection(); insert_task_with_log(&ctx, &conn, "task-status", "status log"); let handler = GhostServerHandler::new(conn); let response = call_tool(&handler, "ghost_status", json!({ "id": "task-status" })).await; let payload = text_content(&response); let task: Task = serde_json::from_str(&payload).expect("valid task JSON"); assert_eq!(task.id, "task-status"); assert_eq!(task.status, TaskStatus::Exited); } #[tokio::test] async fn ghost_log_returns_task_log_contents() { let ctx = McpTestContext::new(); let conn = ctx.connection(); let log_body = "line 1\nline 2\n"; insert_task_with_log(&ctx, &conn, "task-log", log_body); let handler = GhostServerHandler::new(conn); let response = call_tool(&handler, "ghost_log", json!({ "id": "task-log" })).await; let payload = text_content(&response); assert_eq!(payload, log_body); } #[tokio::test] async fn ghost_run_multiple_commands() { let ctx = McpTestContext::new(); let conn = ctx.connection(); let handler = GhostServerHandler::new(conn); let run_result = call_tool( &handler, "ghost_run", json!({ "commands": ["sleep 5", "sleep 5"] }), ) .await; let run_payload = text_content(&run_result); let response: Value = serde_json::from_str(&run_payload).expect("valid run response JSON"); let tasks = response["tasks"].as_array().expect("tasks array"); let errors = response["errors"].as_array().expect("errors array"); assert_eq!(tasks.len(), 2, "should spawn 2 tasks"); assert!(errors.is_empty(), "should have no errors"); // Verify each task has unique ID and correct command let task1: Task = serde_json::from_value(tasks[0].clone()).expect("valid task JSON"); let task2: Task = serde_json::from_value(tasks[1].clone()).expect("valid task JSON"); assert_ne!(task1.id, task2.id, "tasks should have different IDs"); assert!(task1.command.contains("sleep")); assert!(task2.command.contains("sleep")); // Cleanup: stop running processes (ignore errors for already-exited processes) for task in [&task1, &task2] { let _ = handler .handle_call_tool_request( make_call_request("ghost_stop", json!({ "id": task.id, "force": true })), Arc::new(DummyRuntime::default()), ) .await; } } #[tokio::test] async fn ghost_run_empty_commands_returns_error() { let ctx = McpTestContext::new(); let conn = ctx.connection(); let handler = GhostServerHandler::new(conn); let result = handler .handle_call_tool_request( make_call_request("ghost_run", json!({ "commands": [] })), Arc::new(DummyRuntime::default()), ) .await; assert!(result.is_err(), "empty commands should return error"); } #[tokio::test] async fn ghost_run_and_stop_lifecycle() { let ctx = McpTestContext::new(); let conn = ctx.connection(); let handler = GhostServerHandler::new(conn); let run_result = call_tool( &handler, "ghost_run", json!({ "commands": ["sleep 5"], "env": [] }), ) .await; let run_payload = text_content(&run_result); let response: Value = serde_json::from_str(&run_payload).expect("valid run response JSON"); let tasks = response["tasks"].as_array().expect("tasks array"); assert_eq!(tasks.len(), 1); let task: Task = serde_json::from_value(tasks[0].clone()).expect("valid task JSON"); let task_id = task.id.clone(); assert_eq!(task.status, TaskStatus::Running); assert!(task.command.contains("sleep")); assert!(!task.log_path.is_empty()); tokio::time::sleep(Duration::from_millis(100)).await; let stop_id = task_id.clone(); let stop_result = call_tool( &handler, "ghost_stop", json!({ "id": stop_id, "force": true }), ) .await; let stop_message = text_content(&stop_result); assert!(stop_message.contains("stopped successfully")); tokio::time::sleep(Duration::from_millis(100)).await; let status_result = call_tool(&handler, "ghost_status", json!({ "id": task_id })).await; let status_payload = text_content(&status_result); let updated_task: Task = serde_json::from_str(&status_payload).expect("valid status task JSON"); assert_eq!(updated_task.status, TaskStatus::Killed); }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/tests/port_detector_test.rs
tests/port_detector_test.rs
use ghost::app::port_detector::detect_listening_ports; use std::process::Command; use std::thread; use std::time::Duration; #[test] fn test_detect_listening_ports_with_server() { // Compile the TCP server helper let output = Command::new("rustc") .args([ "tests/tcp_server_helper.rs", "-o", "target/tcp_server_helper", ]) .output() .expect("Failed to compile TCP server helper"); if !output.status.success() { panic!( "Failed to compile TCP server helper: {}", String::from_utf8_lossy(&output.stderr) ); } // Start the TCP server let mut child = Command::new("target/tcp_server_helper") .spawn() .expect("Failed to start TCP server"); let server_pid = child.id(); // Give the server more time to start and establish listening state thread::sleep(Duration::from_millis(1000)); // Test port detection let result = detect_listening_ports(server_pid); assert!(result.is_ok()); let ports = result.unwrap(); assert!( !ports.is_empty(), "Should detect at least one listening port for PID {server_pid}" ); // Verify port information let port = &ports[0]; assert_eq!(port.protocol, "tcp"); assert!(port.local_addr.contains(":")); assert_eq!(port.state, "LISTEN"); // Clean up let _ = child.kill(); let _ = child.wait(); } #[test] fn test_detect_listening_ports_no_ports() { // Use a simple command that doesn't listen on any ports let mut sleep_process = Command::new("sleep") .arg("5") .spawn() .expect("Failed to start sleep process"); let sleep_pid = sleep_process.id(); // Test port detection let result = detect_listening_ports(sleep_pid); assert!(result.is_ok()); let ports = result.unwrap(); assert!( ports.is_empty(), "Should not detect any listening ports for sleep command" ); // Clean up let _ = sleep_process.kill(); let _ = sleep_process.wait(); } #[test] fn test_detect_listening_ports_invalid_pid() { // Use a PID that doesn't exist let result = detect_listening_ports(99999); assert!(result.is_ok()); let ports = result.unwrap(); assert!( ports.is_empty(), "Should return empty vec for non-existent PID" ); }
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
false
skanehira/ghost
https://github.com/skanehira/ghost/blob/2608a5ac4d31e61433a039387325199a22701da6/tests/tui_tests.rs
tests/tui_tests.rs
use ghost::app::config::Config; use ghost::app::storage::task::Task; use ghost::app::storage::task_status::TaskStatus; use ghost::app::tui::{App, TaskFilter, ViewMode}; use pretty_assertions::assert_eq; use ratatui::{Terminal, backend::TestBackend}; use std::fs; use tempfile::TempDir; /// Helper function to load expected output from file fn load_expected(filename: &str) -> String { let path = format!("tests/expected/{filename}"); fs::read_to_string(&path).unwrap_or_else(|_| panic!("Failed to read expected file: {path}")) } /// Helper struct to manage test environment with temporary data directory struct TestEnvironment { _temp_dir: TempDir, original_env: Option<String>, pub config: Config, } impl TestEnvironment { fn new() -> Self { // Save original GHOST_DATA_DIR if set let original_env = std::env::var("GHOST_DATA_DIR").ok(); // Create temporary directory for test data let temp_dir = TempDir::new().expect("Failed to create temp dir"); // Set GHOST_DATA_DIR to temp directory unsafe { std::env::set_var("GHOST_DATA_DIR", temp_dir.path()); } let config = Config::with_data_dir(temp_dir.path().to_path_buf()); Self { _temp_dir: temp_dir, original_env, config, } } } impl Drop for TestEnvironment { fn drop(&mut self) { // Restore original GHOST_DATA_DIR or remove it unsafe { match &self.original_env { Some(val) => std::env::set_var("GHOST_DATA_DIR", val), None => std::env::remove_var("GHOST_DATA_DIR"), } } } } /// Helper function to create test tasks fn create_test_tasks() -> Vec<Task> { vec![ Task { id: "abc12345-6789-1234-5678-123456789abc".to_string(), pid: 12345, pgid: Some(12345), command: r#"["echo","hello"]"#.to_string(), env: None, cwd: None, status: TaskStatus::Running, exit_code: None, started_at: 1704109200, // 2024-01-01 10:00 UTC finished_at: None, log_path: "/tmp/test.log".to_string(), }, Task { id: "def67890-1234-5678-9abc-def123456789".to_string(), pid: 67890, pgid: Some(67890), command: r#"["cargo","build"]"#.to_string(), env: None, cwd: None, status: TaskStatus::Exited, exit_code: Some(0), started_at: 1704107400, // 2024-01-01 09:30 UTC finished_at: Some(1704107460), // 2024-01-01 09:31 UTC log_path: "/tmp/test2.log".to_string(), }, Task { id: "ghi11111-5678-9abc-def1-23456789abcd".to_string(), pid: 11111, pgid: Some(11111), command: r#"["python","script.py"]"#.to_string(), env: None, cwd: None, status: TaskStatus::Killed, exit_code: Some(1), started_at: 1704105600, // 2024-01-01 09:00 UTC finished_at: Some(1704105660), // 2024-01-01 09:01 UTC log_path: "/tmp/test3.log".to_string(), }, ] } /// Helper function to convert buffer to string fn buffer_to_string(buffer: &ratatui::buffer::Buffer) -> String { let mut result = String::new(); for y in 0..buffer.area.height { for x in 0..buffer.area.width { let cell = &buffer[(x, y)]; result.push_str(cell.symbol()); } if y < buffer.area.height - 1 { result.push('\n'); } } result } /// Helper function to normalize whitespace for comparison fn normalize_buffer_output(output: &str) -> String { output .lines() .map(|line| line.trim_end()) // Remove trailing whitespace .collect::<Vec<_>>() .join("\n") } /// Helper function to normalize dynamic values like runtime for comparison fn normalize_dynamic_output(output: &str) -> String { // Replace runtime patterns like "13286h 37m 52s" with a placeholder let re = regex::Regex::new(r"\d+h \d+m \d+s|\d+m \d+s|\d+s").unwrap(); re.replace_all(output, "<RUNTIME>").to_string() } /// Helper function to remove status line from process details output fn normalize_without_status_line(output: &str) -> String { output .lines() .filter(|line| !line.contains("│Status:")) .collect::<Vec<_>>() .join("\n") } #[test] fn test_empty_task_list_display() { let backend = TestBackend::new(75, 12); let mut terminal = Terminal::new(backend).unwrap(); let mut app = App::new(); terminal .draw(|f| { app.render_task_list(f, f.area()); }) .unwrap(); let buffer_output = buffer_to_string(terminal.backend().buffer()); let normalized_output = normalize_buffer_output(&buffer_output); let expected = load_expected("task_list_empty.txt"); let normalized_expected = normalize_buffer_output(&expected); assert_eq!( normalized_output, normalized_expected, "Empty task list display does not match expected output" ); } #[test] fn test_task_list_with_tasks_display() { let backend = TestBackend::new(75, 12); let mut terminal = Terminal::new(backend).unwrap(); let tasks = create_test_tasks(); let mut app = App::with_tasks(tasks); terminal .draw(|f| { app.render_task_list(f, f.area()); }) .unwrap(); let buffer_output = buffer_to_string(terminal.backend().buffer()); let normalized_output = normalize_buffer_output(&buffer_output); let expected = load_expected("task_list_with_tasks.txt"); let normalized_expected = normalize_buffer_output(&expected); assert_eq!( normalized_output, normalized_expected, "Task list with tasks display does not match expected output" ); } #[test] fn test_task_list_selection() { let backend = TestBackend::new(75, 12); let mut terminal = Terminal::new(backend).unwrap(); let tasks = create_test_tasks(); let mut app = App::with_tasks(tasks); app.selected_index = 1; // Select second task terminal .draw(|f| { app.render_task_list(f, f.area()); }) .unwrap(); let buffer_output = buffer_to_string(terminal.backend().buffer()); // Check that the output contains the tasks with truncated IDs due to column width // The selection highlighting will be tested once we have the expected file assert!(buffer_output.contains("abc123")); assert!(buffer_output.contains("def678")); assert!(buffer_output.contains("ghi111")); } #[test] fn test_task_filter_display() { let backend = TestBackend::new(75, 8); let mut terminal = Terminal::new(backend).unwrap(); let tasks = create_test_tasks(); let mut app = App::with_tasks(tasks); app.filter = TaskFilter::Running; terminal .draw(|f| { app.render_task_list(f, f.area()); }) .unwrap(); let buffer_output = buffer_to_string(terminal.backend().buffer()); // Check that the filter is displayed in the header assert!(buffer_output.contains("[Filter: Running]")); } #[test] fn test_footer_keybinds_display() { let backend = TestBackend::new(75, 12); let mut terminal = Terminal::new(backend).unwrap(); let mut app = App::new(); terminal .draw(|f| { app.render_task_list(f, f.area()); }) .unwrap(); let buffer_output = buffer_to_string(terminal.backend().buffer()); let normalized_output = normalize_buffer_output(&buffer_output); let expected = load_expected("task_list_empty.txt"); let normalized_expected = normalize_buffer_output(&expected); assert_eq!( normalized_output, normalized_expected, "Footer keybinds display does not match expected output" ); } #[test] fn test_footer_keybinds_with_tasks() { let backend = TestBackend::new(75, 12); let mut terminal = Terminal::new(backend).unwrap(); let tasks = create_test_tasks(); let mut app = App::with_tasks(tasks); terminal .draw(|f| { app.render_task_list(f, f.area()); }) .unwrap(); let buffer_output = buffer_to_string(terminal.backend().buffer()); let normalized_output = normalize_buffer_output(&buffer_output); let expected = load_expected("task_list_with_tasks.txt"); let normalized_expected = normalize_buffer_output(&expected); assert_eq!( normalized_output, normalized_expected, "Footer keybinds with tasks display does not match expected output" ); } #[test] fn test_footer_contains_keybinds() { let backend = TestBackend::new(75, 8); let mut terminal = Terminal::new(backend).unwrap(); let mut app = App::new(); terminal .draw(|f| { app.render_task_list(f, f.area()); }) .unwrap(); let buffer_output = buffer_to_string(terminal.backend().buffer()); // Check that footer contains essential keybinds assert!(buffer_output.contains("j/k:Move")); assert!(buffer_output.contains("l:Log")); assert!(buffer_output.contains("s/C-k:Stop")); assert!(buffer_output.contains("q:Quit")); assert!(buffer_output.contains("g/G:Top/Bot")); } #[test] fn test_task_list_vertical_layout() { let backend = TestBackend::new(75, 10); let mut terminal = Terminal::new(backend).unwrap(); let tasks = create_test_tasks(); let mut app = App::with_tasks(tasks); terminal .draw(|f| { app.render_task_list(f, f.area()); }) .unwrap(); let buffer_output = buffer_to_string(terminal.backend().buffer()); // Check that the layout has proper structure with separate blocks // The layout should be: // 1. Content block (variable height) // 2. Footer block (3 lines) let lines: Vec<&str> = buffer_output.lines().collect(); // Content block should contain title and table assert!(lines[0].starts_with("┌")); // Content top border assert!(lines[0].contains("Ghost v")); assert!(lines[1].contains("ID")); assert!(lines[1].contains("PID")); assert!(lines[1].contains("Status")); // Footer block should be separate assert!(lines[lines.len() - 3].starts_with("├")); // Footer top border assert!(lines[lines.len() - 2].contains("j/k:Move")); assert!(lines[lines.len() - 2].contains("l:Log")); assert!(lines[lines.len() - 1].starts_with("└")); // Footer bottom border } #[test] fn test_table_scroll_functionality() { use crossterm::event::{KeyCode, KeyEvent, KeyModifiers}; use ghost::app::tui::app::TuiApp; let env = TestEnvironment::new(); let mut app = TuiApp::new_with_config(env.config.clone()).unwrap(); // Create many tasks to enable scrolling let mut tasks = Vec::new(); for i in 0..20 { tasks.push(Task { id: format!("task_{i:03}"), pid: 1000 + i as u32, pgid: Some(1000 + i), command: format!(r#"["echo","task_{i}"]"#), env: None, cwd: None, status: TaskStatus::Running, exit_code: None, started_at: 1704109200 + i as i64, finished_at: None, log_path: format!("/tmp/test_{i}.log"), }); } app.tasks = tasks; app.table_scroll.set_total_items(20); app.set_selected_index(0); // Test scrolling down let key_j = KeyEvent::new(KeyCode::Char('j'), KeyModifiers::NONE); app.handle_key(key_j).unwrap(); // After first j, selection should move but scroll should not change yet assert_eq!(app.selected_index(), 1); assert_eq!(app.table_scroll_offset(), 0); // Test scrolling up let key_k = KeyEvent::new(KeyCode::Char('k'), KeyModifiers::NONE); app.handle_key(key_k).unwrap(); assert_eq!(app.selected_index(), 0); assert_eq!(app.table_scroll_offset(), 0); // Test going to bottom triggers scroll let key_shift_g = KeyEvent::new(KeyCode::Char('G'), KeyModifiers::NONE); app.handle_key(key_shift_g).unwrap(); assert_eq!(app.selected_index(), 19); // Last task // Scroll offset should be adjusted to show the selected item assert!(app.table_scroll_offset() > 0); // Test going to top resets scroll let key_g = KeyEvent::new(KeyCode::Char('g'), KeyModifiers::NONE); app.handle_key(key_g).unwrap(); assert_eq!(app.selected_index(), 0); assert_eq!(app.table_scroll_offset(), 0); } #[test] fn test_table_scroll_display() { let backend = TestBackend::new(75, 10); let mut terminal = Terminal::new(backend).unwrap(); // Create many tasks let mut tasks = Vec::new(); for i in 0..15 { tasks.push(Task { id: format!("task_{i:03}"), pid: 1000 + i as u32, pgid: Some(1000 + i), command: format!(r#"["echo","task_{i}"]"#), env: None, cwd: None, status: TaskStatus::Running, exit_code: None, started_at: 1704109200 + i as i64, finished_at: None, log_path: format!("/tmp/test_{i}.log"), }); } let mut app = App::with_tasks_and_scroll(tasks, 5); // Start scrolled down 5 rows terminal .draw(|f| { app.render_task_list(f, f.area()); }) .unwrap(); let buffer_output = buffer_to_string(terminal.backend().buffer()); // Should show tasks starting from around task_005 due to scroll offset // With full task IDs, the first visible task should be task_005 or later assert!(buffer_output.contains("task_")); // Should not show first few tasks due to scrolling // Check that task_000 is not visible (it would be scrolled out of view) assert!(!buffer_output.contains(" task_000 ")); } #[test] fn test_task_termination_keys() { use crossterm::event::{KeyCode, KeyEvent, KeyModifiers}; use ghost::app::storage::task_status::TaskStatus; use ghost::app::tui::app::TuiApp; let env = TestEnvironment::new(); let mut app = TuiApp::new_with_config(env.config.clone()).unwrap(); // Add a running task let tasks = vec![Task { id: "test_task".to_string(), pid: 12345, pgid: Some(12345), command: r#"["echo","test"]"#.to_string(), env: None, cwd: None, status: TaskStatus::Running, exit_code: None, started_at: 1704109200, finished_at: None, log_path: "/tmp/test.log".to_string(), }]; app.tasks = tasks; app.table_scroll.set_total_items(1); app.set_selected_index(0); // Test 's' key for SIGTERM let key_s = KeyEvent::new(KeyCode::Char('s'), KeyModifiers::NONE); // We can't actually test the signal sending, but we can test that the handler is called let result = app.handle_key(key_s); assert!(result.is_ok()); // Test Ctrl+K for SIGKILL let key_ctrl_k = KeyEvent::new(KeyCode::Char('k'), KeyModifiers::CONTROL); let result = app.handle_key(key_ctrl_k); assert!(result.is_ok()); } #[test] fn test_task_filter_cycling_with_tab() { use crossterm::event::{KeyCode, KeyEvent, KeyModifiers}; use ghost::app::tui::app::TuiApp; let env = TestEnvironment::new(); let mut app = TuiApp::new_with_config(env.config.clone()).unwrap(); // Add tasks with different statuses let tasks = vec![ Task { id: "running_task".to_string(), pid: 12345, pgid: Some(12345), command: r#"["echo","running"]"#.to_string(), env: None, cwd: None, status: TaskStatus::Running, exit_code: None, started_at: 1704109200, finished_at: None, log_path: "/tmp/running.log".to_string(), }, Task { id: "exited_task".to_string(), pid: 12346, pgid: Some(12346), command: r#"["echo","exited"]"#.to_string(), env: None, cwd: None, status: TaskStatus::Exited, exit_code: Some(0), started_at: 1704109200, finished_at: Some(1704109260), log_path: "/tmp/exited.log".to_string(), }, Task { id: "killed_task".to_string(), pid: 12347, pgid: Some(12347), command: r#"["echo","killed"]"#.to_string(), env: None, cwd: None, status: TaskStatus::Killed, exit_code: Some(1), started_at: 1704109200, finished_at: Some(1704109260), log_path: "/tmp/killed.log".to_string(), }, ]; app.tasks = tasks; app.table_scroll.set_total_items(3); // Test initial filter is All assert_eq!(app.filter, TaskFilter::All); // Press Tab to cycle to Running let key_tab = KeyEvent::new(KeyCode::Tab, KeyModifiers::NONE); app.handle_key(key_tab).unwrap(); assert_eq!(app.filter, TaskFilter::Running); // Press Tab to cycle to Exited app.handle_key(key_tab).unwrap(); assert_eq!(app.filter, TaskFilter::Exited); // Press Tab to cycle to Killed app.handle_key(key_tab).unwrap(); assert_eq!(app.filter, TaskFilter::Killed); // Press Tab to cycle back to All app.handle_key(key_tab).unwrap(); assert_eq!(app.filter, TaskFilter::All); } #[test] fn test_process_details_navigation() { use crossterm::event::{KeyCode, KeyEvent, KeyModifiers}; use ghost::app::tui::app::TuiApp; let env = TestEnvironment::new(); let mut app = TuiApp::new_with_config(env.config.clone()).unwrap(); // Add a test task with environment variables let tasks = vec![Task { id: "12345678-1234-1234-1234-123456789012".to_string(), pid: 1234, pgid: Some(1234), command: r#"["npm", "run", "dev"]"#.to_string(), env: Some(r#"[["NODE_ENV","development"],["PORT","3000"]]"#.to_string()), cwd: Some("/home/user/project".to_string()), status: TaskStatus::Running, exit_code: None, started_at: 1000000000, finished_at: None, log_path: "/tmp/ghost/logs/12345678.log".to_string(), }]; app.tasks = tasks; app.table_scroll.set_total_items(1); // Initial view should be TaskList assert_eq!(app.view_mode, ViewMode::TaskList); // Press Enter to view process details let key_enter = KeyEvent::new(KeyCode::Enter, KeyModifiers::NONE); app.handle_key(key_enter).unwrap(); assert_eq!(app.view_mode, ViewMode::ProcessDetails); assert_eq!( app.selected_task_id, Some("12345678-1234-1234-1234-123456789012".to_string()) ); // Press Esc to go back to task list let key_esc = KeyEvent::new(KeyCode::Esc, KeyModifiers::NONE); app.handle_key(key_esc).unwrap(); assert_eq!(app.view_mode, ViewMode::TaskList); assert!(!app.should_quit()); // Go back to process details and test q key app.handle_key(key_enter).unwrap(); assert_eq!(app.view_mode, ViewMode::ProcessDetails); // Press q to quit let key_q = KeyEvent::new(KeyCode::Char('q'), KeyModifiers::NONE); app.handle_key(key_q).unwrap(); assert!(app.should_quit()); } #[test] fn test_repeat_command_with_r_key() { use crossterm::event::{KeyCode, KeyEvent, KeyModifiers}; use ghost::app::tui::app::TuiApp; let env = TestEnvironment::new(); let mut app = TuiApp::new_with_config(env.config.clone()).unwrap(); // Add test tasks let tasks = create_test_tasks(); app.tasks = tasks; app.table_scroll.set_total_items(app.tasks.len()); // Select the first task (index 0) - "echo hello" app.set_selected_index(0); // Press 'r' key to repeat command let key_r = KeyEvent::new(KeyCode::Char('r'), KeyModifiers::NONE); let result = app.handle_key(key_r); // Should succeed assert!(result.is_ok()); // Since this is a test environment without a real database, // we can't verify that a new task was created. // The real behavior is tested by the fact that the operation doesn't panic // and returns Ok(()). } #[test] fn test_process_details_display() { use ghost::app::tui::app::TuiApp; let env = TestEnvironment::new(); let mut app = TuiApp::new_with_config(env.config.clone()).unwrap(); // Add a test task let tasks = vec![Task { id: "test-task-id".to_string(), pid: 5678, pgid: Some(5678), command: r#"["echo", "hello world"]"#.to_string(), env: Some(r#"[["TEST_VAR","test_value"]]"#.to_string()), cwd: Some("/tmp/test".to_string()), status: TaskStatus::Exited, exit_code: Some(0), started_at: 1000000000, finished_at: Some(1000001000), log_path: "/tmp/ghost/logs/test.log".to_string(), }]; app.tasks = tasks; app.table_scroll.set_total_items(1); app.view_mode = ViewMode::ProcessDetails; app.selected_task_id = Some("test-task-id".to_string()); // Create a terminal and render the process details view let backend = TestBackend::new(80, 20); let mut terminal = Terminal::new(backend).unwrap(); terminal.draw(|f| app.render(f)).unwrap(); let buffer_output = buffer_to_string(terminal.backend().buffer()); let normalized_output = normalize_buffer_output(&buffer_output); let normalized_output = normalize_dynamic_output(&normalized_output); let normalized_output = normalize_without_status_line(&normalized_output); let expected = load_expected("process_details_display.txt"); let normalized_expected = normalize_buffer_output(&expected); let normalized_expected = normalize_dynamic_output(&normalized_expected); let normalized_expected = normalize_without_status_line(&normalized_expected); assert_eq!( normalized_output, normalized_expected, "Process details display does not match expected output" ); } #[test] fn test_log_viewer_navigation() { use crossterm::event::{KeyCode, KeyEvent, KeyModifiers}; use ghost::app::tui::app::TuiApp; let env = TestEnvironment::new(); let mut app = TuiApp::new_with_config(env.config.clone()).unwrap(); // Add a test task with a log file let tasks = vec![Task { id: "log-test-task".to_string(), pid: 9999, pgid: Some(9999), command: r#"["echo", "test"]"#.to_string(), env: None, cwd: Some("/tmp".to_string()), status: TaskStatus::Exited, exit_code: Some(0), started_at: 1000000000, finished_at: Some(1000001000), log_path: "/tmp/ghost/logs/test.log".to_string(), }]; app.tasks = tasks; app.table_scroll.set_total_items(1); // Initial view should be TaskList assert_eq!(app.view_mode, ViewMode::TaskList); // Press 'l' to view logs let key_l = KeyEvent::new(KeyCode::Char('l'), KeyModifiers::NONE); app.handle_key(key_l).unwrap(); assert_eq!(app.view_mode, ViewMode::LogView); // Test log viewer navigation keys // Press 'j' to scroll down let key_j = KeyEvent::new(KeyCode::Char('j'), KeyModifiers::NONE); app.handle_key(key_j).unwrap(); // Press 'k' to scroll up let key_k = KeyEvent::new(KeyCode::Char('k'), KeyModifiers::NONE); app.handle_key(key_k).unwrap(); // Press 'g' to go to top let key_g = KeyEvent::new(KeyCode::Char('g'), KeyModifiers::NONE); app.handle_key(key_g).unwrap(); // Press 'G' to go to bottom let key_g_upper = KeyEvent::new(KeyCode::Char('G'), KeyModifiers::NONE); app.handle_key(key_g_upper).unwrap(); // Press 'h' to scroll left let key_h = KeyEvent::new(KeyCode::Char('h'), KeyModifiers::NONE); app.handle_key(key_h).unwrap(); // Press 'l' to scroll right let key_l_scroll = KeyEvent::new(KeyCode::Char('l'), KeyModifiers::NONE); app.handle_key(key_l_scroll).unwrap(); // Press Esc to go back to task list let key_esc = KeyEvent::new(KeyCode::Esc, KeyModifiers::NONE); app.handle_key(key_esc).unwrap(); assert_eq!(app.view_mode, ViewMode::TaskList); } #[test] fn test_integrated_navigation_flow() { use crossterm::event::{KeyCode, KeyEvent, KeyModifiers}; use ghost::app::tui::app::TuiApp; let env = TestEnvironment::new(); let mut app = TuiApp::new_with_config(env.config.clone()).unwrap(); // Add multiple tasks with different statuses let tasks = vec![ Task { id: "task-1".to_string(), pid: 1001, pgid: Some(1001), command: r#"["sleep", "60"]"#.to_string(), env: Some(r#"[["VAR1","value1"]]"#.to_string()), cwd: Some("/home/user".to_string()), status: TaskStatus::Running, exit_code: None, started_at: 1000000000, finished_at: None, log_path: "/tmp/ghost/logs/task-1.log".to_string(), }, Task { id: "task-2".to_string(), pid: 1002, pgid: Some(1002), command: r#"["echo", "done"]"#.to_string(), env: Some(r#"[["VAR2","value2"]]"#.to_string()), cwd: Some("/tmp".to_string()), status: TaskStatus::Exited, exit_code: Some(0), started_at: 1000000100, finished_at: Some(1000000200), log_path: "/tmp/ghost/logs/task-2.log".to_string(), }, ]; app.tasks = tasks; app.table_scroll.set_total_items(2); // Start in task list assert_eq!(app.view_mode, ViewMode::TaskList); assert_eq!(app.selected_index(), 0); // Navigate down to second task let key_j = KeyEvent::new(KeyCode::Char('j'), KeyModifiers::NONE); app.handle_key(key_j).unwrap(); assert_eq!(app.selected_index(), 1); // View process details of second task let key_enter = KeyEvent::new(KeyCode::Enter, KeyModifiers::NONE); app.handle_key(key_enter).unwrap(); assert_eq!(app.view_mode, ViewMode::ProcessDetails); assert_eq!(app.selected_task_id, Some("task-2".to_string())); // Go back to task list let key_esc = KeyEvent::new(KeyCode::Esc, KeyModifiers::NONE); app.handle_key(key_esc).unwrap(); assert_eq!(app.view_mode, ViewMode::TaskList); // Navigate back to first task let key_k = KeyEvent::new(KeyCode::Char('k'), KeyModifiers::NONE); app.handle_key(key_k).unwrap(); assert_eq!(app.selected_index(), 0); // View logs of first task let key_l = KeyEvent::new(KeyCode::Char('l'), KeyModifiers::NONE); app.handle_key(key_l).unwrap(); assert_eq!(app.view_mode, ViewMode::LogView); // Go back and view process details app.handle_key(key_esc).unwrap(); app.handle_key(key_enter).unwrap(); assert_eq!(app.view_mode, ViewMode::ProcessDetails); assert_eq!(app.selected_task_id, Some("task-1".to_string())); // Test filter cycling from details view app.handle_key(key_esc).unwrap(); let key_tab = KeyEvent::new(KeyCode::Tab, KeyModifiers::NONE); app.handle_key(key_tab).unwrap(); assert_eq!(app.filter, TaskFilter::Running); // Quit from task list let key_q = KeyEvent::new(KeyCode::Char('q'), KeyModifiers::NONE); app.handle_key(key_q).unwrap(); assert!(app.should_quit()); } #[test] fn test_log_viewer_display() { use ghost::app::tui::app::TuiApp; let env = TestEnvironment::new(); let mut app = TuiApp::new_with_config(env.config.clone()).unwrap(); // Create test log file let log_dir = env._temp_dir.path().join("logs"); std::fs::create_dir_all(&log_dir).unwrap(); let log_path = log_dir.join("test-log.log"); std::fs::write( &log_path, "Line 1: Starting process\nLine 2: Processing...\nLine 3: Complete\n", ) .unwrap(); // Add a task with the log file let tasks = vec![Task { id: "log-display-test".to_string(), pid: 8888, pgid: Some(8888), command: r#"["test", "command"]"#.to_string(), env: None, cwd: Some("/tmp".to_string()), status: TaskStatus::Exited, exit_code: Some(0), started_at: 1000000000, finished_at: Some(1000001000), log_path: log_path.to_string_lossy().to_string(), }]; app.tasks = tasks; app.table_scroll.set_total_items(1); app.view_mode = ViewMode::LogView; // Create a terminal and render the log view let backend = TestBackend::new(80, 15); let mut terminal = Terminal::new(backend).unwrap(); terminal.draw(|f| app.render(f)).unwrap(); let buffer_output = buffer_to_string(terminal.backend().buffer()); let normalized_output = normalize_buffer_output(&buffer_output); let expected = load_expected("log_viewer_display.txt"); let normalized_expected = normalize_buffer_output(&expected); assert_eq!( normalized_output, normalized_expected, "Log viewer display does not match expected output" ); } #[test] fn test_process_details_with_many_env_vars() { use ghost::app::tui::app::TuiApp; let env = TestEnvironment::new(); let mut app = TuiApp::new_with_config(env.config.clone()).unwrap(); // Add a test task with many environment variables let tasks = vec![Task { id: "abc12345-6789-1234-5678-123456789abc".to_string(), pid: 12345, pgid: Some(12345), command: r#"["npm", "run", "dev"]"#.to_string(), env: Some(r#"[["NODE_ENV","development"],["PORT","3000"],["DATABASE_URL","postgresql://localhost:5432/mydb"],["API_KEY","secret123"],["DEBUG","true"],["LOG_LEVEL","verbose"]]"#.to_string()), cwd: Some("/home/user/projects/myapp".to_string()), status: TaskStatus::Running, exit_code: None, started_at: 1704109200, // 2024-01-01 10:00 UTC finished_at: None, log_path: "/tmp/ghost/logs/test.log".to_string(), }]; app.tasks = tasks; app.table_scroll.set_total_items(1); app.view_mode = ViewMode::ProcessDetails; app.selected_task_id = Some("abc12345-6789-1234-5678-123456789abc".to_string()); // Create a terminal and render the process details view let backend = TestBackend::new(80, 20); let mut terminal = Terminal::new(backend).unwrap(); terminal.draw(|f| app.render(f)).unwrap(); let buffer_output = buffer_to_string(terminal.backend().buffer()); let normalized_output = normalize_buffer_output(&buffer_output); let normalized_output = normalize_dynamic_output(&normalized_output); let normalized_output = normalize_without_status_line(&normalized_output); let expected = load_expected("process_details_many_env_vars.txt"); let normalized_expected = normalize_buffer_output(&expected); let normalized_expected = normalize_dynamic_output(&normalized_expected); let normalized_expected = normalize_without_status_line(&normalized_expected); assert_eq!( normalized_output, normalized_expected, "Process details with many env vars display does not match expected output" ); } #[test] fn test_log_viewer_with_many_lines() { use ghost::app::tui::app::TuiApp; let env = TestEnvironment::new(); let mut app = TuiApp::new_with_config(env.config.clone()).unwrap(); // Create test log file with many lines let log_dir = env._temp_dir.path().join("logs"); std::fs::create_dir_all(&log_dir).unwrap(); let log_path = log_dir.join("test-many-lines.log"); let mut log_content = String::new(); for i in 1..=100 { log_content.push_str(&format!("Line {i}: Log message number {i}\n")); } std::fs::write(&log_path, log_content).unwrap(); // Add a task with the log file let tasks = vec![Task { id: "many-lines-test".to_string(), pid: 7777, pgid: Some(7777), command: r#"["tail", "-f", "app.log"]"#.to_string(), env: None, cwd: Some("/var/log".to_string()), status: TaskStatus::Running, exit_code: None, started_at: 1000000000, finished_at: None, log_path: log_path.to_string_lossy().to_string(), }]; app.tasks = tasks; app.table_scroll.set_total_items(1); app.view_mode = ViewMode::LogView; // Create a terminal and render the log view let backend = TestBackend::new(80, 15); let mut terminal = Terminal::new(backend).unwrap(); terminal.draw(|f| app.render(f)).unwrap(); let buffer_output = buffer_to_string(terminal.backend().buffer()); let normalized_output = normalize_buffer_output(&buffer_output);
rust
MIT
2608a5ac4d31e61433a039387325199a22701da6
2026-01-04T20:19:35.234323Z
true
esp-rs/rfc2217-rs
https://github.com/esp-rs/rfc2217-rs/blob/6dd355b57726fcf211a5f06fb866eafda176c431/src/lib.rs
src/lib.rs
#![cfg_attr(not(any(test, feature = "std")), no_std)] mod codes; pub mod command; pub mod negotiation; pub mod parser; #[cfg(feature = "std")] mod serialport_conversions; #[cfg(feature = "std")] pub mod server; pub mod subnegotiation; // Public API pub use command::Command; pub use negotiation::Negotiation; pub use parser::Parser; #[cfg(feature = "std")] pub use server::Server; pub use subnegotiation::Subnegotiation;
rust
Apache-2.0
6dd355b57726fcf211a5f06fb866eafda176c431
2026-01-04T20:19:28.343083Z
false
esp-rs/rfc2217-rs
https://github.com/esp-rs/rfc2217-rs/blob/6dd355b57726fcf211a5f06fb866eafda176c431/src/command.rs
src/command.rs
use crate::codes; pub const SIZE: usize = 2; // Telnet commands without the ones related to negotiation and subnegotiation, // defined here: https://www.rfc-editor.org/rfc/rfc854.txt #[derive(Debug, PartialEq, Eq, Clone, Copy)] pub enum Command { NoOp, DataMark, Break, InterruptProcess, AbortOutput, AreYouThere, EraseCharacter, EraseLine, GoAhead, Unsupported(u8), } impl Command { pub fn serialize(&self, buf: &mut [u8]) { buf[0] = codes::IAC; buf[1] = match *self { Self::NoOp => 241, Self::DataMark => 242, Self::Break => 243, Self::InterruptProcess => 244, Self::AbortOutput => 245, Self::AreYouThere => 246, Self::EraseCharacter => 247, Self::EraseLine => 248, Self::GoAhead => 249, Self::Unsupported(byte) => byte, } } pub const fn deserialize(buf: &[u8]) -> Self { assert!(buf[0] == codes::IAC); match buf[1] { 241 => Self::NoOp, 242 => Self::DataMark, 243 => Self::Break, 244 => Self::InterruptProcess, 245 => Self::AbortOutput, 246 => Self::AreYouThere, 247 => Self::EraseCharacter, 248 => Self::EraseLine, 249 => Self::GoAhead, _ => Self::Unsupported(buf[1]), } } }
rust
Apache-2.0
6dd355b57726fcf211a5f06fb866eafda176c431
2026-01-04T20:19:28.343083Z
false
esp-rs/rfc2217-rs
https://github.com/esp-rs/rfc2217-rs/blob/6dd355b57726fcf211a5f06fb866eafda176c431/src/parser.rs
src/parser.rs
use crate::{codes, command, negotiation, subnegotiation, Command, Negotiation, Subnegotiation}; #[derive(Debug, PartialEq, Eq, Clone, Copy)] pub enum Event { Data(u8), Command(Command), Negotiation(Negotiation), Subnegotiation(Subnegotiation), } #[derive(Debug, PartialEq, Eq, Clone, Copy)] pub enum Error { SubnegotiationParsing, BufferOverflow, } enum State { Data, Command, Negotiation, SubnegotiationOption, SubnegotiationSubOption, SubnegotiationData, SubnegotiationEnd, } pub struct Parser { state: State, buf: [u8; subnegotiation::MAX_SIZE], buf_cnt: usize, } impl Parser { pub const fn new() -> Self { Self { state: State::Data, buf: [0; subnegotiation::MAX_SIZE], buf_cnt: 0, } } pub fn reset(&mut self) { *self = Self::new(); } pub fn process_byte(&mut self, byte: u8) -> Result<Option<Event>, Error> { match self.state { State::Data => { if byte == codes::IAC { self.state = State::Command; return Ok(None); } Ok(Some(Event::Data(byte))) } State::Command => { if byte == codes::IAC { self.state = State::Data; return Ok(Some(Event::Data(byte))); } self.buf_cnt = 0; self.write_to_buf(codes::IAC)?; self.write_to_buf(byte)?; Ok(self.process_command_byte(byte)) } State::Negotiation => { self.write_to_buf(byte)?; self.state = State::Data; Ok(Some(Event::Negotiation(Negotiation::deserialize( &self.buf[..negotiation::SIZE], )))) } State::SubnegotiationOption => { self.write_to_buf(byte)?; self.state = State::SubnegotiationSubOption; Ok(None) } State::SubnegotiationSubOption => { self.write_to_buf(byte)?; self.state = State::SubnegotiationData; Ok(None) } State::SubnegotiationData => { self.write_to_buf(byte)?; if byte == codes::IAC { self.state = State::SubnegotiationEnd; } Ok(None) } State::SubnegotiationEnd => { match byte { // If the IAC byte repeats it's data codes::IAC => { self.write_to_buf(byte)?; self.state = State::SubnegotiationData; Ok(None) } codes::SE => { self.write_to_buf(byte)?; self.state = State::Data; Ok(Some(Event::Subnegotiation(Subnegotiation::deserialize( &self.buf[..self.buf_cnt], )))) } _ => Err(Error::SubnegotiationParsing), } } } } fn process_command_byte(&mut self, command_code: u8) -> Option<Event> { match command_code { codes::WILL | codes::WONT | codes::DO | codes::DONT => { self.state = State::Negotiation; None } codes::SB => { self.state = State::SubnegotiationOption; None } _ => { self.state = State::Data; Some(Event::Command(Command::deserialize( &self.buf[..command::SIZE], ))) } } } fn write_to_buf(&mut self, byte: u8) -> Result<(), Error> { if self.buf_cnt == self.buf.len() { return Err(Error::BufferOverflow); } self.buf[self.buf_cnt] = byte; self.buf_cnt += 1; Ok(()) } }
rust
Apache-2.0
6dd355b57726fcf211a5f06fb866eafda176c431
2026-01-04T20:19:28.343083Z
false
esp-rs/rfc2217-rs
https://github.com/esp-rs/rfc2217-rs/blob/6dd355b57726fcf211a5f06fb866eafda176c431/src/codes.rs
src/codes.rs
// Telnet command codes needed for command, negotiation and // subnegotiation serializing/deserializing pub const IAC: u8 = 255; pub const WILL: u8 = 251; pub const WONT: u8 = 252; pub const DO: u8 = 253; pub const DONT: u8 = 254; pub const SB: u8 = 250; pub const SE: u8 = 240; pub const COM_PORT_OPTION: u8 = 44;
rust
Apache-2.0
6dd355b57726fcf211a5f06fb866eafda176c431
2026-01-04T20:19:28.343083Z
false
esp-rs/rfc2217-rs
https://github.com/esp-rs/rfc2217-rs/blob/6dd355b57726fcf211a5f06fb866eafda176c431/src/serialport_conversions.rs
src/serialport_conversions.rs
use serialport::{DataBits, FlowControl, Parity, StopBits}; // Required functions for conversions between serialport Enums and rfc2217 option values pub(crate) const fn data_bits_to_u8(data_bits: DataBits) -> u8 { match data_bits { DataBits::Five => 5, DataBits::Six => 6, DataBits::Seven => 7, DataBits::Eight => 8, } } pub(crate) const fn u8_to_data_bits(value: u8) -> Option<DataBits> { match value { 5 => Some(DataBits::Five), 6 => Some(DataBits::Six), 7 => Some(DataBits::Seven), 8 => Some(DataBits::Eight), _ => None, } } pub(crate) const fn parity_to_u8(parity: Parity) -> u8 { match parity { Parity::None => 1, Parity::Odd => 2, Parity::Even => 3, } } pub(crate) const fn u8_to_parity(value: u8) -> Option<Parity> { match value { 1 => Some(Parity::None), 2 => Some(Parity::Odd), 3 => Some(Parity::Even), _ => None, } } pub(crate) const fn stop_bits_to_u8(stop_bits: StopBits) -> u8 { match stop_bits { StopBits::One => 1, StopBits::Two => 2, } } pub(crate) const fn u8_to_stop_bits(value: u8) -> Option<StopBits> { match value { 1 => Some(StopBits::One), 2 => Some(StopBits::Two), _ => None, } } pub(crate) const fn flow_control_to_u8(flow_control: FlowControl) -> u8 { match flow_control { FlowControl::None => 1, FlowControl::Software => 2, FlowControl::Hardware => 3, } } pub(crate) const fn u8_to_flow_control(value: u8) -> Option<FlowControl> { match value { 1 => Some(FlowControl::None), 2 => Some(FlowControl::Software), 3 => Some(FlowControl::Hardware), _ => None, } }
rust
Apache-2.0
6dd355b57726fcf211a5f06fb866eafda176c431
2026-01-04T20:19:28.343083Z
false
esp-rs/rfc2217-rs
https://github.com/esp-rs/rfc2217-rs/blob/6dd355b57726fcf211a5f06fb866eafda176c431/src/subnegotiation.rs
src/subnegotiation.rs
use crate::codes; pub const MAX_DATA_SIZE: usize = 256; pub const NONDATA_SIZE: usize = 6; pub const MAX_SIZE: usize = MAX_DATA_SIZE + NONDATA_SIZE; // RFC2217 subnegotiation options, defined here: https://www.rfc-editor.org/rfc/rfc2217.html #[derive(Debug, PartialEq, Eq, Clone, Copy)] pub enum Subnegotiation { SetSignature { data: [u8; MAX_DATA_SIZE], size: u8, }, SetBaudRate(u32), SetDataSize(u8), SetParity(u8), SetStopSize(u8), SetControl(u8), NotifyLineState(u8), NotifyModemState(u8), FlowControlSuspend, FlowControlResume, SetLinestateMask(u8), SetModemStateMask(u8), PurgeData(u8), Unsupported { base_option_code: u8, option_code: u8, data: [u8; MAX_DATA_SIZE], data_cnt: u8, }, } // The codes for client to server and server to client ComPort options differ by 100, // this indicates which one the serializer should pick #[derive(Clone, Copy)] enum OptionKind { ClientToServer, ServerToClient, } impl Subnegotiation { pub fn serialize_client(&self, buf: &mut [u8]) -> usize { self.serialize(buf, OptionKind::ClientToServer) } pub fn serialize_server(&self, buf: &mut [u8]) -> usize { self.serialize(buf, OptionKind::ServerToClient) } fn serialize(&self, buf: &mut [u8], option_kind: OptionKind) -> usize { let start = |option_code: u8| -> [u8; 4] { [ codes::IAC, codes::SB, codes::COM_PORT_OPTION, match option_kind { OptionKind::ClientToServer => option_code, OptionKind::ServerToClient => option_code + 100, }, ] }; let end = [codes::IAC, codes::SE]; let mut subnegotiate = |option_code: u8, data: &[u8]| -> usize { buf[..4].copy_from_slice(&start(option_code)); buf[4..4 + data.len()].copy_from_slice(data); buf[4 + data.len()..NONDATA_SIZE + data.len()].copy_from_slice(&end); NONDATA_SIZE + data.len() }; match *self { Self::SetSignature { data, size } => { buf[..4].copy_from_slice(&start(0)); let data_slice = &data[..size as usize]; let mut i = 4; for byte in data_slice { buf[i] = *byte; i += 1; // Make sure to escape IAC bytes in the signature if *byte == codes::IAC { buf[i] = *byte; i += 1; } } buf[i..i + 2].copy_from_slice(&end); i + 2 } Self::SetBaudRate(baud) => subnegotiate(1, &u32::to_be_bytes(baud)), Self::SetDataSize(data_size) => subnegotiate(2, &[data_size]), Self::SetParity(parity) => subnegotiate(3, &[parity]), Self::SetStopSize(stopsize) => subnegotiate(4, &[stopsize]), Self::SetControl(control) => subnegotiate(5, &[control]), Self::NotifyLineState(linestate) => subnegotiate(6, &[linestate]), Self::NotifyModemState(modemstate) => subnegotiate(7, &[modemstate]), Self::FlowControlSuspend => subnegotiate(8, &[]), Self::FlowControlResume => subnegotiate(9, &[]), Self::SetLinestateMask(linestate_mask) => subnegotiate(10, &[linestate_mask]), Self::SetModemStateMask(modemstate_mask) => subnegotiate(11, &[modemstate_mask]), Self::PurgeData(purge_data) => subnegotiate(12, &[purge_data]), Self::Unsupported { base_option_code, option_code, data, data_cnt, } => { buf[..4].copy_from_slice(&[codes::IAC, codes::SB, base_option_code, option_code]); buf[4..4 + data_cnt as usize].copy_from_slice(&data[..data_cnt as usize]); buf[4 + data_cnt as usize..NONDATA_SIZE + data_cnt as usize].copy_from_slice(&end); NONDATA_SIZE + data_cnt as usize } } } pub fn deserialize(buf: &[u8]) -> Self { assert!( buf[0] == codes::IAC && buf[1] == codes::SB && buf[buf.len() - 2] == codes::IAC && buf[buf.len() - 1] == codes::SE ); let base_option_code = buf[2]; let option_code = buf[3]; let data_len = buf.len() - NONDATA_SIZE; let data = &buf[4..4 + data_len]; match base_option_code { codes::COM_PORT_OPTION => match option_code { 0 | 100 => { let mut data_no_escapes = [0; MAX_DATA_SIZE]; let mut i = 0; let mut iac_occured = false; for &byte in data { if byte == codes::IAC { if iac_occured { iac_occured = false; continue; } else { iac_occured = true; } } data_no_escapes[i] = byte; i += 1; } Self::SetSignature { data: data_no_escapes, size: i as u8, } } 1 | 101 => { let baud_rate = u32::from_be_bytes([data[0], data[1], data[2], data[3]]); Self::SetBaudRate(baud_rate) } 2 | 102 => Self::SetDataSize(data[0]), 3 | 103 => Self::SetParity(data[0]), 4 | 104 => Self::SetStopSize(data[0]), 5 | 105 => Self::SetControl(data[0]), 6 | 106 => Self::NotifyLineState(data[0]), 7 | 107 => Self::NotifyModemState(data[0]), 8 | 108 => Self::FlowControlSuspend, 9 | 109 => Self::FlowControlResume, 10 | 110 => Self::SetLinestateMask(data[0]), 11 | 111 => Self::SetModemStateMask(data[0]), 12 | 112 => Self::PurgeData(data[0]), _ => panic!("Option code is not a Com Port option code"), }, _ => { let mut data_arr = [0; MAX_DATA_SIZE]; data_arr.copy_from_slice(data); Self::Unsupported { base_option_code: base_option_code, option_code: option_code, data: data_arr, data_cnt: data_len as u8, } } } } }
rust
Apache-2.0
6dd355b57726fcf211a5f06fb866eafda176c431
2026-01-04T20:19:28.343083Z
false
esp-rs/rfc2217-rs
https://github.com/esp-rs/rfc2217-rs/blob/6dd355b57726fcf211a5f06fb866eafda176c431/src/server.rs
src/server.rs
use crate::serialport_conversions::*; use crate::{ codes, negotiation, parser, subnegotiation, Command, Negotiation, Parser, Subnegotiation, }; use serialport::{ClearBuffer, FlowControl, SerialPort}; use std::io::{self, BufWriter, Read, Write}; use std::net::{TcpListener, TcpStream, ToSocketAddrs}; #[derive(Debug)] pub enum Error { Parsing(parser::Error), SerialInit(serialport::Error), Serial(io::Error), Tcp(io::Error), } pub struct Server { port: Box<dyn SerialPort>, port_writer: BufWriter<Box<dyn SerialPort>>, tcp_conn: TcpStream, tcp_writer: BufWriter<TcpStream>, tcp_answer_buf: [u8; subnegotiation::MAX_SIZE], parser: Parser, signature: Vec<u8>, suspended_flow_control: FlowControl, break_state: bool, } impl Server { pub fn new<A: ToSocketAddrs>(serial_port_name: &str, tcp_addr: A) -> Result<Self, Error> { let port = serialport::new(serial_port_name, 9600) .open() .map_err(Error::SerialInit)?; let port_clone = port.try_clone().map_err(Error::SerialInit)?; let listener = TcpListener::bind(tcp_addr).map_err(Error::Tcp)?; let (connection, _) = listener.accept().map_err(Error::Tcp)?; connection.set_nonblocking(true).map_err(Error::Tcp)?; let cloned_connection = connection.try_clone().map_err(Error::Tcp)?; Ok(Server { port: port, parser: Parser::new(), port_writer: BufWriter::new(port_clone), tcp_conn: connection, tcp_writer: BufWriter::new(cloned_connection), tcp_answer_buf: [0; subnegotiation::MAX_SIZE], signature: Vec::new(), suspended_flow_control: FlowControl::None, break_state: false, }) } pub fn run(&mut self) -> Result<(), Error> { // Read and handle the data from the TCP connection let mut tcp_data = [0; 256]; match self.tcp_conn.read(&mut tcp_data) { Ok(bytes_read) => { self.process_tcp_data(&tcp_data[..bytes_read])?; } Err(error) => match error.kind() { io::ErrorKind::WouldBlock => {} _ => return Err(Error::Tcp(error)), }, } // Read and handle the data from the serial port let mut port_data = [0; 256]; match self.port.read(&mut port_data) { Ok(bytes_read) => { for &byte in &port_data[..bytes_read] { // Escape all IAC bytes self.tcp_writer.write_all(&[byte]).map_err(Error::Tcp)?; if byte == codes::IAC { self.tcp_writer.write_all(&[byte]).map_err(Error::Tcp)?; } } } Err(error) => match error.kind() { io::ErrorKind::TimedOut => {} _ => return Err(Error::Serial(error)), }, } // Flush the buffered data to be sent self.port_writer.flush().map_err(Error::Serial)?; self.tcp_writer.flush().map_err(Error::Tcp)?; Ok(()) } fn process_tcp_data(&mut self, bytes: &[u8]) -> Result<(), Error> { for &byte in bytes { if let Some(event) = self.parser.process_byte(byte).map_err(Error::Parsing)? { let answer_size = self.process_event(event).map_err(Error::Serial)?; self.tcp_writer .write_all(&self.tcp_answer_buf[..answer_size]) .map_err(Error::Tcp)?; } } Ok(()) } fn process_event(&mut self, event: parser::Event) -> Result<usize, io::Error> { match event { parser::Event::Data(byte) => { self.port_writer.write_all(&[byte])?; Ok(0) } parser::Event::Command(command) => self.process_command(command), parser::Event::Negotiation(negotiation) => self.process_negotiation(negotiation), parser::Event::Subnegotiation(subnegotiation) => { self.process_subnegotiation(subnegotiation) } } } fn process_command(&mut self, command: Command) -> Result<usize, io::Error> { match command { _ => Ok(0), } } fn process_negotiation(&mut self, negotiation: Negotiation) -> Result<usize, io::Error> { match negotiation.get_answer() { Some(answer) => { answer.serialize(&mut self.tcp_answer_buf[..negotiation::SIZE]); Ok(negotiation::SIZE) } None => Ok(0), } } fn process_subnegotiation( &mut self, subnegotiation: Subnegotiation, ) -> Result<usize, io::Error> { let answer_opt = match subnegotiation { Subnegotiation::SetSignature { data, size } => { // An empty signature constitutes a signature query if size == 0 { let mut data = [0; subnegotiation::MAX_DATA_SIZE]; let size = self.signature.len() as u8; data.copy_from_slice(&self.signature); Some(Subnegotiation::SetSignature { data, size }) } else { self.signature.copy_from_slice(&data[..size as usize]); Some(subnegotiation) } } Subnegotiation::SetBaudRate(val) => { if val == 0 { Some(Subnegotiation::SetBaudRate(self.port.baud_rate()?)) } else { self.port.set_baud_rate(val)?; Some(subnegotiation) } } Subnegotiation::SetDataSize(val) => match u8_to_data_bits(val) { Some(data_bits) => { self.port.set_data_bits(data_bits)?; Some(subnegotiation) } None => Some(Subnegotiation::SetDataSize(data_bits_to_u8( self.port.data_bits()?, ))), }, Subnegotiation::SetParity(val) => match u8_to_parity(val) { Some(parity) => { self.port.set_parity(parity)?; Some(subnegotiation) } None => Some(Subnegotiation::SetParity(parity_to_u8(self.port.parity()?))), }, Subnegotiation::SetStopSize(val) => match u8_to_stop_bits(val) { Some(stop_bits) => { self.port.set_stop_bits(stop_bits)?; Some(subnegotiation) } None => Some(Subnegotiation::SetStopSize(stop_bits_to_u8( self.port.stop_bits()?, ))), }, Subnegotiation::SetControl(val) => self.handle_set_control(val)?, Subnegotiation::FlowControlSuspend => { self.suspended_flow_control = self.port.flow_control()?; self.port.set_flow_control(FlowControl::None)?; Some(subnegotiation) } Subnegotiation::FlowControlResume => { self.port.set_flow_control(self.suspended_flow_control)?; Some(subnegotiation) } Subnegotiation::PurgeData(val) => self.handle_purge_data(val)?, _ => None, }; match answer_opt { Some(answer) => Ok(answer.serialize_server(&mut self.tcp_answer_buf)), None => Ok(0), } } fn handle_set_control(&mut self, val: u8) -> Result<Option<Subnegotiation>, io::Error> { match val { 0 => Ok(Some(Subnegotiation::SetControl(flow_control_to_u8( self.port.flow_control()?, )))), 1 | 2 | 3 => { self.port .set_flow_control(u8_to_flow_control(val).unwrap())?; Ok(Some(Subnegotiation::SetControl(val))) } 4 => match self.break_state { true => Ok(Some(Subnegotiation::SetControl(5))), false => Ok(Some(Subnegotiation::SetControl(6))), }, 5 => { self.port.set_break()?; self.break_state = true; Ok(Some(Subnegotiation::SetControl(val))) } 6 => { self.port.clear_break()?; self.break_state = false; Ok(Some(Subnegotiation::SetControl(val))) } 7 => match self.port.read_data_set_ready()? { true => Ok(Some(Subnegotiation::SetControl(8))), false => Ok(Some(Subnegotiation::SetControl(9))), }, 8 => { self.port.write_data_terminal_ready(true)?; Ok(Some(Subnegotiation::SetControl(val))) } 9 => { self.port.write_data_terminal_ready(false)?; Ok(Some(Subnegotiation::SetControl(val))) } 10 => match self.port.read_clear_to_send()? { true => Ok(Some(Subnegotiation::SetControl(11))), false => Ok(Some(Subnegotiation::SetControl(12))), }, 11 => { self.port.write_request_to_send(true)?; Ok(Some(Subnegotiation::SetControl(val))) } 12 => { self.port.write_request_to_send(false)?; Ok(Some(Subnegotiation::SetControl(val))) } _ => Ok(None), } } fn handle_purge_data(&mut self, val: u8) -> Result<Option<Subnegotiation>, io::Error> { match val { 1 => { self.port.clear(ClearBuffer::Input)?; Ok(Some(Subnegotiation::PurgeData(val))) } 2 => { self.port.clear(ClearBuffer::Output)?; Ok(Some(Subnegotiation::PurgeData(val))) } 3 => { self.port.clear(ClearBuffer::Input)?; self.port.clear(ClearBuffer::Output)?; Ok(Some(Subnegotiation::PurgeData(val))) } _ => Ok(None), } } } impl Negotiation { fn get_answer(&self) -> Option<Negotiation> { match (self.intent, self.option) { ( negotiation::Intent::Will, negotiation::Option::Binary | negotiation::Option::ComPort | negotiation::Option::SuppressGoAhead, ) => Some(Negotiation { intent: negotiation::Intent::Do, option: self.option, }), ( negotiation::Intent::Do, negotiation::Option::Binary | negotiation::Option::ComPort | negotiation::Option::SuppressGoAhead, ) => None, (negotiation::Intent::Will, _) => Some(Negotiation { intent: negotiation::Intent::Dont, option: self.option, }), (negotiation::Intent::Do, _) => Some(Negotiation { intent: negotiation::Intent::Wont, option: self.option, }), _ => panic!(), } } }
rust
Apache-2.0
6dd355b57726fcf211a5f06fb866eafda176c431
2026-01-04T20:19:28.343083Z
false
esp-rs/rfc2217-rs
https://github.com/esp-rs/rfc2217-rs/blob/6dd355b57726fcf211a5f06fb866eafda176c431/src/negotiation.rs
src/negotiation.rs
use crate::codes; pub const SIZE: usize = 3; #[derive(Debug, PartialEq, Eq, Clone, Copy)] pub struct Negotiation { pub intent: Intent, pub option: Option, } // Telnet options #[derive(Debug, PartialEq, Eq, Clone, Copy)] pub enum Option { Binary, Echo, SuppressGoAhead, ComPort, Unsupported(u8), } #[derive(Debug, PartialEq, Eq, Clone, Copy)] pub enum Intent { Will, Wont, Do, Dont, } impl Negotiation { pub fn serialize(&self, buf: &mut [u8]) { buf[0] = codes::IAC; buf[1] = self.intent.to_u8(); buf[2] = self.option.to_u8(); } pub const fn deserialize(buf: &[u8]) -> Self { assert!(buf[0] == codes::IAC); Self { intent: Intent::from_u8(buf[1]), option: Option::from_u8(buf[2]), } } } impl Option { const fn from_u8(byte: u8) -> Option { match byte { 0 => Self::Binary, 1 => Self::Echo, 3 => Self::SuppressGoAhead, 44 => Self::ComPort, _ => Self::Unsupported(byte), } } const fn to_u8(self) -> u8 { match self { Self::Binary => 0, Self::Echo => 1, Self::SuppressGoAhead => 3, Self::ComPort => 44, Self::Unsupported(byte) => byte, } } } impl Intent { const fn from_u8(byte: u8) -> Intent { match byte { codes::WILL => Self::Will, codes::WONT => Self::Wont, codes::DO => Self::Do, codes::DONT => Self::Dont, _ => panic!("Not a command code for negotiation intent"), } } const fn to_u8(self) -> u8 { match self { Self::Will => codes::WILL, Self::Wont => codes::WONT, Self::Do => codes::DO, Self::Dont => codes::DONT, } } }
rust
Apache-2.0
6dd355b57726fcf211a5f06fb866eafda176c431
2026-01-04T20:19:28.343083Z
false
esp-rs/rfc2217-rs
https://github.com/esp-rs/rfc2217-rs/blob/6dd355b57726fcf211a5f06fb866eafda176c431/src/bin/server.rs
src/bin/server.rs
use clap::Parser; use std::net::IpAddr; use rfc2217_rs::Server; #[derive(Parser, Debug)] struct Args { #[clap(long = "serial_port", short = 'p', default_value = "/dev/ttyUSB0")] serial_port: String, #[clap(long = "address", short = 'a', default_value = "127.0.0.1")] address: IpAddr, #[clap(long = "tcp_port", default_value = "7878")] tcp_port: u16, } fn main() { let Args { address, tcp_port, serial_port } = Args::parse(); let mut server = Server::new(&serial_port, (address, tcp_port)).unwrap(); loop { server.run().unwrap(); } }
rust
Apache-2.0
6dd355b57726fcf211a5f06fb866eafda176c431
2026-01-04T20:19:28.343083Z
false
esp-rs/rfc2217-rs
https://github.com/esp-rs/rfc2217-rs/blob/6dd355b57726fcf211a5f06fb866eafda176c431/tests/unit_tests.rs
tests/unit_tests.rs
use parser::{Error, Event}; use rfc2217_rs::*; #[test] fn test_negotiation() { let mut neg: [u8; 3] = [0; negotiation::SIZE]; Negotiation { intent: negotiation::Intent::Will, option: negotiation::Option::Binary, } .serialize(&mut neg); let mut parser = Parser::new(); let mut result: Result<Option<Event>, Error> = Ok(None); for byte in neg { result = parser.process_byte(byte); } assert_eq!( result, Ok(Some(Event::Negotiation(Negotiation { intent: negotiation::Intent::Will, option: negotiation::Option::Binary, }))) ); } #[test] fn test_negotiation_unsupported() { let mut neg: [u8; 3] = [0; negotiation::SIZE]; Negotiation { intent: negotiation::Intent::Wont, option: negotiation::Option::Unsupported(66), } .serialize(&mut neg); let mut parser = Parser::new(); let mut result: Result<Option<Event>, Error> = Ok(None); for byte in neg { result = parser.process_byte(byte); } assert_eq!( result, Ok(Some(Event::Negotiation(Negotiation { intent: negotiation::Intent::Wont, option: negotiation::Option::Unsupported(neg[2]) }))) ); } #[test] fn test_command_unsupported() { let mut command = [0; command::SIZE]; Command::Unsupported(239).serialize(&mut command); let mut parser = Parser::new(); let mut result: Result<Option<Event>, Error> = Ok(None); for byte in command { result = parser.process_byte(byte); } assert_eq!( result, Ok(Some(Event::Command(Command::Unsupported(command[1])))) ); } #[test] fn test_signature_subnegotiation_containing_iac() { let mut subneg = [0; 13]; let mut signature = [0; subnegotiation::MAX_DATA_SIZE]; signature[..6].copy_from_slice(&[63, 111, 32, 255, 10, 44]); Subnegotiation::SetSignature { data: signature, size: 6, } .serialize_client(&mut subneg); let mut parser = Parser::new(); let mut result: Result<Option<Event>, Error> = Ok(None); for byte in subneg { result = parser.process_byte(byte); } assert_eq!( result, Ok(Some(Event::Subnegotiation(Subnegotiation::SetSignature { data: signature, size: 6 }))) ); } #[test] fn test_baud_subnegotiation_generated() { let mut subneg: [u8; 10] = [0; 10]; let expected_baudrate = 9600; Subnegotiation::SetBaudRate(expected_baudrate).serialize_client(&mut subneg); let mut parser = Parser::new(); let mut result: Result<Option<Event>, Error> = Ok(None); for byte in subneg { result = parser.process_byte(byte); } assert_eq!( result, Ok(Some(Event::Subnegotiation(Subnegotiation::SetBaudRate( expected_baudrate )))) ); } #[test] fn test_baud_subnegotiation_containing_iac() { let mut subneg: [u8; 10] = [0; 10]; let expected_baudrate = 0x0000FFFF; Subnegotiation::SetBaudRate(expected_baudrate).serialize_client(&mut subneg); for byte in subneg { print!("byte: {}", byte) } let mut parser = Parser::new(); let mut result: Result<Option<Event>, Error> = Ok(None); for byte in subneg { result = parser.process_byte(byte); } assert_eq!( result, Ok(Some(Event::Subnegotiation(Subnegotiation::SetBaudRate( expected_baudrate )))) ); } #[test] fn test_parity_subnegotiation() { let mut subneg = [0; 7]; Subnegotiation::SetParity(1).serialize_client(&mut subneg); let mut parser = Parser::new(); let mut result: Result<Option<Event>, Error> = Ok(None); for byte in subneg { result = parser.process_byte(byte); } assert_eq!( result, Ok(Some(Event::Subnegotiation(Subnegotiation::SetParity(1)))) ); } #[test] fn test_flow_control_suspend_subnegotiation() { let mut subneg = [0; 6]; Subnegotiation::FlowControlSuspend.serialize_client(&mut subneg); let mut parser = Parser::new(); let mut result: Result<Option<Event>, Error> = Ok(None); for byte in subneg { result = parser.process_byte(byte); } assert_eq!( result, Ok(Some(Event::Subnegotiation( Subnegotiation::FlowControlSuspend ))) ); }
rust
Apache-2.0
6dd355b57726fcf211a5f06fb866eafda176c431
2026-01-04T20:19:28.343083Z
false
bluecatengineering/dhcproto
https://github.com/bluecatengineering/dhcproto/blob/b4ea30defc01e7ae66e7075d6a0533d9bb9503dc/src/lib.rs
src/lib.rs
#![cfg_attr(not(test), no_std)] #![warn( missing_debug_implementations, // missing_docs, // some variants still missing docs missing_copy_implementations, rust_2018_idioms, unreachable_pub, non_snake_case, non_upper_case_globals )] #![allow(clippy::cognitive_complexity)] #![deny(rustdoc::broken_intra_doc_links)] #![doc(test( no_crate_inject, attr(deny(warnings, rust_2018_idioms), allow(dead_code, unused_variables)) ))] //! # dhcproto //! //! A DHCP parser and encoder for DHCPv4 and DHCPv6. `dhcproto` aims to be a functionally complete DHCP implementation. //! //! ## DHCPv4 //! //! ```rust //! use dhcproto::v4::{Message, Encoder, Decoder, Decodable, Encodable}; //! # fn main() -> Result<(), Box<dyn std::error::Error>> { //! // decode //! let bytes = dhcp_offer(); //! let msg = Message::decode(&mut Decoder::new(&bytes))?; //! // now encode //! let mut buf = Vec::new(); //! let mut e = Encoder::new(&mut buf); //! msg.encode(&mut e)?; //! # Ok(()) //! # } //! # fn dhcp_offer() -> Vec<u8> { //! # vec![ //! # 0x02, 0x01, 0x06, 0x00, 0x00, 0x00, 0x15, 0x5c, 0x00, 0x00, 0x80, 0x00, 0x00, 0x00, //! # 0x00, 0x00, 0xc0, 0xa8, 0x00, 0x03, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, //! # 0xcc, 0x00, 0x0a, 0xc4, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, //! # 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, //! # 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, //! # 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, //! # 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, //! # 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, //! # 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, //! # 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, //! # 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, //! # 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, //! # 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, //! # 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, //! # 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, //! # 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, //! # 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x63, 0x82, //! # 0x53, 0x63, 0x35, 0x01, 0x02, 0x36, 0x04, 0xc0, 0xa8, 0x00, 0x01, 0x33, 0x04, 0x00, //! # 0x00, 0x00, 0x3c, 0x3a, 0x04, 0x00, 0x00, 0x00, 0x1e, 0x3b, 0x04, 0x00, 0x00, 0x00, //! # 0x34, 0x01, 0x04, 0xff, 0xff, 0xff, 0x00, 0x03, 0x04, 0xc0, 0xa8, 0x00, 0x01, 0x06, //! # 0x08, 0xc0, 0xa8, 0x00, 0x01, 0xc0, 0xa8, 0x01, 0x01, 0xff, 0x00, 0x00, 0x00, 0x00, //! # 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, //! # ] //! # } //! ``` //! //! ## DHCPv6 //! //! ```rust //! use dhcproto::v6::{Message, Encoder, Decoder, Decodable, Encodable}; //! # fn main() -> Result<(), Box<dyn std::error::Error>> { //! // decode //! let bytes = solicit(); //! let msg = Message::decode(&mut Decoder::new(&bytes))?; //! // now encode //! let mut buf = Vec::new(); //! let mut e = Encoder::new(&mut buf); //! msg.encode(&mut e)?; //! # Ok(()) //! # } //! # fn solicit() -> Vec<u8> { //! # vec![ //! # 0x01, 0x10, 0x08, 0x74, 0x00, 0x01, 0x00, 0x0e, 0x00, 0x01, 0x00, 0x01, 0x1c, 0x39, //! # 0xcf, 0x88, 0x08, 0x00, 0x27, 0xfe, 0x8f, 0x95, 0x00, 0x06, 0x00, 0x04, 0x00, 0x17, //! # 0x00, 0x18, 0x00, 0x08, 0x00, 0x02, 0x00, 0x00, 0x00, 0x19, 0x00, 0x0c, 0x27, 0xfe, //! # 0x8f, 0x95, 0x00, 0x00, 0x0e, 0x10, 0x00, 0x00, 0x15, 0x18, //! # ] //! # } //! ``` extern crate alloc; pub use decoder::{Decodable, Decoder}; pub use encoder::{Encodable, Encoder}; pub mod decoder; pub mod encoder; pub mod error; pub mod v4; pub mod v6; pub use hickory_proto::ProtoError as NameError; pub use hickory_proto::rr::Name;
rust
MIT
b4ea30defc01e7ae66e7075d6a0533d9bb9503dc
2026-01-04T20:19:33.979507Z
false
bluecatengineering/dhcproto
https://github.com/bluecatengineering/dhcproto/blob/b4ea30defc01e7ae66e7075d6a0533d9bb9503dc/src/encoder.rs
src/encoder.rs
//! Encodable trait & Encoder use alloc::vec::Vec; use crate::error::{EncodeError, EncodeResult}; /// A trait for types which are deserializable to DHCP binary formats pub trait Encodable { /// encode type to buffer in Encoder fn encode(&self, e: &mut Encoder<'_>) -> EncodeResult<()>; /// encode this type into its binary form in a new `Vec` fn to_vec(&self) -> EncodeResult<Vec<u8>> { let mut buffer = Vec::with_capacity(512); let mut encoder = Encoder::new(&mut buffer); self.encode(&mut encoder)?; Ok(buffer) } } /// Encoder type, holds a mut ref to a buffer /// that it will write data to and an offset /// of the next position to write. /// /// This will start writing from the beginning of the buffer, *not* from the end. /// The buffer will be grown as needed. #[derive(Debug)] pub struct Encoder<'a> { buffer: &'a mut Vec<u8>, offset: usize, } impl<'a> Encoder<'a> { /// Create a new Encoder from a mutable buffer pub fn new(buffer: &'a mut Vec<u8>) -> Self { Self { buffer, offset: 0 } } /// Get a reference to the underlying buffer pub fn buffer(&self) -> &[u8] { self.buffer } /// Returns the slice of the underlying buffer that has been filled. pub fn buffer_filled(&self) -> &[u8] { &self.buffer[..self.offset] } /// Returns the number of bytes that have been written to the buffer. pub fn len_filled(&self) -> usize { self.offset } /// write bytes to buffer /// Return: /// number of bytes written pub fn write_slice(&mut self, bytes: &[u8]) -> EncodeResult<()> { let additional = bytes.len(); // space already reserved, we may not need this if self.offset + additional <= self.buffer.len() { // if self.offset == self.buffer.len() indexing can panic for (byte, b) in self.buffer[self.offset..].iter_mut().zip(bytes.iter()) { *byte = *b; } } else { let expected_len = self.buffer.len() + additional; self.buffer.reserve(additional); self.buffer.extend_from_slice(bytes); debug_assert!(self.buffer.len() == expected_len); } let index = self .offset .checked_add(additional) .ok_or(EncodeError::AddOverflow)?; self.offset = index; Ok(()) } /// Write const number of bytes to buffer pub fn write<const N: usize>(&mut self, bytes: [u8; N]) -> EncodeResult<()> { // TODO: refactor this and above method? // only difference is zip & extend let additional = bytes.len(); // space already reserved, we may not need this if self.offset + additional <= self.buffer.len() { // if self.offset == self.buffer.len() indexing can panic for (byte, b) in self.buffer[self.offset..].iter_mut().zip(bytes) { *byte = b; } } else { let expected_len = self.buffer.len() + additional; self.buffer.reserve(additional); self.buffer.extend(bytes); debug_assert!(self.buffer.len() == expected_len); } let index = self .offset .checked_add(additional) .ok_or(EncodeError::AddOverflow)?; self.offset = index; Ok(()) } /// write a u8 pub fn write_u8(&mut self, data: u8) -> EncodeResult<()> { self.write(data.to_be_bytes()) } /// write a u16 pub fn write_u16(&mut self, data: u16) -> EncodeResult<()> { self.write(data.to_be_bytes()) } /// write a u32 pub fn write_u32(&mut self, data: u32) -> EncodeResult<()> { self.write(data.to_be_bytes()) } /// write a u128 pub fn write_u128(&mut self, data: u128) -> EncodeResult<()> { self.write(data.to_be_bytes()) } /// write a u64 pub fn write_u64(&mut self, data: u64) -> EncodeResult<()> { self.write(data.to_be_bytes()) } /// write a i32 pub fn write_i32(&mut self, data: i32) -> EncodeResult<()> { self.write(data.to_be_bytes()) } /// Writes bytes to buffer and pads with 0 bytes up to some fill_len /// /// Returns /// Err - if bytes.len() is greater then fill_len pub fn write_fill_bytes(&mut self, bytes: &[u8], fill_len: usize) -> EncodeResult<()> { if bytes.len() > fill_len { return Err(EncodeError::StringSizeTooBig { len: bytes.len() }); } let nul_len = fill_len - bytes.len(); self.write_slice(bytes)?; for _ in 0..nul_len { self.write_u8(0)?; } Ok(()) } /// Writes value to buffer and pads with 0 bytes up to some fill_len /// if String is None then write fill_len 0 bytes /// /// Returns /// Err - if bytes.len() is greater then fill_len pub fn write_fill<T: AsRef<[u8]>>( &mut self, s: &Option<T>, fill_len: usize, ) -> EncodeResult<()> { match s { Some(sname) => { let bytes = sname.as_ref(); self.write_fill_bytes(bytes, fill_len)?; } None => { // should we keep some static [0;64] arrays around // to fill quickly? for _ in 0..fill_len { self.write_u8(0)?; } } } Ok(()) } } #[cfg(test)] mod tests { use super::*; #[test] fn basic_encode() -> EncodeResult<()> { let mut buf = alloc::vec![0, 1, 2, 3, 4, 5]; let mut enc = Encoder::new(&mut buf); enc.offset = 4; // write already reserved space enc.write_slice(&[5, 6])?; assert_eq!(enc.buffer, &mut alloc::vec![0, 1, 2, 3, 5, 6]); assert_eq!(enc.offset, 6); // reserve extra space enc.write_slice(&[7, 8])?; assert_eq!(enc.buffer, &mut alloc::vec![0, 1, 2, 3, 5, 6, 7, 8]); assert_eq!(enc.offset, 8); // start w/ empty buf let mut buf = alloc::vec![]; let mut enc = Encoder::new(&mut buf); // reserve space & write enc.write_slice(&[0, 1, 2, 3])?; assert_eq!(enc.buffer, &mut alloc::vec![0, 1, 2, 3]); assert_eq!(enc.offset, 4); Ok(()) } }
rust
MIT
b4ea30defc01e7ae66e7075d6a0533d9bb9503dc
2026-01-04T20:19:33.979507Z
false
bluecatengineering/dhcproto
https://github.com/bluecatengineering/dhcproto/blob/b4ea30defc01e7ae66e7075d6a0533d9bb9503dc/src/decoder.rs
src/decoder.rs
//! Decodable trait & Decoder use hickory_proto::{ rr::Name, serialize::binary::{BinDecodable, BinDecoder}, }; use crate::error::{DecodeError, DecodeResult}; use alloc::{borrow::ToOwned, ffi::CString, string::String, vec::Vec}; use core::{ array::TryFromSliceError, convert::TryInto, ffi::CStr, mem, net::{Ipv4Addr, Ipv6Addr}, str, }; /// A trait for types which are serializable to and from DHCP binary formats pub trait Decodable: Sized { /// Read the type from the stream fn decode(decoder: &mut Decoder<'_>) -> DecodeResult<Self>; /// Returns the object in binary form fn from_bytes(bytes: &[u8]) -> DecodeResult<Self> { let mut decoder = Decoder::new(bytes); Self::decode(&mut decoder) } } /// Decoder type. Wraps a buffer which only contains bytes that have not been read yet #[derive(Debug, Clone, Copy)] pub struct Decoder<'a> { buffer: &'a [u8], } impl<'a> Decoder<'a> { /// Create a new Decoder pub fn new(buffer: &'a [u8]) -> Self { Decoder { buffer } } /// peek at the next byte without advancing the internal pointer pub fn peek_u8(&self) -> DecodeResult<u8> { Ok(u8::from_be_bytes(self.peek::<{ mem::size_of::<u8>() }>()?)) } /// read a u8 pub fn read_u8(&mut self) -> DecodeResult<u8> { Ok(u8::from_be_bytes(self.read::<{ mem::size_of::<u8>() }>()?)) } /// read a u32 pub fn read_u32(&mut self) -> DecodeResult<u32> { Ok(u32::from_be_bytes( self.read::<{ mem::size_of::<u32>() }>()?, )) } /// read a i32 pub fn read_i32(&mut self) -> DecodeResult<i32> { Ok(i32::from_be_bytes( self.read::<{ mem::size_of::<i32>() }>()?, )) } /// read a u16 pub fn read_u16(&mut self) -> DecodeResult<u16> { Ok(u16::from_be_bytes( self.read::<{ mem::size_of::<u16>() }>()?, )) } /// read a u64 pub fn read_u64(&mut self) -> DecodeResult<u64> { Ok(u64::from_be_bytes( self.read::<{ mem::size_of::<u64>() }>()?, )) } /// read a `N` bytes into slice pub fn read<const N: usize>(&mut self) -> DecodeResult<[u8; N]> { if N > self.buffer.len() { return Err(DecodeError::NotEnoughBytes); } let (slice, remaining) = self.buffer.split_at(N); self.buffer = remaining; // can't panic-- condition checked above Ok(slice.try_into().unwrap()) } /// peek a `N` bytes into slice pub fn peek<const N: usize>(&self) -> DecodeResult<[u8; N]> { if N > self.buffer.len() { return Err(DecodeError::NotEnoughBytes); } // can't panic-- condition checked above Ok(self.buffer[..N].try_into().unwrap()) } /// read a `MAX` length bytes into nul terminated `CString` pub fn read_cstring<const MAX: usize>(&mut self) -> DecodeResult<Option<CString>> { let bytes = self.read::<MAX>()?; let nul_idx = bytes.iter().position(|&b| b == 0); match nul_idx { Some(0) => Ok(None), Some(n) => Ok(Some(CStr::from_bytes_with_nul(&bytes[..=n])?.to_owned())), // TODO: error? None => Ok(None), } } pub fn read_nul_bytes<const MAX: usize>(&mut self) -> DecodeResult<Option<Vec<u8>>> { let bytes = self.read::<MAX>()?; let nul_idx = bytes.iter().position(|&b| b == 0); match nul_idx { Some(0) => Ok(None), Some(n) => Ok(Some(bytes[..=n].to_vec())), // TODO: error? None => Ok(None), } } /// read `MAX` length bytes and read into utf-8 encoded `String` pub fn read_nul_string<const MAX: usize>(&mut self) -> DecodeResult<Option<String>> { Ok(self .read_nul_bytes::<MAX>()? .map(|ref bytes| str::from_utf8(bytes).map(|s| s.to_owned())) .transpose()?) } /// read a slice of bytes determined at runtime pub fn read_slice(&mut self, len: usize) -> DecodeResult<&'a [u8]> { if len > self.buffer.len() { return Err(DecodeError::NotEnoughBytes); } let (slice, remaining) = self.buffer.split_at(len); self.buffer = remaining; Ok(slice) } /// Read a utf-8 encoded String pub fn read_string(&mut self, len: usize) -> DecodeResult<String> { Ok(self.read_str(len)?.to_owned()) } /// Read a utf-8 encoded String pub fn read_str(&mut self, len: usize) -> DecodeResult<&str> { Ok(str::from_utf8(self.read_slice(len)?)?) } /// Read an ipv4 addr pub fn read_ipv4(&mut self, length: usize) -> DecodeResult<Ipv4Addr> { if length != 4 { return Err(DecodeError::NotEnoughBytes); } let bytes = self.read::<4>()?; Ok(bytes.into()) } /// Read a list of ipv4 addrs pub fn read_ipv4s(&mut self, length: usize) -> DecodeResult<Vec<Ipv4Addr>> { // must be multiple of 4 if !length.is_multiple_of(4) { return Err(DecodeError::NotEnoughBytes); } let ips = self.read_slice(length)?; Ok(ips .chunks(4) .map(|bytes| [bytes[0], bytes[1], bytes[2], bytes[3]].into()) .collect()) } /// Read a list of ipv6 addrs pub fn read_ipv6s(&mut self, length: usize) -> DecodeResult<Vec<Ipv6Addr>> { // must be multiple of 16 if !length.is_multiple_of(16) { return Err(DecodeError::NotEnoughBytes); } let ips = self.read_slice(length)?; // type annotations needed below Ok(ips .chunks(16) .map(|bytes| Ok::<_, TryFromSliceError>(TryInto::<[u8; 16]>::try_into(bytes)?.into())) .collect::<Result<Vec<Ipv6Addr>, _>>()?) } /// Read a list of ipv4 pairs pub fn read_pair_ipv4s(&mut self, length: usize) -> DecodeResult<Vec<(Ipv4Addr, Ipv4Addr)>> { // must be multiple of 8 if !length.is_multiple_of(8) { return Err(DecodeError::NotEnoughBytes); } let ips = self.read_slice(length)?; Ok(ips .chunks(8) .map(|bytes| { ( [bytes[0], bytes[1], bytes[2], bytes[3]].into(), [bytes[4], bytes[5], bytes[6], bytes[7]].into(), ) }) .collect()) } /// Read a list of domain `Name`s pub fn read_domains(&mut self, length: usize) -> DecodeResult<Vec<Name>> { let mut name_decoder = BinDecoder::new(self.read_slice(length)?); let mut names = Vec::new(); while let Ok(name) = Name::read(&mut name_decoder) { names.push(name); } Ok(names) } /// Read a bool pub fn read_bool(&mut self) -> DecodeResult<bool> { Ok(self.read_u8()? == 1) } /// return slice of buffer start at index of unread data pub fn buffer(&self) -> &[u8] { self.buffer } }
rust
MIT
b4ea30defc01e7ae66e7075d6a0533d9bb9503dc
2026-01-04T20:19:33.979507Z
false
bluecatengineering/dhcproto
https://github.com/bluecatengineering/dhcproto/blob/b4ea30defc01e7ae66e7075d6a0533d9bb9503dc/src/error.rs
src/error.rs
//! Error types for Encoding/Decoding use alloc::boxed::Box; use thiserror::Error; /// Convenience type for decode errors pub type DecodeResult<T> = Result<T, DecodeError>; /// Returned from types that decode #[derive(Error, Debug)] pub enum DecodeError { /// add overflow #[error("decoder checked_add failed")] AddOverflow, /// ran out of bytes #[error("parser ran out of data-- not enough bytes")] NotEnoughBytes, /// error converting from slice #[error("error converting from slice {0}")] SliceError(#[from] core::array::TryFromSliceError), /// error finding nul in string #[error("error getting null terminated string {0}")] NulError(#[from] core::ffi::FromBytesWithNulError), /// error converting to utf-8 #[error("error converting to UTF-8 {0}")] Utf8Error(#[from] core::str::Utf8Error), /// invalid data error #[error("invalid data error {0} msg {1}")] InvalidData(u32, &'static str), /// domain parse error #[error("domain parse error {0}")] DomainParseError(#[from] hickory_proto::ProtoError), /// Unknown decode error #[error("unknown error")] Unknown(Box<dyn core::error::Error + Send + Sync + 'static>), } /// Returned from types that encode #[derive(Error, Debug)] pub enum EncodeError { /// addition overflow #[error("encoder checked_add failed")] AddOverflow, /// string exceeds bounds #[error( "message is trying to write a string to the message that exceeds the max size of {len}" )] StringSizeTooBig { /// size of string len: usize, }, // /// io error // #[error("io error {0}")] // IoError(#[from] io::Error), /// DNS encoding error from hickory-dns #[error("domain encoding error {0}")] DomainEncodeError(#[from] hickory_proto::ProtoError), } /// Convenience type for encode errors pub type EncodeResult<T> = Result<T, EncodeError>;
rust
MIT
b4ea30defc01e7ae66e7075d6a0533d9bb9503dc
2026-01-04T20:19:33.979507Z
false
bluecatengineering/dhcproto
https://github.com/bluecatengineering/dhcproto/blob/b4ea30defc01e7ae66e7075d6a0533d9bb9503dc/src/v4/opcode.rs
src/v4/opcode.rs
use crate::{ decoder::{Decodable, Decoder}, encoder::{Encodable, Encoder}, error::{DecodeResult, EncodeResult}, }; #[cfg(feature = "serde")] use serde::{Deserialize, Serialize}; /// Opcode of Message #[cfg_attr(feature = "serde", derive(Serialize, Deserialize))] #[derive(Debug, Clone, Copy, PartialEq, Eq)] pub enum Opcode { /// BootRequest - <https://datatracker.ietf.org/doc/html/rfc1534#section-2> BootRequest, /// BootReply - <https://datatracker.ietf.org/doc/html/rfc1534#section-2> BootReply, /// Unknown or not yet implemented Unknown(u8), } impl Decodable for Opcode { fn decode(decoder: &mut Decoder<'_>) -> DecodeResult<Self> { Ok(decoder.read_u8()?.into()) } } impl Encodable for Opcode { fn encode(&self, e: &'_ mut Encoder<'_>) -> EncodeResult<()> { e.write_u8((*self).into()) } } impl From<u8> for Opcode { fn from(opcode: u8) -> Self { match opcode { 1 => Opcode::BootRequest, 2 => Opcode::BootReply, _ => Opcode::Unknown(opcode), } } } impl From<Opcode> for u8 { fn from(opcode: Opcode) -> Self { match opcode { Opcode::BootRequest => 1, Opcode::BootReply => 2, Opcode::Unknown(opcode) => opcode, } } }
rust
MIT
b4ea30defc01e7ae66e7075d6a0533d9bb9503dc
2026-01-04T20:19:33.979507Z
false
bluecatengineering/dhcproto
https://github.com/bluecatengineering/dhcproto/blob/b4ea30defc01e7ae66e7075d6a0533d9bb9503dc/src/v4/bulk_query.rs
src/v4/bulk_query.rs
use core::fmt; #[cfg(feature = "serde")] use serde::{Deserialize, Serialize}; /// Lease query data source flags #[cfg_attr(feature = "serde", derive(Serialize, Deserialize))] #[derive(Copy, Default, Clone, PartialEq, Eq)] pub struct DataSourceFlags(u8); impl fmt::Debug for DataSourceFlags { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { f.debug_struct("DataSourceFlags") .field("remote", &self.remote()) .finish() } } impl fmt::Display for DataSourceFlags { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { write!(f, "{self:?}") } } impl DataSourceFlags { /// Create new DataSourceFlags from u8 pub fn new(n: u8) -> Self { Self(n) } /// get the status of the remote flag pub fn remote(&self) -> bool { (self.0 & 0x01) == 1 } /// set the remote bit, returns a new DataSourceFlags pub fn set_remote(mut self) -> Self { self.0 |= 0x01; self } } impl From<u8> for DataSourceFlags { fn from(n: u8) -> Self { Self(n) } } impl From<DataSourceFlags> for u8 { fn from(f: DataSourceFlags) -> Self { f.0 } } #[cfg_attr(feature = "serde", derive(Serialize, Deserialize))] #[derive(Debug, Clone, Copy, PartialEq, Eq, Hash)] pub enum QueryState { Available, Active, Expired, Release, Abandoned, Reset, Remote, Transitioning, Unknown(u8), } impl From<u8> for QueryState { fn from(n: u8) -> Self { use QueryState::*; match n { 1 => Available, 2 => Active, 3 => Expired, 4 => Release, 5 => Abandoned, 6 => Reset, 7 => Remote, 8 => Transitioning, _ => Unknown(n), } } } impl From<QueryState> for u8 { fn from(state: QueryState) -> Self { use QueryState as Q; match state { Q::Available => 1, Q::Active => 2, Q::Expired => 3, Q::Release => 4, Q::Abandoned => 5, Q::Reset => 6, Q::Remote => 7, Q::Transitioning => 8, Q::Unknown(code) => code, } } } #[cfg_attr(feature = "serde", derive(Serialize, Deserialize))] #[derive(Debug, Clone, Copy, PartialEq, Eq, Hash)] pub enum Code { Success, UnspecFail, QueryTerminated, MalformedQuery, NotAllowed, Unknown(u8), } impl From<u8> for Code { fn from(n: u8) -> Self { use Code::*; match n { 0 => Success, 1 => UnspecFail, 2 => QueryTerminated, 3 => MalformedQuery, 4 => NotAllowed, _ => Unknown(n), } } } impl From<Code> for u8 { fn from(code: Code) -> Self { use Code as C; match code { C::Success => 0, C::UnspecFail => 1, C::QueryTerminated => 2, C::MalformedQuery => 3, C::NotAllowed => 4, C::Unknown(code) => code, } } } #[cfg(test)] mod tests { use super::*; #[test] fn test_datasourceflags() { let flag = DataSourceFlags::default(); assert_eq!(flag.0, 0); let flag = flag.set_remote(); assert_eq!(flag.0, 0x01); assert!(flag.remote()); let flag = DataSourceFlags::new(0x80).set_remote(); assert_eq!(flag.0, 0x81); } }
rust
MIT
b4ea30defc01e7ae66e7075d6a0533d9bb9503dc
2026-01-04T20:19:33.979507Z
false
bluecatengineering/dhcproto
https://github.com/bluecatengineering/dhcproto/blob/b4ea30defc01e7ae66e7075d6a0533d9bb9503dc/src/v4/htype.rs
src/v4/htype.rs
use crate::{ decoder::{Decodable, Decoder}, encoder::{Encodable, Encoder}, error::{DecodeResult, EncodeResult}, }; #[cfg(feature = "serde")] use serde::{Deserialize, Serialize}; /// Hardware type of message #[cfg_attr(feature = "serde", derive(Serialize, Deserialize))] #[derive(Debug, Copy, Hash, Clone, PartialEq, Eq)] pub enum HType { /// 1 Ethernet Eth, /// 2 Experimental Ethernet ExperimentalEth, /// 3 Amateur Radio AX25 AmRadioAX25, /// 4 Proteon Token Ring ProteonTokenRing, /// 5 Chaos Chaos, /// 6 IEEE.802 IEEE802, /// 7 ARCNET ARCNET, /// 8 Hyperchannel Hyperchannel, /// 9 LANSTAR Lanstar, /// 10 Autonet Short Addr AutonetShortAddr, /// 11 LocalTalk LocalTalk, /// 12 LocalNet LocalNet, /// 13 Ultralink Ultralink, /// 14 SMDS SMDS, /// 15 FrameRelay FrameRelay, /// 17 HDLC HDLC, /// 18 FibreChannel FibreChannel, /// 20 SerialLine SerialLine, /// 22 Mil STD MilStd188220, /// 23 Metricom Metricom, /// 25 MAPOS MAPOS, /// 26 Twinaxial Twinaxial, /// 30 ARPSec ARPSec, /// 31 IPsec tunnel IPsecTunnel, /// 32 Infiniband Infiniband, /// 34 WeigandInt WiegandInt, /// 35 PureIP PureIP, /// Unknown or not yet implemented htype Unknown(u8), } impl From<u8> for HType { fn from(n: u8) -> Self { use HType::*; match n { 1 => Eth, 2 => ExperimentalEth, 3 => AmRadioAX25, 4 => ProteonTokenRing, 5 => Chaos, 6 => IEEE802, 7 => ARCNET, 8 => Hyperchannel, 9 => Lanstar, 10 => AutonetShortAddr, 11 => LocalTalk, 12 => LocalNet, 13 => Ultralink, 14 => SMDS, 15 => FrameRelay, 17 => HDLC, 18 => FibreChannel, 20 => SerialLine, 22 => MilStd188220, 23 => Metricom, 25 => MAPOS, 26 => Twinaxial, 30 => ARPSec, 31 => IPsecTunnel, 32 => Infiniband, 34 => WiegandInt, 35 => PureIP, n => Unknown(n), } } } impl From<HType> for u8 { fn from(n: HType) -> Self { use HType as H; match n { H::Eth => 1, H::ExperimentalEth => 2, H::AmRadioAX25 => 3, H::ProteonTokenRing => 4, H::Chaos => 5, H::IEEE802 => 6, H::ARCNET => 7, H::Hyperchannel => 8, H::Lanstar => 9, H::AutonetShortAddr => 10, H::LocalTalk => 11, H::LocalNet => 12, H::Ultralink => 13, H::SMDS => 14, H::FrameRelay => 15, H::HDLC => 17, H::FibreChannel => 18, H::SerialLine => 20, H::MilStd188220 => 22, H::Metricom => 23, H::MAPOS => 25, H::Twinaxial => 26, H::ARPSec => 30, H::IPsecTunnel => 31, H::Infiniband => 32, H::WiegandInt => 34, H::PureIP => 35, H::Unknown(n) => n, } } } impl Decodable for HType { fn decode(decoder: &mut Decoder<'_>) -> DecodeResult<Self> { Ok(decoder.read_u8()?.into()) } } impl Encodable for HType { fn encode(&self, e: &mut Encoder<'_>) -> EncodeResult<()> { e.write_u8((*self).into()) } }
rust
MIT
b4ea30defc01e7ae66e7075d6a0533d9bb9503dc
2026-01-04T20:19:33.979507Z
false
bluecatengineering/dhcproto
https://github.com/bluecatengineering/dhcproto/blob/b4ea30defc01e7ae66e7075d6a0533d9bb9503dc/src/v4/flags.rs
src/v4/flags.rs
use core::fmt; use crate::{ decoder::{Decodable, Decoder}, encoder::{Encodable, Encoder}, error::{DecodeResult, EncodeResult}, }; #[cfg(feature = "serde")] use serde::{Deserialize, Serialize}; /// Represents available flags on message #[cfg_attr(feature = "serde", derive(Serialize, Deserialize))] #[derive(Copy, Default, Clone, PartialEq, Eq)] pub struct Flags(u16); impl fmt::Debug for Flags { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { f.debug_struct("Flags") .field("broadcast", &self.broadcast()) .finish() } } impl fmt::Display for Flags { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { write!(f, "{self:?}") } } impl Flags { /// Create new Flags from u16 pub fn new(n: u16) -> Self { Self(n) } /// get the status of the broadcast flag pub fn broadcast(&self) -> bool { (self.0 & 0x80_00) >> (u16::BITS - 1) == 1 } /// set the broadcast bit, returns a new Flags pub fn set_broadcast(mut self) -> Self { self.0 |= 0x80_00; self } } impl From<u16> for Flags { fn from(n: u16) -> Self { Self(n) } } impl From<Flags> for u16 { fn from(f: Flags) -> Self { f.0 } } impl Decodable for Flags { fn decode(decoder: &mut Decoder<'_>) -> DecodeResult<Self> { Ok(decoder.read_u16()?.into()) } } impl Encodable for Flags { fn encode(&self, e: &mut Encoder<'_>) -> EncodeResult<()> { e.write_u16((*self).into()) } } #[cfg(test)] mod tests { use super::*; #[test] fn test_broadcast() { let flag = Flags::default(); assert_eq!(flag.0, 0); let flag = flag.set_broadcast(); assert_eq!(flag.0, 0x80_00); assert!(flag.broadcast()); let flag = Flags::new(0x00_20).set_broadcast(); assert_eq!(flag.0, 0x80_20); } }
rust
MIT
b4ea30defc01e7ae66e7075d6a0533d9bb9503dc
2026-01-04T20:19:33.979507Z
false
bluecatengineering/dhcproto
https://github.com/bluecatengineering/dhcproto/blob/b4ea30defc01e7ae66e7075d6a0533d9bb9503dc/src/v4/relay.rs
src/v4/relay.rs
//! # relay use alloc::{collections::BTreeMap, vec::Vec}; use core::{fmt, net::Ipv4Addr}; use crate::{Decodable, Encodable}; #[cfg(feature = "serde")] use serde::{Deserialize, Serialize}; /// Collection of relay agent information /// /// You can create/modify it, then insert into a message opts section /// in [`DhcpOption::RelayAgentInformation`] /// /// ```rust /// use dhcproto::v4::{self, relay::{RelayInfo, RelayAgentInformation}}; /// /// let mut info = RelayAgentInformation::default(); /// info.insert(RelayInfo::LinkSelection("1.2.3.4".parse().unwrap())); /// let mut msg = v4::Message::default(); /// msg.opts_mut() /// .insert(v4::DhcpOption::RelayAgentInformation(info)); /// ``` /// /// [`DhcpOption::RelayAgentInformation`]: crate::v4::DhcpOption::RelayAgentInformation #[cfg_attr(feature = "serde", derive(Serialize, Deserialize))] #[derive(Debug, Default, Clone, PartialEq, Eq)] pub struct RelayAgentInformation(BTreeMap<RelayCode, RelayInfo>); impl RelayAgentInformation { /// Get the data for a particular [`RelayCode`] /// /// [`RelayCode`]: crate::v4::relay::RelayCode pub fn get(&self, code: RelayCode) -> Option<&RelayInfo> { self.0.get(&code) } /// Get the mutable data for a particular [`RelayCode`] /// /// [`RelayCode`]: crate::v4::relay::RelayCode pub fn get_mut(&mut self, code: RelayCode) -> Option<&mut RelayInfo> { self.0.get_mut(&code) } /// remove sub option pub fn remove(&mut self, code: RelayCode) -> Option<RelayInfo> { self.0.remove(&code) } /// insert a new [`RelayInfo`] /// /// [`RelayInfo`]: crate::v4::relay::RelayInfo pub fn insert(&mut self, info: RelayInfo) -> Option<RelayInfo> { self.0.insert((&info).into(), info) } /// iterate over entries pub fn iter(&self) -> impl Iterator<Item = (&RelayCode, &RelayInfo)> { self.0.iter() } /// iterate mutably over entries pub fn iter_mut(&mut self) -> impl Iterator<Item = (&RelayCode, &mut RelayInfo)> { self.0.iter_mut() } /// clear all options pub fn clear(&mut self) { self.0.clear() } /// Returns `true` if there are no options pub fn is_empty(&self) -> bool { self.0.is_empty() } /// Retans only the elements specified by the predicate pub fn retain<F>(&mut self, pred: F) where F: FnMut(&RelayCode, &mut RelayInfo) -> bool, { self.0.retain(pred) } } impl Decodable for RelayAgentInformation { fn decode(d: &mut crate::Decoder<'_>) -> super::DecodeResult<Self> { let mut opts = BTreeMap::new(); while let Ok(opt) = RelayInfo::decode(d) { opts.insert(RelayCode::from(&opt), opt); } Ok(RelayAgentInformation(opts)) } } impl Encodable for RelayAgentInformation { fn encode(&self, e: &mut crate::Encoder<'_>) -> super::EncodeResult<()> { self.0.iter().try_for_each(|(_, info)| info.encode(e)) } } #[cfg_attr(feature = "serde", derive(Serialize, Deserialize))] #[derive(Debug, Clone, PartialEq, Eq)] pub enum RelayInfo { /// 1 - <https://datatracker.ietf.org/doc/html/rfc3046> AgentCircuitId(Vec<u8>), /// 2 - <https://datatracker.ietf.org/doc/html/rfc3046> AgentRemoteId(Vec<u8>), /// 4 - <https://datatracker.ietf.org/doc/html/rfc3256> DocsisDeviceClass(u32), /// 5 - <https://datatracker.ietf.org/doc/html/rfc3527> LinkSelection(Ipv4Addr), /// 6 - <https://datatracker.ietf.org/doc/html/rfc3993#section-3.1> SubscriberId(Vec<u8>), /// 10 - <https://datatracker.ietf.org/doc/html/rfc5010#section-3> RelayAgentFlags(RelayFlags), /// 11 - <https://datatracker.ietf.org/doc/html/rfc5107#section-4> ServerIdentifierOverride(Ipv4Addr), Unknown(UnknownInfo), // TODO: not tackling this at the moment // 7 - <https://datatracker.ietf.org/doc/html/rfc4014> // RadiusAttributes, // 8 - <https://datatracker.ietf.org/doc/html/rfc4030#section-4> // 9 // VendorSpecificInformation(Vec<u8>), // Authentication(Authentication), // 151 - <https://datatracker.ietf.org/doc/html/rfc6607> // VirtualSubnet(VirtualSubnet), // 152 // VirtualSubnetControl(u8), } impl Decodable for RelayInfo { fn decode(d: &mut crate::Decoder<'_>) -> super::DecodeResult<Self> { use RelayInfo::*; // read the code first, determines the variant Ok(match d.read_u8()?.into() { RelayCode::AgentCircuitId => { let len = d.read_u8()? as usize; let data = d.read_slice(len)?.to_vec(); AgentCircuitId(data) } RelayCode::AgentRemoteId => { let len = d.read_u8()? as usize; let data = d.read_slice(len)?.to_vec(); AgentRemoteId(data) } RelayCode::DocsisDeviceClass => { let _ = d.read_u8()?; let device_id = d.read_u32()?; DocsisDeviceClass(device_id) } RelayCode::LinkSelection => { let len = d.read_u8()? as usize; LinkSelection(d.read_ipv4(len)?) } RelayCode::SubscriberId => { let len = d.read_u8()? as usize; let data = d.read_slice(len)?.to_vec(); SubscriberId(data) } RelayCode::RelayAgentFlags => { let _len = d.read_u8()?; let flags = d.read_u8()?; RelayAgentFlags(flags.into()) } RelayCode::ServerIdentifierOverride => { let len = d.read_u8()? as usize; ServerIdentifierOverride(d.read_ipv4(len)?) } // we have codes for these but not full type definitions yet code @ (RelayCode::Authentication | RelayCode::VirtualSubnet | RelayCode::VirtualSubnetControl | RelayCode::RadiusAttributes | RelayCode::VendorSpecificInformation) => { let length = d.read_u8()?; let bytes = d.read_slice(length as usize)?.to_vec(); Unknown(UnknownInfo { code: code.into(), data: bytes, }) } // not yet implemented RelayCode::Unknown(code) => { let length = d.read_u8()?; let bytes = d.read_slice(length as usize)?.to_vec(); Unknown(UnknownInfo { code, data: bytes }) } }) } } impl Encodable for RelayInfo { fn encode(&self, e: &mut crate::Encoder<'_>) -> super::EncodeResult<()> { use RelayInfo as R; let code: RelayCode = self.into(); e.write_u8(code.into())?; match self { R::AgentCircuitId(id) | R::AgentRemoteId(id) | R::SubscriberId(id) => { // length of bytes stored in Vec e.write_u8(id.len() as u8)?; e.write_slice(id)? } R::DocsisDeviceClass(n) => { e.write_u8(4)?; e.write_u32(*n)? } R::LinkSelection(addr) | R::ServerIdentifierOverride(addr) => { e.write_u8(4)?; e.write_u32((*addr).into())? } R::RelayAgentFlags(flags) => { e.write_u8(1)?; e.write_u8((*flags).into())? } // not yet implemented R::Unknown(opt) => { // length of bytes stored in Vec e.write_u8(opt.data.len() as u8)?; e.write_slice(&opt.data)? } }; Ok(()) } } #[cfg_attr(feature = "serde", derive(Serialize, Deserialize))] #[derive(Copy, Default, Clone, PartialEq, Eq)] pub struct RelayFlags(u8); impl fmt::Debug for RelayFlags { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { f.debug_struct("RelayFlags") .field("unicast", &self.unicast()) .finish() } } impl fmt::Display for RelayFlags { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { write!(f, "{self:?}") } } impl RelayFlags { /// Create new RelayFlags from u8 pub fn new(n: u8) -> Self { Self(n) } /// get the status of the unicast flag pub fn unicast(&self) -> bool { (self.0 & 0x80) >> (u8::BITS - 1) == 1 } /// set the unicast bit, returns a new Flags pub fn set_unicast(mut self) -> Self { self.0 |= 0x80; self } } impl From<u8> for RelayFlags { fn from(n: u8) -> Self { Self(n) } } impl From<RelayFlags> for u8 { fn from(f: RelayFlags) -> Self { f.0 } } /// An as-of-yet unimplemented relay info #[cfg_attr(feature = "serde", derive(Serialize, Deserialize))] #[derive(Debug, Clone, PartialEq, Eq, Hash)] pub struct UnknownInfo { code: u8, data: Vec<u8>, } impl UnknownInfo { pub fn new(code: RelayCode, data: Vec<u8>) -> Self { Self { code: code.into(), data, } } /// return the relay code pub fn code(&self) -> RelayCode { self.code.into() } /// return the data for this code pub fn data(&self) -> &[u8] { &self.data } /// take ownership and return the parts of this pub fn into_parts(self) -> (RelayCode, Vec<u8>) { (self.code.into(), self.data) } } /// relay code, represented as a u8 #[cfg_attr(feature = "serde", derive(Serialize, Deserialize))] #[derive(Debug, Copy, Clone, PartialEq, Eq, Hash)] pub enum RelayCode { AgentCircuitId, AgentRemoteId, DocsisDeviceClass, LinkSelection, SubscriberId, RadiusAttributes, Authentication, VendorSpecificInformation, RelayAgentFlags, ServerIdentifierOverride, VirtualSubnet, VirtualSubnetControl, /// unknown/unimplemented message type Unknown(u8), } impl PartialOrd for RelayCode { fn partial_cmp(&self, other: &Self) -> Option<core::cmp::Ordering> { Some(self.cmp(other)) } } impl Ord for RelayCode { fn cmp(&self, other: &Self) -> core::cmp::Ordering { u8::from(*self).cmp(&u8::from(*other)) } } impl From<u8> for RelayCode { fn from(n: u8) -> Self { use RelayCode::*; match n { 1 => AgentCircuitId, 2 => AgentRemoteId, 4 => DocsisDeviceClass, 5 => LinkSelection, 6 => SubscriberId, 7 => RadiusAttributes, 8 => Authentication, 9 => VendorSpecificInformation, 10 => RelayAgentFlags, 11 => ServerIdentifierOverride, 151 => VirtualSubnet, 152 => VirtualSubnetControl, _ => Unknown(n), } } } impl From<RelayCode> for u8 { fn from(code: RelayCode) -> Self { use RelayCode as R; match code { R::AgentCircuitId => 1, R::AgentRemoteId => 2, R::DocsisDeviceClass => 4, R::LinkSelection => 5, R::SubscriberId => 6, R::RadiusAttributes => 7, R::Authentication => 8, R::VendorSpecificInformation => 9, R::RelayAgentFlags => 10, R::ServerIdentifierOverride => 11, R::VirtualSubnet => 151, R::VirtualSubnetControl => 152, R::Unknown(n) => n, } } } impl From<&RelayInfo> for RelayCode { fn from(info: &RelayInfo) -> Self { use RelayInfo as R; match info { R::AgentCircuitId(_) => RelayCode::AgentCircuitId, R::AgentRemoteId(_) => RelayCode::AgentRemoteId, R::DocsisDeviceClass(_) => RelayCode::DocsisDeviceClass, R::LinkSelection(_) => RelayCode::LinkSelection, R::SubscriberId(_) => RelayCode::SubscriberId, R::RelayAgentFlags(_) => RelayCode::RelayAgentFlags, R::ServerIdentifierOverride(_) => RelayCode::ServerIdentifierOverride, R::Unknown(unknown) => RelayCode::Unknown(unknown.code), } } } #[cfg(test)] mod tests { use super::*; type Result<T> = std::result::Result<T, Box<dyn std::error::Error>>; #[test] fn test_unicast() { let flag = RelayFlags::default(); assert_eq!(flag.0, 0); let flag = flag.set_unicast(); assert_eq!(flag.0, 0x80); let flag = RelayFlags::new(0x00).set_unicast(); assert_eq!(flag.0, 0x80); assert!(flag.unicast()); } fn test_opt(opt: RelayInfo, actual: Vec<u8>) -> Result<()> { let mut out = vec![]; let mut enc = crate::Encoder::new(&mut out); opt.encode(&mut enc)?; println!("{:?}", enc.buffer()); assert_eq!(out, actual); let buf = RelayInfo::decode(&mut crate::Decoder::new(&out))?; assert_eq!(buf, opt); Ok(()) } #[test] fn test_ip() -> Result<()> { test_opt( RelayInfo::LinkSelection("192.168.0.1".parse::<Ipv4Addr>().unwrap()), vec![5, 4, 192, 168, 0, 1], )?; Ok(()) } #[test] fn test_str() -> Result<()> { test_opt( RelayInfo::AgentCircuitId(vec![0, 1, 2, 3, 4]), vec![1, 5, 0, 1, 2, 3, 4], )?; Ok(()) } #[test] fn test_remote() -> Result<()> { test_opt( RelayInfo::AgentRemoteId(vec![0, 1, 2, 3, 4]), vec![2, 5, 0, 1, 2, 3, 4], )?; Ok(()) } #[test] fn test_flags() -> Result<()> { test_opt( RelayInfo::RelayAgentFlags(RelayFlags::default().set_unicast()), vec![10, 1, 0x80], )?; Ok(()) } #[test] fn test_unknown() -> Result<()> { test_opt( RelayInfo::Unknown(UnknownInfo::new(RelayCode::Unknown(149), vec![1, 2, 3, 4])), vec![149, 4, 1, 2, 3, 4], )?; Ok(()) } }
rust
MIT
b4ea30defc01e7ae66e7075d6a0533d9bb9503dc
2026-01-04T20:19:33.979507Z
false
bluecatengineering/dhcproto
https://github.com/bluecatengineering/dhcproto/blob/b4ea30defc01e7ae66e7075d6a0533d9bb9503dc/src/v4/options.rs
src/v4/options.rs
use alloc::{ borrow::Cow, collections::BTreeMap, string::{String, ToString}, vec::Vec, }; use core::{iter, net::Ipv4Addr}; use crate::{ decoder::{Decodable, Decoder}, encoder::{Encodable, Encoder}, error::{DecodeResult, EncodeResult}, v4::bulk_query, v4::{fqdn, relay}, }; use hickory_proto::{ rr::Name, serialize::binary::{BinDecodable, BinDecoder, BinEncodable, BinEncoder, EncodeMode}, }; use ipnet::Ipv4Net; #[cfg(feature = "serde")] use serde::{Deserialize, Serialize}; // declares DHCP Option codes. // generates: // * the `OptionCode` enum and its From<u8>, Into<u8> // * the DhcpOption enum // * From<&DhcpOption> for OptionCode // // Syntax is {N, Name, "DocString" [,(T0,..TN,)]} // where: // * N is the numeric code associated with this option // * Name is the name to use for the enum variants // * "Docstring" is the documentation string that will be added to the variant in the OptionCode enum // * (T0,..TN) is the associated variables (if any). e.g. Ipv4Addr for "SubnetMask" or bool for "IpForwarding". // can contain more than one type but needs to be enclosed in parenthesis even if it's just a single variable. dhcproto_macros::declare_codes!( {0, Pad, "Padding"}, {1, SubnetMask, "Subnet Mask", (Ipv4Addr)}, {2, TimeOffset, "Time Offset", (i32)}, {3, Router, "Router", (Vec<Ipv4Addr>)}, {4, TimeServer, "Router", (Vec<Ipv4Addr>)}, {5, NameServer, "Name Server", (Vec<Ipv4Addr>)}, {6, DomainNameServer, "Name Server", (Vec<Ipv4Addr>)}, {7, LogServer, "Log Server", (Vec<Ipv4Addr>)}, {8, QuoteServer, "Quote Server", (Vec<Ipv4Addr>)}, {9, LprServer, "LPR Server", (Vec<Ipv4Addr>)}, {10, ImpressServer, "Impress server", (Vec<Ipv4Addr>)}, {11, ResourceLocationServer, "Resource Location Server", (Vec<Ipv4Addr>)}, {12, Hostname, "Host name", (String)}, {13, BootFileSize, "Boot file size", (u16)}, {14, MeritDumpFile, "Merit Dump File", (String)}, {15, DomainName, "Domain Name", (String)}, {16, SwapServer, "Swap server", (Ipv4Addr)}, {17, RootPath, "Root Path", (String)}, {18, ExtensionsPath, "Extensions path", (String)}, {19, IpForwarding, "IP forwarding", (bool)}, {20, NonLocalSrcRouting, "Non-local source routing", (bool)}, {21, PolicyFilter, "Policy Filter", (Vec<(Ipv4Addr, Ipv4Addr)>)}, {22, MaxDatagramSize, "Max Datagram reassembly size", (u16)}, {23, DefaultIpTtl, "Ip TTL", (u8)}, {24, PathMtuAgingTimeout, "Path MTU Aging Timeout", (u32)}, {25, PathMtuPlateauTable, "Path MTU Plateau Table", (Vec<u16>)}, {26, InterfaceMtu, "Interface MTU", (u16)}, {27, AllSubnetsLocal, "All Subnets Local", (bool)}, {28, BroadcastAddr, "Broadcast address", (Ipv4Addr)}, {29, PerformMaskDiscovery, "Perform mask discovery", (bool)}, {30, MaskSupplier, "Mask supplier", (bool)}, {31, PerformRouterDiscovery, "Perform router discovery", (bool)}, {32, RouterSolicitationAddr, "Router solicitation address", (Ipv4Addr)}, {33, StaticRoutingTable, "Static routing table", (Vec<(Ipv4Addr, Ipv4Addr)>)}, {34, TrailerEncapsulated, "Trailer Encapsulated", (bool)}, {35, ArpCacheTimeout, "ARP timeout", (u32)}, {36, EthernetEncapsulation, "Ethernet encapsulation", (bool)}, {37, DefaultTcpTtl, "Default TCP TTL", (u8)}, {38, TcpKeepaliveInterval, "TCP keepalive interval", (u32)}, {39, TcpKeepaliveGarbage, "TCP keealive garbage", (bool)}, {40, NisDomain, "Network information service domain", (String)}, {41, NisServers, "NIS servers", (Vec<Ipv4Addr>)}, {42, NtpServers, "NTP servers", (Vec<Ipv4Addr>)}, {43, VendorExtensions, "Vendor Extensions - can contain encapsulated options", (Vec<u8>)}, // TODO: Hashmap<u8, UnknownOption>? {44, NetBiosNameServers, "NetBIOS over TCP/IP name server", (Vec<Ipv4Addr>)}, {45, NetBiosDatagramDistributionServer, "NetBIOS over TCP/IP Datagram Distribution Server", (Vec<Ipv4Addr>)}, {46, NetBiosNodeType, "NetBIOS over TCP/IP Node Type", (NodeType)}, {47, NetBiosScope, "NetBIOS over TCP/IP Scope", (String)}, {48, XFontServer, "X Window System Font Server", (Vec<Ipv4Addr>)}, {49, XDisplayManager, "Window System Display Manager", (Vec<Ipv4Addr>)}, {50, RequestedIpAddress, "Requested IP Address", (Ipv4Addr)}, {51, AddressLeaseTime, "IP Address Lease Time", (u32)}, {52, OptionOverload, "Option Overload", (u8)}, {53, MessageType, "Message Type", (MessageType)}, {54, ServerIdentifier, "Server Identifier", (Ipv4Addr)}, {55, ParameterRequestList, "Parameter Request List", (Vec<OptionCode>)}, {56, Message, "Message", (String)}, {57, MaxMessageSize, "Maximum DHCP Message Size", (u16)}, {58, Renewal, "Renewal (T1) Time Value", (u32)}, {59, Rebinding, "Rebinding (T2) Time Value", (u32)}, {60, ClassIdentifier, "Class-identifier", (Vec<u8>)}, {61, ClientIdentifier, "Client Identifier", (Vec<u8>)}, {62, NwipDomainName, "Netware/IP Domain Name", (String)}, {63, NwipInformation, "Netware/IP Information - <https://www.rfc-editor.org/rfc/rfc2242.html>", (Vec<u8>)}, // TODO: https://www.rfc-editor.org/rfc/rfc2242.html sub opts {64, NispServiceDomain, "NIS+ Domain Option", (String)}, {65, NispServers, "NIS+ Server Addr", (Vec<Ipv4Addr>)}, {66, TFTPServerName, "TFTP Server Name - <https://www.rfc-editor.org/rfc/rfc2132.html>", (Vec<u8>)}, {67, BootfileName, "Bootfile Name - <https://www.rfc-editor.org/rfc/rfc2132.html>", (Vec<u8>)}, {68, MobileIpHomeAgent, "Mobile IP Home Agent", (Vec<Ipv4Addr>)}, {69, SmtpServer, "SMTP Server Option", (Vec<Ipv4Addr>)}, {70, Pop3Server, "Pop3 Server Option", (Vec<Ipv4Addr>)}, {71, NntpServer, "NNTP Server Option", (Vec<Ipv4Addr>)}, {72, WwwServer, "WWW Server Option", (Vec<Ipv4Addr>)}, {73, DefaultFingerServer, "Default Finger Option", (Vec<Ipv4Addr>)}, {74, IrcServer, "IRC Server Option", (Vec<Ipv4Addr>)}, {75, StreetTalkServer, "StreetTalk Server Option", (Vec<Ipv4Addr>)}, {76, StreetTalkDirectoryAssistance, "StreetTalk Directory Insistance (STDA) Option", (Vec<Ipv4Addr>)}, // TODO: split user-class into individual classes [len | <class>, ...] {77, UserClass, "User Class Option - <https://www.rfc-editor.org/rfc/rfc3004.html>", (Vec<u8>)}, {80, RapidCommit, "Rapid Commit - <https://www.rfc-editor.org/rfc/rfc4039.html>"}, {81, ClientFQDN, "FQDN - <https://datatracker.ietf.org/doc/html/rfc4702>", (fqdn::ClientFQDN)}, {82, RelayAgentInformation, "Relay Agent Information - <https://datatracker.ietf.org/doc/html/rfc3046>", (relay::RelayAgentInformation)}, {88, BcmsControllerNames, "Broadcast Multicast Controller Names - <https://www.rfc-editor.org/rfc/rfc4280.html#section-4.1>", (Vec<Name>)}, {89, BcmsControllerAddrs, "Broadcast Mutlicast Controller Address - <https://www.rfc-editor.org/rfc/rfc4280.html#section-4.3>", (Vec<Ipv4Addr>)}, {91, ClientLastTransactionTime, "client-last-transaction-time - <https://www.rfc-editor.org/rfc/rfc4388.html#section-6.1>", (u32)}, {92, AssociatedIp, "associated-ip - <https://www.rfc-editor.org/rfc/rfc4388.html#section-6.1>", (Vec<Ipv4Addr>)}, {93, ClientSystemArchitecture, "Client System Architecture - <https://www.rfc-editor.org/rfc/rfc4578.html>", (Architecture)}, {94, ClientNetworkInterface, "Client Network Interface - <https://www.rfc-editor.org/rfc/rfc4578.html>", (u8, u8, u8)}, {97, ClientMachineIdentifier, "Client Machine Identifier - <https://www.rfc-editor.org/rfc/rfc4578.html>", (Vec<u8>)}, {106, Ipv6OnlyPreferred, "IPv6-Only Preferred - <https://datatracker.ietf.org/doc/html/rfc8925>", (u32)}, {114, CaptivePortal, "Captive Portal - <https://datatracker.ietf.org/doc/html/rfc8910>", (String)}, {116, DisableSLAAC, "Disable Stateless Autoconfig for Ipv4 - <https://datatracker.ietf.org/doc/html/rfc2563>", (AutoConfig)}, {118, SubnetSelection, "Subnet selection - <https://datatracker.ietf.org/doc/html/rfc3011>", (Ipv4Addr)}, {119, DomainSearch, "Domain Search - <https://www.rfc-editor.org/rfc/rfc3397.html>", (Vec<Name>)}, {121, ClasslessStaticRoute, "Classless Static Route - <https://www.rfc-editor.org/rfc/rfc3442>", (Vec<(Ipv4Net, Ipv4Addr)>)}, {150, TFTPServerAddress, "TFTP Server Address - <https://www.rfc-editor.org/rfc/rfc5859.html>", (Ipv4Addr)}, {151, BulkLeaseQueryStatusCode, "BLQ status-code - <https://www.rfc-editor.org/rfc/rfc6926.html#section-6.2.2>", (bulk_query::Code, String)}, {152, BulkLeaseQueryBaseTime, "BLQ base time - <https://www.rfc-editor.org/rfc/rfc6926.html#section-6.2.3>", (u32)}, {153, BulkLeasQueryStartTimeOfState, "BLQ start time of state - <https://www.rfc-editor.org/rfc/rfc6926.html#section-6.2.4>", (u32)}, {154, BulkLeaseQueryQueryStartTime, "BLQ query start time - <https://www.rfc-editor.org/rfc/rfc6926.html#section-6.2.5>", (u32)}, {155, BulkLeaseQueryQueryEndTime, "BLQ query end time- <https://www.rfc-editor.org/rfc/rfc6926.html#section-6.2.6>", (u32)}, {156, BulkLeaseQueryDhcpState, "BLQ DHCP state - <https://www.rfc-editor.org/rfc/rfc6926.html#section-6.2.7>", (bulk_query::QueryState)}, {157, BulkLeaseQueryDataSource, "BLQ data source - <https://www.rfc-editor.org/rfc/rfc6926.html#section-6.2.8>", (bulk_query::DataSourceFlags)}, {255, End, "end-of-list marker"} ); /// Holds DHCP options. Options are stored internally in a BTreeMap, so will be stored in /// increasing order according to the opcode. Encoding options to bytes (with [`to_vec()`]) are guaranteed to /// be in the same order for the same set of options /// /// `DhcpOptions` supports long-opt encoding. If the option you insert has data bigger than 255, /// it will be split into multiple max 255 len opts with the same opt code according to RFC3396. /// /// Dhcp options end with the `End` (op code 255) option. This option is automatically inserted /// by the library when encoded to bytes so you don't need to worry about it. /// /// ex /// ```rust /// use dhcproto::v4; /// /// let mut msg = v4::Message::default(); /// msg.opts_mut() /// .insert(v4::DhcpOption::MessageType(v4::MessageType::Discover)); /// msg.opts_mut().insert(v4::DhcpOption::ClientIdentifier( /// vec![0, 1, 2, 3, 4, 5], /// )); /// msg.opts_mut() /// .insert(v4::DhcpOption::ParameterRequestList(vec![ /// v4::OptionCode::SubnetMask, /// v4::OptionCode::Router, /// v4::OptionCode::DomainNameServer, /// v4::OptionCode::DomainName, /// ])); /// ``` /// [`to_vec()`]: crate::encoder::Encoder #[cfg_attr(feature = "serde", derive(Serialize, Deserialize))] #[derive(Debug, Default, Clone, PartialEq, Eq)] pub struct DhcpOptions(BTreeMap<OptionCode, DhcpOption>); impl DhcpOptions { /// Create new [`DhcpOptions`] /// /// [`DhcpOptions`]: crate::v4::DhcpOptions pub fn new() -> Self { Self::default() } /// Get the data for a particular [`OptionCode`] /// /// [`OptionCode`]: crate::v4::OptionCode pub fn get(&self, code: OptionCode) -> Option<&DhcpOption> { self.0.get(&code) } /// Get the mutable data for a particular [`OptionCode`] /// /// [`OptionCode`]: crate::v4::OptionCode pub fn get_mut(&mut self, code: OptionCode) -> Option<&mut DhcpOption> { self.0.get_mut(&code) } /// remove option pub fn remove(&mut self, code: OptionCode) -> Option<DhcpOption> { self.0.remove(&code) } /// insert a new [`DhcpOption`] /// /// ``` /// # use dhcproto::v4::{MessageType, DhcpOption, DhcpOptions}; /// let mut opts = DhcpOptions::new(); /// opts.insert(DhcpOption::MessageType(MessageType::Discover)); /// ``` /// [`DhcpOption`]: crate::v4::DhcpOption pub fn insert(&mut self, opt: DhcpOption) -> Option<DhcpOption> { if opt == DhcpOption::End || opt == DhcpOption::Pad { return None; } self.0.insert((&opt).into(), opt) } /// iterate over entries /// ``` /// # use dhcproto::v4::{MessageType, DhcpOption, DhcpOptions}; /// let mut opts = DhcpOptions::new(); /// opts.insert(DhcpOption::MessageType(MessageType::Offer)); /// opts.insert(DhcpOption::SubnetMask([198, 168, 0, 1].into())); /// for (code, opt) in opts.iter() { /// println!("{code:?} {opt:?}"); /// } /// ``` pub fn iter(&self) -> impl Iterator<Item = (&OptionCode, &DhcpOption)> { self.0.iter() } /// iterate mutably over entries pub fn iter_mut(&mut self) -> impl Iterator<Item = (&OptionCode, &mut DhcpOption)> { self.0.iter_mut() } /// return message type /// ``` /// # use dhcproto::v4::{MessageType, DhcpOption, DhcpOptions}; /// let mut opts = DhcpOptions::new(); /// opts.insert(DhcpOption::MessageType(MessageType::Offer)); /// assert_eq!(opts.msg_type(), Some(MessageType::Offer)); /// ``` pub fn msg_type(&self) -> Option<MessageType> { let opt = self.get(OptionCode::MessageType)?; match opt { DhcpOption::MessageType(mtype) => Some(*mtype), _ => unreachable!("cannot return different option for MessageType"), } } /// determine if options contains a specific message type /// ``` /// # use dhcproto::v4::{MessageType, DhcpOption, DhcpOptions}; /// let mut opts = DhcpOptions::new(); /// opts.insert(DhcpOption::MessageType(MessageType::Offer)); /// assert!(opts.has_msg_type(MessageType::Offer)); /// assert!(!opts.has_msg_type(MessageType::Decline)); /// ``` pub fn has_msg_type(&self, opt: MessageType) -> bool { matches!(self.get(OptionCode::MessageType), Some(DhcpOption::MessageType(msg)) if *msg == opt) } /// clear all options /// ``` /// # use dhcproto::v4::{MessageType, DhcpOption, DhcpOptions}; /// let mut opts = DhcpOptions::new(); /// opts.insert(DhcpOption::MessageType(MessageType::Discover)); /// assert!(opts.len() == 1); /// opts.clear(); // clear options /// assert!(opts.is_empty()); /// ``` pub fn clear(&mut self) { self.0.clear() } /// Returns `true` if there are no options /// ``` /// # use dhcproto::v4::{MessageType, DhcpOption, DhcpOptions}; /// let mut opts = DhcpOptions::new(); /// opts.insert(DhcpOption::MessageType(MessageType::Offer)); /// assert!(!opts.is_empty()); /// ``` pub fn is_empty(&self) -> bool { self.0.is_empty() } /// Retains only the elements specified by the predicate pub fn retain<F>(&mut self, pred: F) where F: FnMut(&OptionCode, &mut DhcpOption) -> bool, { self.0.retain(pred) } /// Returns number of Options /// ``` /// # use dhcproto::v4::{MessageType, DhcpOption, DhcpOptions}; /// let mut opts = DhcpOptions::new(); /// opts.insert(DhcpOption::MessageType(MessageType::Offer)); /// assert_eq!(opts.len(), 1); /// ``` pub fn len(&self) -> usize { self.0.len() } /// Returns true if the option code exists in the map /// /// ``` /// # use dhcproto::v4::{MessageType, DhcpOption, DhcpOptions, OptionCode}; /// let mut opts = DhcpOptions::new(); /// opts.insert(DhcpOption::MessageType(MessageType::Discover)); /// /// assert!(opts.contains(OptionCode::MessageType)); /// assert!(!opts.contains(OptionCode::SubnetMask)); /// ``` pub fn contains(&self, code: OptionCode) -> bool { self.0.contains_key(&code) } } impl IntoIterator for DhcpOptions { type Item = (OptionCode, DhcpOption); type IntoIter = alloc::collections::btree_map::IntoIter<OptionCode, DhcpOption>; fn into_iter(self) -> Self::IntoIter { self.0.into_iter() } } impl<'a> IntoIterator for &'a DhcpOptions { type Item = (&'a OptionCode, &'a DhcpOption); type IntoIter = alloc::collections::btree_map::Iter<'a, OptionCode, DhcpOption>; fn into_iter(self) -> Self::IntoIter { self.0.iter() } } impl FromIterator<DhcpOption> for DhcpOptions { fn from_iter<T: IntoIterator<Item = DhcpOption>>(iter: T) -> Self { DhcpOptions( iter.into_iter() .map(|opt| ((&opt).into(), opt)) .collect::<BTreeMap<OptionCode, DhcpOption>>(), ) } } impl FromIterator<(OptionCode, DhcpOption)> for DhcpOptions { fn from_iter<T: IntoIterator<Item = (OptionCode, DhcpOption)>>(iter: T) -> Self { DhcpOptions(iter.into_iter().collect::<BTreeMap<_, _>>()) } } impl Decodable for DhcpOptions { fn decode(decoder: &mut Decoder<'_>) -> DecodeResult<Self> { // represented as a vector in the actual message let mut opts = BTreeMap::new(); // should we error the whole parser if we fail to parse an // option or just stop parsing options? -- here we will just stop while let Ok(opt) = DhcpOption::decode(decoder) { // we throw away PAD bytes here match opt { DhcpOption::End => { break; } DhcpOption::Pad => {} _ => { opts.insert(OptionCode::from(&opt), opt); } } } Ok(DhcpOptions(opts)) } } impl Encodable for DhcpOptions { fn encode(&self, e: &mut Encoder<'_>) -> EncodeResult<()> { if self.0.is_empty() { return Ok(()); } // encode all opts adding the `End` afterwards self.0 .iter() .filter(|(code, _)| **code != OptionCode::RelayAgentInformation) .chain( self.get(OptionCode::RelayAgentInformation) .map(|opt| (&OptionCode::RelayAgentInformation, opt)), ) .chain(iter::once((&OptionCode::End, &DhcpOption::End))) .try_for_each(|(_, opt)| opt.encode(e)) } } impl PartialOrd for OptionCode { fn partial_cmp(&self, other: &Self) -> Option<core::cmp::Ordering> { Some(self.cmp(other)) } } impl Ord for OptionCode { fn cmp(&self, other: &Self) -> core::cmp::Ordering { u8::from(*self).cmp(&u8::from(*other)) } } impl PartialOrd for DhcpOption { fn partial_cmp(&self, other: &Self) -> Option<core::cmp::Ordering> { Some(self.cmp(other)) } } impl Ord for DhcpOption { fn cmp(&self, other: &Self) -> core::cmp::Ordering { OptionCode::from(self).cmp(&OptionCode::from(other)) } } /// Architecture name from - <https://www.rfc-editor.org/rfc/rfc4578.html> #[cfg_attr(feature = "serde", derive(Serialize, Deserialize))] #[derive(Debug, Clone, Copy, PartialEq, Eq, Hash)] pub enum Architecture { /// Intel x86PC Intelx86PC, /// NEC/PC98 NECPC98, /// EFI Itanium Itanium, /// DEC Alpha DECAlpha, /// Arc x86 Arcx86, /// Intel Lean Client IntelLeanClient, /// EFI IA32 IA32, /// EFI BC BC, /// EFI Xscale Xscale, /// EFI x86-64 X86_64, /// Unknown Unknown(u16), } impl From<u16> for Architecture { fn from(n: u16) -> Self { use Architecture::*; match n { 0 => Intelx86PC, 1 => NECPC98, 2 => Itanium, 3 => DECAlpha, 4 => Arcx86, 5 => IntelLeanClient, 6 => IA32, 7 => BC, 8 => Xscale, 9 => X86_64, _ => Unknown(n), } } } impl From<Architecture> for u16 { fn from(n: Architecture) -> Self { use Architecture as A; match n { A::Intelx86PC => 0, A::NECPC98 => 1, A::Itanium => 2, A::DECAlpha => 3, A::Arcx86 => 4, A::IntelLeanClient => 5, A::IA32 => 6, A::BC => 7, A::Xscale => 8, A::X86_64 => 9, A::Unknown(n) => n, } } } /// NetBIOS allows several different node types #[cfg_attr(feature = "serde", derive(Serialize, Deserialize))] #[derive(Debug, Clone, Copy, PartialEq, Eq, Hash)] pub enum NodeType { /// Broadcast B, /// Peer-to-peer P, /// Mixed (B & P) M, /// Hybrid (P & B) H, /// Unknown Unknown(u8), } impl From<u8> for NodeType { fn from(n: u8) -> Self { use NodeType::*; match n { 1 => B, 2 => P, 4 => M, 8 => H, _ => Unknown(n), } } } impl From<NodeType> for u8 { fn from(n: NodeType) -> Self { use NodeType as N; match n { N::B => 1, N::P => 2, N::M => 4, N::H => 8, N::Unknown(n) => n, } } } /// AutoConfigure option values #[cfg_attr(feature = "serde", derive(Serialize, Deserialize))] #[derive(Debug, Clone, Copy, PartialEq, Eq, Hash)] #[repr(u8)] pub enum AutoConfig { /// Do not autoconfig DoNotAutoConfigure = 0, /// autoconfig AutoConfigure = 1, } impl TryFrom<u8> for AutoConfig { type Error = crate::error::DecodeError; fn try_from(value: u8) -> Result<Self, Self::Error> { match value { 0 => Ok(AutoConfig::DoNotAutoConfigure), 1 => Ok(AutoConfig::AutoConfigure), n => Err(super::DecodeError::InvalidData( n as u32, "invalid number in disable SLAAC autoconfig", )), } } } #[inline] pub(crate) fn decode_inner( code: OptionCode, len: usize, decoder: &mut Decoder<'_>, ) -> DecodeResult<DhcpOption> { use DhcpOption::*; Ok(match code { OptionCode::Pad => Pad, OptionCode::SubnetMask => SubnetMask(decoder.read_ipv4(len)?), OptionCode::TimeOffset => TimeOffset(decoder.read_i32()?), OptionCode::Router => Router(decoder.read_ipv4s(len)?), OptionCode::TimeServer => TimeServer(decoder.read_ipv4s(len)?), OptionCode::NameServer => NameServer(decoder.read_ipv4s(len)?), OptionCode::DomainNameServer => DomainNameServer(decoder.read_ipv4s(len)?), OptionCode::LogServer => LogServer(decoder.read_ipv4s(len)?), OptionCode::QuoteServer => QuoteServer(decoder.read_ipv4s(len)?), OptionCode::LprServer => LprServer(decoder.read_ipv4s(len)?), OptionCode::ImpressServer => ImpressServer(decoder.read_ipv4s(len)?), OptionCode::ResourceLocationServer => ResourceLocationServer(decoder.read_ipv4s(len)?), OptionCode::Hostname => Hostname(decoder.read_string(len)?), OptionCode::BootFileSize => BootFileSize(decoder.read_u16()?), OptionCode::MeritDumpFile => MeritDumpFile(decoder.read_string(len)?), OptionCode::DomainName => DomainName(decoder.read_string(len)?), OptionCode::SwapServer => SwapServer(decoder.read_ipv4(len)?), OptionCode::RootPath => RootPath(decoder.read_string(len)?), OptionCode::ExtensionsPath => ExtensionsPath(decoder.read_string(len)?), OptionCode::IpForwarding => IpForwarding(decoder.read_bool()?), OptionCode::NonLocalSrcRouting => NonLocalSrcRouting(decoder.read_bool()?), OptionCode::PolicyFilter => PolicyFilter(decoder.read_pair_ipv4s(len)?), OptionCode::MaxDatagramSize => MaxDatagramSize(decoder.read_u16()?), OptionCode::DefaultIpTtl => DefaultIpTtl(decoder.read_u8()?), OptionCode::PathMtuAgingTimeout => PathMtuAgingTimeout(decoder.read_u32()?), OptionCode::PathMtuPlateauTable => PathMtuPlateauTable({ decoder .read_slice(len)? .chunks_exact(2) .map(|num| u16::from_be_bytes([num[0], num[1]])) .collect() }), OptionCode::InterfaceMtu => InterfaceMtu(decoder.read_u16()?), OptionCode::AllSubnetsLocal => AllSubnetsLocal(decoder.read_bool()?), OptionCode::BroadcastAddr => BroadcastAddr(decoder.read_ipv4(len)?), OptionCode::PerformMaskDiscovery => PerformMaskDiscovery(decoder.read_bool()?), OptionCode::MaskSupplier => MaskSupplier(decoder.read_bool()?), OptionCode::PerformRouterDiscovery => PerformRouterDiscovery(decoder.read_bool()?), OptionCode::RouterSolicitationAddr => RouterSolicitationAddr(decoder.read_ipv4(len)?), OptionCode::StaticRoutingTable => StaticRoutingTable(decoder.read_pair_ipv4s(len)?), OptionCode::TrailerEncapsulated => TrailerEncapsulated(decoder.read_bool()?), OptionCode::ArpCacheTimeout => ArpCacheTimeout(decoder.read_u32()?), OptionCode::EthernetEncapsulation => EthernetEncapsulation(decoder.read_bool()?), OptionCode::DefaultTcpTtl => DefaultIpTtl(decoder.read_u8()?), OptionCode::TcpKeepaliveInterval => TcpKeepaliveInterval(decoder.read_u32()?), OptionCode::TcpKeepaliveGarbage => TcpKeepaliveGarbage(decoder.read_bool()?), OptionCode::NisDomain => NisDomain(decoder.read_string(len)?), OptionCode::NisServers => NisServers(decoder.read_ipv4s(len)?), OptionCode::NtpServers => NtpServers(decoder.read_ipv4s(len)?), OptionCode::VendorExtensions => VendorExtensions(decoder.read_slice(len)?.to_vec()), OptionCode::NetBiosNameServers => NetBiosNameServers(decoder.read_ipv4s(len)?), OptionCode::NetBiosDatagramDistributionServer => { NetBiosDatagramDistributionServer(decoder.read_ipv4s(len)?) } OptionCode::NetBiosNodeType => NetBiosNodeType(decoder.read_u8()?.into()), OptionCode::NetBiosScope => NetBiosScope(decoder.read_string(len)?), OptionCode::XFontServer => XFontServer(decoder.read_ipv4s(len)?), OptionCode::XDisplayManager => XDisplayManager(decoder.read_ipv4s(len)?), OptionCode::RequestedIpAddress => RequestedIpAddress(decoder.read_ipv4(len)?), OptionCode::AddressLeaseTime => AddressLeaseTime(decoder.read_u32()?), OptionCode::OptionOverload => OptionOverload(decoder.read_u8()?), OptionCode::MessageType => MessageType(decoder.read_u8()?.into()), OptionCode::ServerIdentifier => ServerIdentifier(decoder.read_ipv4(len)?), OptionCode::ParameterRequestList => ParameterRequestList( decoder .read_slice(len)? .iter() .map(|code| (*code).into()) .collect(), ), OptionCode::Message => Message(decoder.read_string(len)?), OptionCode::MaxMessageSize => MaxMessageSize(decoder.read_u16()?), OptionCode::Renewal => Renewal(decoder.read_u32()?), OptionCode::Rebinding => Rebinding(decoder.read_u32()?), OptionCode::ClassIdentifier => ClassIdentifier(decoder.read_slice(len)?.to_vec()), OptionCode::ClientIdentifier => ClientIdentifier(decoder.read_slice(len)?.to_vec()), OptionCode::NwipDomainName => NwipDomainName(decoder.read_string(len)?), OptionCode::NwipInformation => NwipInformation(decoder.read_slice(len)?.to_vec()), OptionCode::NispServiceDomain => NispServiceDomain(decoder.read_string(len)?), OptionCode::NispServers => NispServers(decoder.read_ipv4s(len)?), OptionCode::TFTPServerName => TFTPServerName(decoder.read_slice(len)?.to_vec()), OptionCode::BootfileName => BootfileName(decoder.read_slice(len)?.to_vec()), OptionCode::MobileIpHomeAgent => MobileIpHomeAgent(decoder.read_ipv4s(len)?), OptionCode::SmtpServer => SmtpServer(decoder.read_ipv4s(len)?), OptionCode::Pop3Server => Pop3Server(decoder.read_ipv4s(len)?), OptionCode::NntpServer => NntpServer(decoder.read_ipv4s(len)?), OptionCode::WwwServer => WwwServer(decoder.read_ipv4s(len)?), OptionCode::DefaultFingerServer => DefaultFingerServer(decoder.read_ipv4s(len)?), OptionCode::IrcServer => IrcServer(decoder.read_ipv4s(len)?), OptionCode::StreetTalkServer => StreetTalkServer(decoder.read_ipv4s(len)?), OptionCode::StreetTalkDirectoryAssistance => { StreetTalkDirectoryAssistance(decoder.read_ipv4s(len)?) } OptionCode::UserClass => UserClass(decoder.read_slice(len)?.to_vec()), OptionCode::RapidCommit => { debug_assert!(len == 0); RapidCommit } OptionCode::RelayAgentInformation => { let mut dec = Decoder::new(decoder.read_slice(len)?); RelayAgentInformation(relay::RelayAgentInformation::decode(&mut dec)?) } OptionCode::BcmsControllerNames => BcmsControllerNames(decoder.read_domains(len)?), OptionCode::BcmsControllerAddrs => BcmsControllerAddrs(decoder.read_ipv4s(len)?), OptionCode::ClientLastTransactionTime => ClientLastTransactionTime(decoder.read_u32()?), OptionCode::AssociatedIp => AssociatedIp(decoder.read_ipv4s(len)?), OptionCode::ClientSystemArchitecture => { let ty = decoder.read_u16()?; ClientSystemArchitecture(ty.into()) } OptionCode::ClientNetworkInterface => { debug_assert!(len == 3); ClientNetworkInterface(decoder.read_u8()?, decoder.read_u8()?, decoder.read_u8()?) } OptionCode::ClientMachineIdentifier => { ClientMachineIdentifier(decoder.read_slice(len)?.to_vec()) } OptionCode::Ipv6OnlyPreferred => Ipv6OnlyPreferred(decoder.read_u32()?), OptionCode::CaptivePortal => CaptivePortal(decoder.read_str(len)?.to_string()), OptionCode::DisableSLAAC => DisableSLAAC(decoder.read_u8()?.try_into()?), OptionCode::SubnetSelection => SubnetSelection(decoder.read_ipv4(len)?), OptionCode::DomainSearch => DomainSearch(decoder.read_domains(len)?), OptionCode::TFTPServerAddress => TFTPServerAddress(decoder.read_ipv4(len)?), OptionCode::BulkLeaseQueryStatusCode => { let code = decoder.read_u8()?.into(); // len - 1 because code is included in length let message = decoder.read_string(len - 1)?; BulkLeaseQueryStatusCode(code, message) } OptionCode::BulkLeaseQueryBaseTime => { debug_assert!(len == 4); BulkLeaseQueryBaseTime(decoder.read_u32()?) } OptionCode::BulkLeasQueryStartTimeOfState => { debug_assert!(len == 4); BulkLeasQueryStartTimeOfState(decoder.read_u32()?) } OptionCode::BulkLeaseQueryQueryStartTime => { debug_assert!(len == 4); BulkLeaseQueryQueryStartTime(decoder.read_u32()?) } OptionCode::BulkLeaseQueryQueryEndTime => { debug_assert!(len == 4); BulkLeaseQueryQueryEndTime(decoder.read_u32()?) } OptionCode::BulkLeaseQueryDhcpState => BulkLeaseQueryDhcpState(decoder.read_u8()?.into()), OptionCode::BulkLeaseQueryDataSource => { BulkLeaseQueryDataSource(bulk_query::DataSourceFlags::new(decoder.read_u8()?)) } OptionCode::ClientFQDN => { debug_assert!(len >= 3); let flags = decoder.read_u8()?.into(); let rcode1 = decoder.read_u8()?; let rcode2 = decoder.read_u8()?; let mut name_decoder = BinDecoder::new(decoder.read_slice(len - 3)?); let name = Name::read(&mut name_decoder)?; ClientFQDN(fqdn::ClientFQDN { flags, r1: rcode1, r2: rcode2, domain: name, }) } OptionCode::ClasslessStaticRoute => { let mut routes = Vec::new(); let mut route_dec = Decoder::new(decoder.read_slice(len)?); while let Ok(prefix_len) = route_dec.read_u8() { if prefix_len > 32 { break; } // Significant bytes to hold the prefix let sig_bytes = (prefix_len as usize).div_ceil(8); let mut dest = [0u8; 4]; dest[0..sig_bytes].clone_from_slice(route_dec.read_slice(sig_bytes)?); let dest = Ipv4Net::new(dest.into(), prefix_len).unwrap(); let gw = route_dec.read_ipv4(4)?; routes.push((dest, gw)); } ClasslessStaticRoute(routes) } OptionCode::End => End, // not yet implemented OptionCode::Unknown(code) => { let data = decoder.read_slice(len)?.to_vec(); Unknown(UnknownOption { code, data }) } }) } impl Decodable for DhcpOption { #[inline]
rust
MIT
b4ea30defc01e7ae66e7075d6a0533d9bb9503dc
2026-01-04T20:19:33.979507Z
true