repo
stringlengths
6
65
file_url
stringlengths
81
311
file_path
stringlengths
6
227
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 15:31:58
2026-01-04 20:25:31
truncated
bool
2 classes
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/macros_pin.rs
tokio/tests/macros_pin.rs
#![cfg(feature = "macros")] #[cfg(all(target_family = "wasm", not(target_os = "wasi")))] use wasm_bindgen_test::wasm_bindgen_test as maybe_tokio_test; #[cfg(not(all(target_family = "wasm", not(target_os = "wasi"))))] use tokio::test as maybe_tokio_test; async fn one() {} async fn two() {} #[maybe_tokio_test] async fn multi_pin() { tokio::pin! { let f1 = one(); let f2 = two(); } (&mut f1).await; (&mut f2).await; }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_lines.rs
tokio/tests/io_lines.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use tokio::io::AsyncBufReadExt; use tokio_test::assert_ok; #[tokio::test] async fn lines_inherent() { let rd: &[u8] = b"hello\r\nworld\n\n"; let mut st = rd.lines(); let b = assert_ok!(st.next_line().await).unwrap(); assert_eq!(b, "hello"); let b = assert_ok!(st.next_line().await).unwrap(); assert_eq!(b, "world"); let b = assert_ok!(st.next_line().await).unwrap(); assert_eq!(b, ""); assert!(assert_ok!(st.next_line().await).is_none()); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/fs_file.rs
tokio/tests/fs_file.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // WASI does not support all fs operations use futures::future::FutureExt; use std::io::prelude::*; use std::io::IoSlice; use tempfile::NamedTempFile; use tokio::fs::File; use tokio::io::{AsyncReadExt, AsyncSeekExt, AsyncWriteExt, SeekFrom}; use tokio_test::task; const HELLO: &[u8] = b"hello world..."; #[tokio::test] async fn basic_read() { let mut tempfile = tempfile(); tempfile.write_all(HELLO).unwrap(); let mut file = File::open(tempfile.path()).await.unwrap(); let mut buf = [0; 1024]; let n = file.read(&mut buf).await.unwrap(); assert_eq!(n, HELLO.len()); assert_eq!(&buf[..n], HELLO); } #[tokio::test] async fn basic_write() { let tempfile = tempfile(); let mut file = File::create(tempfile.path()).await.unwrap(); file.write_all(HELLO).await.unwrap(); file.flush().await.unwrap(); let file = std::fs::read(tempfile.path()).unwrap(); assert_eq!(file, HELLO); } #[tokio::test] async fn basic_write_and_shutdown() { let tempfile = tempfile(); let mut file = File::create(tempfile.path()).await.unwrap(); file.write_all(HELLO).await.unwrap(); file.shutdown().await.unwrap(); let file = std::fs::read(tempfile.path()).unwrap(); assert_eq!(file, HELLO); } #[tokio::test] async fn write_vectored() { let tempfile = tempfile(); let mut file = File::create(tempfile.path()).await.unwrap(); let ret = file .write_vectored(&[IoSlice::new(HELLO), IoSlice::new(HELLO)]) .await .unwrap(); assert_eq!(ret, HELLO.len() * 2); file.flush().await.unwrap(); let file = std::fs::read(tempfile.path()).unwrap(); assert_eq!(file, [HELLO, HELLO].concat()); } #[tokio::test] async fn write_vectored_and_shutdown() { let tempfile = tempfile(); let mut file = File::create(tempfile.path()).await.unwrap(); let ret = file .write_vectored(&[IoSlice::new(HELLO), IoSlice::new(HELLO)]) .await .unwrap(); assert_eq!(ret, HELLO.len() * 2); file.shutdown().await.unwrap(); let file = std::fs::read(tempfile.path()).unwrap(); assert_eq!(file, [HELLO, HELLO].concat()); } #[tokio::test] async fn rewind_seek_position() { let tempfile = tempfile(); let mut file = File::create(tempfile.path()).await.unwrap(); file.seek(SeekFrom::Current(10)).await.unwrap(); file.rewind().await.unwrap(); assert_eq!(file.stream_position().await.unwrap(), 0); } #[tokio::test] async fn coop() { let mut tempfile = tempfile(); tempfile.write_all(HELLO).unwrap(); let mut task = task::spawn(async { let mut file = File::open(tempfile.path()).await.unwrap(); let mut buf = [0; 1024]; loop { let _ = file.read(&mut buf).await.unwrap(); file.seek(std::io::SeekFrom::Start(0)).await.unwrap(); } }); for _ in 0..1_000 { if task.poll().is_pending() { return; } } panic!("did not yield"); } #[tokio::test] async fn write_to_clone() { let tempfile = tempfile(); let file = File::create(tempfile.path()).await.unwrap(); let mut clone = file.try_clone().await.unwrap(); clone.write_all(HELLO).await.unwrap(); clone.flush().await.unwrap(); let contents = std::fs::read(tempfile.path()).unwrap(); assert_eq!(contents, HELLO); } #[tokio::test] async fn write_into_std() { let tempfile = tempfile(); let file = File::create(tempfile.path()).await.unwrap(); let mut std_file = file.into_std().await; std_file.write_all(HELLO).unwrap(); let contents = std::fs::read(tempfile.path()).unwrap(); assert_eq!(contents, HELLO); } #[tokio::test] async fn write_into_std_immediate() { let tempfile = tempfile(); let file = File::create(tempfile.path()).await.unwrap(); let mut std_file = file.try_into_std().unwrap(); std_file.write_all(HELLO).unwrap(); let contents = std::fs::read(tempfile.path()).unwrap(); assert_eq!(contents, HELLO); } #[tokio::test] async fn read_file_from_std() { let mut tempfile = tempfile(); tempfile.write_all(HELLO).unwrap(); let std_file = std::fs::File::open(tempfile.path()).unwrap(); let mut file = File::from(std_file); let mut buf = [0; 1024]; let n = file.read(&mut buf).await.unwrap(); assert_eq!(n, HELLO.len()); assert_eq!(&buf[..n], HELLO); } #[tokio::test] async fn empty_read() { let mut tempfile = tempfile(); tempfile.write_all(HELLO).unwrap(); let mut file = File::open(tempfile.path()).await.unwrap(); // Perform an empty read and get a length of zero. assert!(matches!(file.read(&mut []).now_or_never(), Some(Ok(0)))); // Check that we don't get EOF on the next read. let mut buf = [0; 1024]; let n = file.read(&mut buf).await.unwrap(); assert_eq!(n, HELLO.len()); assert_eq!(&buf[..n], HELLO); } fn tempfile() -> NamedTempFile { NamedTempFile::new().unwrap() } #[tokio::test] async fn set_max_buf_size_read() { let mut tempfile = tempfile(); tempfile.write_all(HELLO).unwrap(); let mut file = File::open(tempfile.path()).await.unwrap(); let mut buf = [0; 1024]; file.set_max_buf_size(1); // A single read operation reads a maximum of 1 byte. assert_eq!(file.read(&mut buf).await.unwrap(), 1); } #[tokio::test] async fn set_max_buf_size_write() { let tempfile = tempfile(); let mut file = File::create(tempfile.path()).await.unwrap(); file.set_max_buf_size(1); // A single write operation writes a maximum of 1 byte. assert_eq!(file.write(HELLO).await.unwrap(), 1); } #[tokio::test] #[cfg_attr(miri, ignore)] #[cfg(unix)] async fn file_debug_fmt() { let tempfile = tempfile(); let file = File::open(tempfile.path()).await.unwrap(); assert_eq!( &format!("{file:?}")[0..33], "tokio::fs::File { std: File { fd:" ); } #[tokio::test] #[cfg(windows)] async fn file_debug_fmt() { let tempfile = tempfile(); let file = File::open(tempfile.path()).await.unwrap(); assert_eq!( &format!("{:?}", file)[0..37], "tokio::fs::File { std: File { handle:" ); } #[tokio::test] #[cfg(unix)] async fn unix_fd_is_valid() { use std::os::unix::io::AsRawFd; let tempfile = tempfile(); let file = File::create(tempfile.path()).await.unwrap(); assert!(file.as_raw_fd() as u64 > 0); } #[tokio::test] #[cfg(unix)] async fn read_file_from_unix_fd() { use std::os::unix::io::{FromRawFd, IntoRawFd}; let mut tempfile = tempfile(); tempfile.write_all(HELLO).unwrap(); let file1 = File::open(tempfile.path()).await.unwrap(); let raw_fd = file1.into_std().await.into_raw_fd(); assert!(raw_fd > 0); let mut file2 = unsafe { File::from_raw_fd(raw_fd) }; let mut buf = [0; 1024]; let n = file2.read(&mut buf).await.unwrap(); assert_eq!(n, HELLO.len()); assert_eq!(&buf[..n], HELLO); } #[tokio::test] #[cfg(windows)] async fn windows_handle() { use std::os::windows::io::AsRawHandle; let tempfile = tempfile(); let file = File::create(tempfile.path()).await.unwrap(); assert!(file.as_raw_handle() as u64 > 0); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_repeat.rs
tokio/tests/io_repeat.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(miri)))] use tokio::io::AsyncReadExt; #[tokio::test] async fn repeat_poll_read_is_cooperative() { tokio::select! { biased; _ = async { loop { let mut buf = [0u8; 4096]; tokio::io::repeat(0b101).read_exact(&mut buf).await.unwrap(); } } => {}, _ = tokio::task::yield_now() => {} } }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/signal_realtime.rs
tokio/tests/signal_realtime.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] #![cfg(any(target_os = "linux", target_os = "illumos"))] #![cfg(not(miri))] // No `sigaction` in Miri. mod support { pub mod signal; } use libc::c_int; use support::signal::send_signal; use futures::stream::{FuturesUnordered, StreamExt}; use std::collections::HashMap; use tokio::signal::unix::{signal, SignalKind}; use tokio::time::{sleep, Duration}; use tokio_test::assert_ok; #[tokio::test] async fn signal_realtime() { // Attempt to register a real-time signal for everything between SIGRTMIN // and SIGRTMAX. let signals = (libc::SIGRTMIN()..=sigrt_max()) .map(|signum| { let sig = assert_ok!( signal(SignalKind::from_raw(signum)), "failed to create signal for {}", sigrtnum_to_string(signum), ); (signum, sig) }) .collect::<Vec<_>>(); eprintln!( "registered {} signals in the range {}..={}", signals.len(), libc::SIGRTMIN(), libc::SIGRTMAX() ); // Now send signals to each of the registered signals. for signum in libc::SIGRTMIN()..=sigrt_max() { send_signal(signum); } let futures = signals .into_iter() .map(|(signum, mut sig)| async move { let res = sig.recv().await; (signum, res) }) .collect::<FuturesUnordered<_>>(); // Ensure that all signals are received in time -- attempt to get whatever // we can. let sleep = std::pin::pin!(sleep(Duration::from_secs(5))); let done = futures.take_until(sleep).collect::<HashMap<_, _>>().await; let mut none = Vec::new(); let mut missing = Vec::new(); for signum in libc::SIGRTMIN()..=sigrt_max() { match done.get(&signum) { Some(Some(())) => {} Some(None) => none.push(signum), None => missing.push(signum), } } if none.is_empty() && missing.is_empty() { return; } let mut msg = String::new(); if !none.is_empty() { msg.push_str("no signals received for:\n"); for signum in none { msg.push_str(&format!("- {}\n", sigrtnum_to_string(signum))); } } if !missing.is_empty() { msg.push_str("missing signals for:\n"); for signum in missing { msg.push_str(&format!("- {}\n", sigrtnum_to_string(signum))); } } panic!("{}", msg); } fn sigrt_max() -> c_int { // Generally, you would expect this to be SIGRTMAX. But QEMU only supports // 28 real-time signals even though it might report SIGRTMAX to be higher. // See https://wiki.qemu.org/ChangeLog/9.2#signals. // // The goal of this test is to test that real-time signals are supported in // general, not necessarily that every single signal is supported (which, as // this example suggests, depends on the execution environment). So cap this // at SIGRTMIN+27 (i.e. SIGRTMIN..=SIGRTMIN+27, so 28 signals inclusive). libc::SIGRTMAX().min(libc::SIGRTMIN() + 27) } fn sigrtnum_to_string(signum: i32) -> String { format!("SIGRTMIN+{} (signal {})", signum - libc::SIGRTMIN(), signum) }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/sync_notify_owned.rs
tokio/tests/sync_notify_owned.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "sync")] #[cfg(all(target_family = "wasm", not(target_os = "wasi")))] use wasm_bindgen_test::wasm_bindgen_test as test; use std::sync::Arc; use tokio::sync::Notify; use tokio_test::task::spawn; use tokio_test::*; #[allow(unused)] trait AssertSend: Send + Sync {} impl AssertSend for Notify {} #[test] fn notify_notified_one() { let notify = Arc::new(Notify::new()); let mut notified = spawn(async { notify.clone().notified_owned().await }); notify.notify_one(); assert_ready!(notified.poll()); } #[test] fn notify_multi_notified_one() { let notify = Arc::new(Notify::new()); let mut notified1 = spawn(async { notify.clone().notified_owned().await }); let mut notified2 = spawn(async { notify.clone().notified_owned().await }); // add two waiters into the queue assert_pending!(notified1.poll()); assert_pending!(notified2.poll()); // should wakeup the first one notify.notify_one(); assert_ready!(notified1.poll()); assert_pending!(notified2.poll()); } #[test] fn notify_multi_notified_last() { let notify = Arc::new(Notify::new()); let mut notified1 = spawn(async { notify.clone().notified_owned().await }); let mut notified2 = spawn(async { notify.clone().notified_owned().await }); // add two waiters into the queue assert_pending!(notified1.poll()); assert_pending!(notified2.poll()); // should wakeup the last one notify.notify_last(); assert_pending!(notified1.poll()); assert_ready!(notified2.poll()); } #[test] fn notified_one_notify() { let notify = Arc::new(Notify::new()); let mut notified = spawn(async { notify.clone().notified_owned().await }); assert_pending!(notified.poll()); notify.notify_one(); assert!(notified.is_woken()); assert_ready!(notified.poll()); } #[test] fn notified_multi_notify() { let notify = Arc::new(Notify::new()); let mut notified1 = spawn(async { notify.clone().notified_owned().await }); let mut notified2 = spawn(async { notify.clone().notified_owned().await }); assert_pending!(notified1.poll()); assert_pending!(notified2.poll()); notify.notify_one(); assert!(notified1.is_woken()); assert!(!notified2.is_woken()); assert_ready!(notified1.poll()); assert_pending!(notified2.poll()); } #[test] fn notify_notified_multi() { let notify = Arc::new(Notify::new()); notify.notify_one(); let mut notified1 = spawn(async { notify.clone().notified_owned().await }); let mut notified2 = spawn(async { notify.clone().notified_owned().await }); assert_ready!(notified1.poll()); assert_pending!(notified2.poll()); notify.notify_one(); assert!(notified2.is_woken()); assert_ready!(notified2.poll()); } #[test] fn notified_drop_notified_notify() { let notify = Arc::new(Notify::new()); let mut notified1 = spawn(async { notify.clone().notified_owned().await }); let mut notified2 = spawn(async { notify.clone().notified_owned().await }); assert_pending!(notified1.poll()); drop(notified1); assert_pending!(notified2.poll()); notify.notify_one(); assert!(notified2.is_woken()); assert_ready!(notified2.poll()); } #[test] fn notified_multi_notify_drop_one() { let notify = Arc::new(Notify::new()); let mut notified1 = spawn(async { notify.clone().notified_owned().await }); let mut notified2 = spawn(async { notify.clone().notified_owned().await }); assert_pending!(notified1.poll()); assert_pending!(notified2.poll()); notify.notify_one(); assert!(notified1.is_woken()); assert!(!notified2.is_woken()); drop(notified1); assert!(notified2.is_woken()); assert_ready!(notified2.poll()); } #[test] fn notified_multi_notify_one_drop() { let notify = Arc::new(Notify::new()); let mut notified1 = spawn(async { notify.clone().notified_owned().await }); let mut notified2 = spawn(async { notify.clone().notified_owned().await }); let mut notified3 = spawn(async { notify.clone().notified_owned().await }); // add waiters by order of poll execution assert_pending!(notified1.poll()); assert_pending!(notified2.poll()); assert_pending!(notified3.poll()); // by default fifo notify.notify_one(); drop(notified1); // next waiter should be the one to be to woken up assert_ready!(notified2.poll()); assert_pending!(notified3.poll()); } #[test] fn notified_multi_notify_last_drop() { let notify = Arc::new(Notify::new()); let mut notified1 = spawn(async { notify.clone().notified_owned().await }); let mut notified2 = spawn(async { notify.clone().notified_owned().await }); let mut notified3 = spawn(async { notify.clone().notified_owned().await }); // add waiters by order of poll execution assert_pending!(notified1.poll()); assert_pending!(notified2.poll()); assert_pending!(notified3.poll()); notify.notify_last(); drop(notified3); // latest waiter added should be the one to woken up assert_ready!(notified2.poll()); assert_pending!(notified1.poll()); } #[test] fn notify_in_drop_after_wake() { use futures::task::ArcWake; use std::future::Future; use std::sync::Arc; let notify = Arc::new(Notify::new()); struct NotifyOnDrop(Arc<Notify>); impl ArcWake for NotifyOnDrop { fn wake_by_ref(_arc_self: &Arc<Self>) {} } impl Drop for NotifyOnDrop { fn drop(&mut self) { self.0.notify_waiters(); } } let mut fut = Box::pin(async { notify.clone().notified_owned().await; }); { let waker = futures::task::waker(Arc::new(NotifyOnDrop(notify.clone()))); let mut cx = std::task::Context::from_waker(&waker); assert!(fut.as_mut().poll(&mut cx).is_pending()); } // Now, notifying **should not** deadlock notify.notify_waiters(); } #[test] fn notify_one_after_dropped_all() { let notify = Arc::new(Notify::new()); let mut notified1 = spawn(async { notify.clone().notified_owned().await }); assert_pending!(notified1.poll()); notify.notify_waiters(); notify.notify_one(); drop(notified1); let mut notified2 = spawn(async { notify.clone().notified_owned().await }); assert_ready!(notified2.poll()); } #[test] fn test_notify_one_not_enabled() { let notify = Arc::new(Notify::new()); let mut future = spawn(notify.clone().notified_owned()); notify.notify_one(); assert_ready!(future.poll()); } #[test] fn test_notify_one_after_enable() { let notify = Arc::new(Notify::new()); let mut future = spawn(notify.clone().notified_owned()); future.enter(|_, fut| assert!(!fut.enable())); notify.notify_one(); assert_ready!(future.poll()); future.enter(|_, fut| assert!(fut.enable())); } #[test] fn test_poll_after_enable() { let notify = Arc::new(Notify::new()); let mut future = spawn(notify.clone().notified_owned()); future.enter(|_, fut| assert!(!fut.enable())); assert_pending!(future.poll()); } #[test] fn test_enable_after_poll() { let notify = Arc::new(Notify::new()); let mut future = spawn(notify.clone().notified_owned()); assert_pending!(future.poll()); future.enter(|_, fut| assert!(!fut.enable())); } #[test] fn test_enable_consumes_permit() { let notify = Arc::new(Notify::new()); // Add a permit. notify.notify_one(); let mut future1 = spawn(notify.clone().notified_owned()); future1.enter(|_, fut| assert!(fut.enable())); let mut future2 = spawn(notify.clone().notified_owned()); future2.enter(|_, fut| assert!(!fut.enable())); } #[test] fn test_waker_update() { use futures::task::noop_waker; use std::future::Future; use std::task::Context; let notify = Arc::new(Notify::new()); let mut future = spawn(notify.clone().notified_owned()); let noop = noop_waker(); future.enter(|_, fut| assert_pending!(fut.poll(&mut Context::from_waker(&noop)))); assert_pending!(future.poll()); notify.notify_one(); assert!(future.is_woken()); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/sync_semaphore_owned.rs
tokio/tests/sync_semaphore_owned.rs
#![cfg(feature = "sync")] #[cfg(all(target_family = "wasm", not(target_os = "wasi")))] use wasm_bindgen_test::wasm_bindgen_test as test; use std::sync::Arc; use tokio::sync::Semaphore; #[test] fn try_acquire() { let sem = Arc::new(Semaphore::new(1)); { let p1 = sem.clone().try_acquire_owned(); assert!(p1.is_ok()); let p2 = sem.clone().try_acquire_owned(); assert!(p2.is_err()); } let p3 = sem.try_acquire_owned(); assert!(p3.is_ok()); } #[test] fn try_acquire_many() { let sem = Arc::new(Semaphore::new(42)); { let p1 = sem.clone().try_acquire_many_owned(42); assert!(p1.is_ok()); let p2 = sem.clone().try_acquire_owned(); assert!(p2.is_err()); } let p3 = sem.clone().try_acquire_many_owned(32); assert!(p3.is_ok()); let p4 = sem.clone().try_acquire_many_owned(10); assert!(p4.is_ok()); assert!(sem.try_acquire_owned().is_err()); } #[tokio::test] #[cfg(feature = "full")] async fn acquire() { let sem = Arc::new(Semaphore::new(1)); let p1 = sem.clone().try_acquire_owned().unwrap(); let sem_clone = sem.clone(); let j = tokio::spawn(async move { let _p2 = sem_clone.acquire_owned().await; }); drop(p1); j.await.unwrap(); } #[tokio::test] #[cfg(feature = "full")] async fn acquire_many() { let semaphore = Arc::new(Semaphore::new(42)); let permit32 = semaphore.clone().try_acquire_many_owned(32).unwrap(); let (sender, receiver) = tokio::sync::oneshot::channel(); let join_handle = tokio::spawn(async move { let _permit10 = semaphore.clone().acquire_many_owned(10).await.unwrap(); sender.send(()).unwrap(); let _permit32 = semaphore.acquire_many_owned(32).await.unwrap(); }); receiver.await.unwrap(); drop(permit32); join_handle.await.unwrap(); } #[tokio::test] #[cfg(feature = "full")] async fn add_permits() { let sem = Arc::new(Semaphore::new(0)); let sem_clone = sem.clone(); let j = tokio::spawn(async move { let _p2 = sem_clone.acquire_owned().await; }); sem.add_permits(1); j.await.unwrap(); } #[test] fn forget() { let sem = Arc::new(Semaphore::new(1)); { let p = sem.clone().try_acquire_owned().unwrap(); assert_eq!(sem.available_permits(), 0); p.forget(); assert_eq!(sem.available_permits(), 0); } assert_eq!(sem.available_permits(), 0); assert!(sem.try_acquire_owned().is_err()); } #[test] fn merge() { let sem = Arc::new(Semaphore::new(3)); { let mut p1 = sem.clone().try_acquire_owned().unwrap(); assert_eq!(sem.available_permits(), 2); let p2 = sem.clone().try_acquire_many_owned(2).unwrap(); assert_eq!(sem.available_permits(), 0); p1.merge(p2); assert_eq!(sem.available_permits(), 0); } assert_eq!(sem.available_permits(), 3); } #[test] #[cfg(not(target_family = "wasm"))] // No stack unwinding on wasm targets #[should_panic] fn merge_unrelated_permits() { let sem1 = Arc::new(Semaphore::new(3)); let sem2 = Arc::new(Semaphore::new(3)); let mut p1 = sem1.try_acquire_owned().unwrap(); let p2 = sem2.try_acquire_owned().unwrap(); p1.merge(p2) } #[test] fn split() { let sem = Arc::new(Semaphore::new(5)); let mut p1 = sem.clone().try_acquire_many_owned(3).unwrap(); assert_eq!(sem.available_permits(), 2); assert_eq!(p1.num_permits(), 3); let mut p2 = p1.split(1).unwrap(); assert_eq!(sem.available_permits(), 2); assert_eq!(p1.num_permits(), 2); assert_eq!(p2.num_permits(), 1); let p3 = p1.split(0).unwrap(); assert_eq!(p3.num_permits(), 0); drop(p1); assert_eq!(sem.available_permits(), 4); let p4 = p2.split(1).unwrap(); assert_eq!(p2.num_permits(), 0); assert_eq!(p4.num_permits(), 1); assert!(p2.split(1).is_none()); drop(p2); assert_eq!(sem.available_permits(), 4); drop(p3); assert_eq!(sem.available_permits(), 4); drop(p4); assert_eq!(sem.available_permits(), 5); } #[tokio::test] #[cfg(feature = "full")] async fn stress_test() { let sem = Arc::new(Semaphore::new(5)); let mut join_handles = Vec::new(); for _ in 0..1000 { let sem_clone = sem.clone(); join_handles.push(tokio::spawn(async move { let _p = sem_clone.acquire_owned().await; })); } for j in join_handles { j.await.unwrap(); } // there should be exactly 5 semaphores available now let _p1 = sem.clone().try_acquire_owned().unwrap(); let _p2 = sem.clone().try_acquire_owned().unwrap(); let _p3 = sem.clone().try_acquire_owned().unwrap(); let _p4 = sem.clone().try_acquire_owned().unwrap(); let _p5 = sem.clone().try_acquire_owned().unwrap(); assert!(sem.try_acquire_owned().is_err()); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_mem_stream.rs
tokio/tests/io_mem_stream.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use tokio::io::{duplex, AsyncReadExt, AsyncWriteExt}; #[tokio::test] async fn ping_pong() { let (mut a, mut b) = duplex(32); let mut buf = [0u8; 4]; a.write_all(b"ping").await.unwrap(); b.read_exact(&mut buf).await.unwrap(); assert_eq!(&buf, b"ping"); b.write_all(b"pong").await.unwrap(); a.read_exact(&mut buf).await.unwrap(); assert_eq!(&buf, b"pong"); } #[tokio::test] async fn across_tasks() { let (mut a, mut b) = duplex(32); let t1 = tokio::spawn(async move { a.write_all(b"ping").await.unwrap(); let mut buf = [0u8; 4]; a.read_exact(&mut buf).await.unwrap(); assert_eq!(&buf, b"pong"); }); let t2 = tokio::spawn(async move { let mut buf = [0u8; 4]; b.read_exact(&mut buf).await.unwrap(); assert_eq!(&buf, b"ping"); b.write_all(b"pong").await.unwrap(); }); t1.await.unwrap(); t2.await.unwrap(); } #[tokio::test] async fn disconnect() { let (mut a, mut b) = duplex(32); let t1 = tokio::spawn(async move { a.write_all(b"ping").await.unwrap(); // and dropped }); let t2 = tokio::spawn(async move { let mut buf = [0u8; 32]; let n = b.read(&mut buf).await.unwrap(); assert_eq!(&buf[..n], b"ping"); let n = b.read(&mut buf).await.unwrap(); assert_eq!(n, 0); }); t1.await.unwrap(); t2.await.unwrap(); } #[tokio::test] async fn disconnect_reader() { let (a, mut b) = duplex(2); let t1 = tokio::spawn(async move { // this will block, as not all data fits into duplex b.write_all(b"ping").await.unwrap_err(); }); let t2 = tokio::spawn(async move { // here we drop the reader side, and we expect the writer in the other // task to exit with an error drop(a); }); t2.await.unwrap(); t1.await.unwrap(); } #[tokio::test] async fn max_write_size() { let (mut a, mut b) = duplex(32); let t1 = tokio::spawn(async move { let n = a.write(&[0u8; 64]).await.unwrap(); assert_eq!(n, 32); let n = a.write(&[0u8; 64]).await.unwrap(); assert_eq!(n, 4); }); let mut buf = [0u8; 4]; b.read_exact(&mut buf).await.unwrap(); t1.await.unwrap(); // drop b only after task t1 finishes writing drop(b); } #[tokio::test] async fn duplex_is_cooperative() { let (mut tx, mut rx) = tokio::io::duplex(1024 * 8); tokio::select! { biased; _ = async { loop { let buf = [3u8; 4096]; tx.write_all(&buf).await.unwrap(); let mut buf = [0u8; 4096]; let _ = rx.read(&mut buf).await.unwrap(); } } => {}, _ = tokio::task::yield_now() => {} } }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/process_issue_7144.rs
tokio/tests/process_issue_7144.rs
#![cfg(feature = "process")] #![warn(rust_2018_idioms)] #![cfg(target_os = "linux")] #![cfg(not(miri))] use tokio::process::Command; use tokio::time::{sleep, Duration}; #[tokio::test] async fn issue_7144() { let mut threads = vec![]; for _ in 0..20 { threads.push(tokio::spawn(test_one())); } for thread in threads { thread.await.unwrap(); } } async fn test_one() { let mut t = Command::new("strace") .args("-o /dev/null -D sleep 5".split(' ')) .spawn() .unwrap(); sleep(Duration::from_millis(100)).await; unsafe { libc::kill(t.id().unwrap() as _, libc::SIGINT) }; t.wait().await.unwrap(); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_read_buf.rs
tokio/tests/io_read_buf.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use tokio::io::{AsyncRead, AsyncReadExt, ReadBuf}; use tokio_test::assert_ok; use std::io; use std::pin::Pin; use std::task::{Context, Poll}; #[tokio::test] async fn read_buf() { struct Rd { cnt: usize, } impl AsyncRead for Rd { fn poll_read( mut self: Pin<&mut Self>, _cx: &mut Context<'_>, buf: &mut ReadBuf<'_>, ) -> Poll<io::Result<()>> { self.cnt += 1; buf.put_slice(b"hello world"); Poll::Ready(Ok(())) } } let mut buf = vec![]; let mut rd = Rd { cnt: 0 }; let n = assert_ok!(rd.read_buf(&mut buf).await); assert_eq!(1, rd.cnt); assert_eq!(n, 11); assert_eq!(buf[..], b"hello world"[..]); } #[tokio::test] #[cfg(feature = "io-util")] async fn issue_5588() { use bytes::BufMut; // steps to zero let mut buf = [0; 8]; let mut read_buf = ReadBuf::new(&mut buf); assert_eq!(read_buf.remaining_mut(), 8); assert_eq!(read_buf.chunk_mut().len(), 8); unsafe { read_buf.advance_mut(1); } assert_eq!(read_buf.remaining_mut(), 7); assert_eq!(read_buf.chunk_mut().len(), 7); unsafe { read_buf.advance_mut(5); } assert_eq!(read_buf.remaining_mut(), 2); assert_eq!(read_buf.chunk_mut().len(), 2); unsafe { read_buf.advance_mut(2); } assert_eq!(read_buf.remaining_mut(), 0); assert_eq!(read_buf.chunk_mut().len(), 0); // directly to zero let mut buf = [0; 8]; let mut read_buf = ReadBuf::new(&mut buf); assert_eq!(read_buf.remaining_mut(), 8); assert_eq!(read_buf.chunk_mut().len(), 8); unsafe { read_buf.advance_mut(8); } assert_eq!(read_buf.remaining_mut(), 0); assert_eq!(read_buf.chunk_mut().len(), 0); // uninit let mut buf = [std::mem::MaybeUninit::new(1); 8]; let mut uninit = ReadBuf::uninit(&mut buf); assert_eq!(uninit.remaining_mut(), 8); assert_eq!(uninit.chunk_mut().len(), 8); let mut buf = [std::mem::MaybeUninit::uninit(); 8]; let mut uninit = ReadBuf::uninit(&mut buf); unsafe { uninit.advance_mut(4); } assert_eq!(uninit.remaining_mut(), 4); assert_eq!(uninit.chunk_mut().len(), 4); uninit.put_u8(1); assert_eq!(uninit.remaining_mut(), 3); assert_eq!(uninit.chunk_mut().len(), 3); uninit.put_slice(&[1, 2, 3]); assert_eq!(uninit.remaining_mut(), 0); assert_eq!(uninit.chunk_mut().len(), 0); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/fs.rs
tokio/tests/fs.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // Wasi does not support file operations use tokio::fs; use tokio_test::assert_ok; #[tokio::test] async fn path_read_write() { let temp = tempdir(); let dir = temp.path(); assert_ok!(fs::write(dir.join("bar"), b"bytes").await); let out = assert_ok!(fs::read(dir.join("bar")).await); assert_eq!(out, b"bytes"); } #[tokio::test] async fn try_clone_should_preserve_max_buf_size() { let buf_size = 128; let temp = tempdir(); let dir = temp.path(); let mut file = fs::File::create(dir.join("try_clone_should_preserve_max_buf_size")) .await .unwrap(); file.set_max_buf_size(buf_size); let cloned = file.try_clone().await.unwrap(); assert_eq!(cloned.max_buf_size(), buf_size); } fn tempdir() -> tempfile::TempDir { tempfile::tempdir().unwrap() }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/process_issue_42.rs
tokio/tests/process_issue_42.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] #![cfg(unix)] #![cfg(not(miri))] use futures::future::join_all; use std::process::Stdio; use tokio::process::Command; use tokio::task; #[tokio::test] async fn issue_42() { // We spawn a many batches of processes which should exit at roughly the // same time (modulo OS scheduling delays), to make sure that consuming // a readiness event for one process doesn't inadvertently starve another. // We then do this many times (in parallel) in an effort to stress test the // implementation to ensure there are no race conditions. // See alexcrichton/tokio-process#42 for background let join_handles = (0..10usize).map(|_| { task::spawn(async { let processes = (0..10usize).map(|i| { let mut child = Command::new("echo") .arg(format!("I am spawned process #{i}")) .stdin(Stdio::null()) .stdout(Stdio::null()) .stderr(Stdio::null()) .kill_on_drop(true) .spawn() .unwrap(); async move { child.wait().await } }); join_all(processes).await; }) }); join_all(join_handles).await; }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/task_local_set.rs
tokio/tests/task_local_set.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use futures::{ future::{pending, ready}, FutureExt, }; use tokio::runtime; use tokio::sync::{mpsc, oneshot}; use tokio::task::{self, LocalSet}; use tokio::time; #[cfg(not(target_os = "wasi"))] use std::cell::Cell; use std::sync::atomic::AtomicBool; #[cfg(not(target_os = "wasi"))] use std::sync::atomic::AtomicUsize; use std::sync::atomic::Ordering; #[cfg(not(target_os = "wasi"))] use std::sync::atomic::Ordering::SeqCst; use std::time::Duration; #[tokio::test(flavor = "current_thread")] async fn local_current_thread_scheduler() { LocalSet::new() .run_until(async { task::spawn_local(async {}).await.unwrap(); }) .await; } #[cfg(not(target_os = "wasi"))] // Wasi doesn't support threads #[tokio::test(flavor = "multi_thread")] async fn local_threadpool() { thread_local! { static ON_RT_THREAD: Cell<bool> = const { Cell::new(false) }; } ON_RT_THREAD.with(|cell| cell.set(true)); LocalSet::new() .run_until(async { assert!(ON_RT_THREAD.with(|cell| cell.get())); task::spawn_local(async { assert!(ON_RT_THREAD.with(|cell| cell.get())); }) .await .unwrap(); }) .await; } #[cfg(not(target_os = "wasi"))] // Wasi doesn't support threads #[tokio::test(flavor = "multi_thread")] async fn localset_future_threadpool() { thread_local! { static ON_LOCAL_THREAD: Cell<bool> = const { Cell::new(false) }; } ON_LOCAL_THREAD.with(|cell| cell.set(true)); let local = LocalSet::new(); local.spawn_local(async move { assert!(ON_LOCAL_THREAD.with(|cell| cell.get())); }); local.await; } #[cfg(not(target_os = "wasi"))] // Wasi doesn't support threads #[tokio::test(flavor = "multi_thread")] async fn localset_future_timers() { static RAN1: AtomicBool = AtomicBool::new(false); static RAN2: AtomicBool = AtomicBool::new(false); let local = LocalSet::new(); local.spawn_local(async move { time::sleep(Duration::from_millis(5)).await; RAN1.store(true, Ordering::SeqCst); }); local.spawn_local(async move { time::sleep(Duration::from_millis(10)).await; RAN2.store(true, Ordering::SeqCst); }); local.await; assert!(RAN1.load(Ordering::SeqCst)); assert!(RAN2.load(Ordering::SeqCst)); } #[tokio::test] async fn localset_future_drives_all_local_futs() { static RAN1: AtomicBool = AtomicBool::new(false); static RAN2: AtomicBool = AtomicBool::new(false); static RAN3: AtomicBool = AtomicBool::new(false); let local = LocalSet::new(); local.spawn_local(async move { task::spawn_local(async { task::yield_now().await; RAN3.store(true, Ordering::SeqCst); }); task::yield_now().await; RAN1.store(true, Ordering::SeqCst); }); local.spawn_local(async move { task::yield_now().await; RAN2.store(true, Ordering::SeqCst); }); local.await; assert!(RAN1.load(Ordering::SeqCst)); assert!(RAN2.load(Ordering::SeqCst)); assert!(RAN3.load(Ordering::SeqCst)); } #[cfg(not(target_os = "wasi"))] // Wasi doesn't support threads #[tokio::test(flavor = "multi_thread")] async fn local_threadpool_timer() { // This test ensures that runtime services like the timer are properly // set for the local task set. thread_local! { static ON_RT_THREAD: Cell<bool> = const { Cell::new(false) }; } ON_RT_THREAD.with(|cell| cell.set(true)); LocalSet::new() .run_until(async { assert!(ON_RT_THREAD.with(|cell| cell.get())); let join = task::spawn_local(async move { assert!(ON_RT_THREAD.with(|cell| cell.get())); time::sleep(Duration::from_millis(10)).await; assert!(ON_RT_THREAD.with(|cell| cell.get())); }); join.await.unwrap(); }) .await; } #[test] fn enter_guard_spawn() { let local = LocalSet::new(); let _guard = local.enter(); // Run the local task set. let join = task::spawn_local(async { true }); let rt = runtime::Builder::new_current_thread() .enable_all() .build() .unwrap(); local.block_on(&rt, async move { assert!(join.await.unwrap()); }); } #[cfg(not(target_os = "wasi"))] mod block_in_place_cases { use super::*; use std::future::Future; use std::pin::Pin; use std::task::{Context, Poll}; struct BlockInPlaceOnDrop; impl Future for BlockInPlaceOnDrop { type Output = (); fn poll(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<Self::Output> { Poll::Pending } } impl Drop for BlockInPlaceOnDrop { fn drop(&mut self) { tokio::task::block_in_place(|| {}); } } async fn complete(jh: tokio::task::JoinHandle<()>) { match jh.await { Ok(()) => {} Err(err) if err.is_panic() => std::panic::resume_unwind(err.into_panic()), Err(err) if err.is_cancelled() => panic!("task cancelled"), Err(err) => panic!("{err:?}"), } } #[test] #[should_panic = "can call blocking only when running on the multi-threaded runtime"] fn local_threadpool_blocking_in_place() { thread_local! { static ON_RT_THREAD: Cell<bool> = const { Cell::new(false) }; } ON_RT_THREAD.with(|cell| cell.set(true)); let rt = runtime::Builder::new_current_thread() .enable_all() .build() .unwrap(); LocalSet::new().block_on(&rt, async { assert!(ON_RT_THREAD.with(|cell| cell.get())); let join = task::spawn_local(async move { assert!(ON_RT_THREAD.with(|cell| cell.get())); task::block_in_place(|| {}); assert!(ON_RT_THREAD.with(|cell| cell.get())); }); complete(join).await; }); } #[tokio::test(flavor = "multi_thread")] #[should_panic = "can call blocking only when running on the multi-threaded runtime"] async fn block_in_place_in_run_until_mt() { let local_set = LocalSet::new(); local_set .run_until(async { tokio::task::block_in_place(|| {}); }) .await; } #[tokio::test(flavor = "multi_thread")] #[should_panic = "can call blocking only when running on the multi-threaded runtime"] async fn block_in_place_in_spawn_local_mt() { let local_set = LocalSet::new(); let jh = local_set.spawn_local(async { tokio::task::block_in_place(|| {}); }); local_set.await; complete(jh).await; } #[tokio::test(flavor = "multi_thread")] #[should_panic = "can call blocking only when running on the multi-threaded runtime"] async fn block_in_place_in_spawn_local_drop_mt() { let local_set = LocalSet::new(); let jh = local_set.spawn_local(BlockInPlaceOnDrop); local_set.run_until(tokio::task::yield_now()).await; drop(local_set); complete(jh).await; } #[tokio::test(flavor = "current_thread")] #[should_panic = "can call blocking only when running on the multi-threaded runtime"] async fn block_in_place_in_run_until_ct() { let local_set = LocalSet::new(); local_set .run_until(async { tokio::task::block_in_place(|| {}); }) .await; } #[tokio::test(flavor = "current_thread")] #[should_panic = "can call blocking only when running on the multi-threaded runtime"] async fn block_in_place_in_spawn_local_ct() { let local_set = LocalSet::new(); let jh = local_set.spawn_local(async { tokio::task::block_in_place(|| {}); }); local_set.await; complete(jh).await; } #[tokio::test(flavor = "current_thread")] #[should_panic = "can call blocking only when running on the multi-threaded runtime"] async fn block_in_place_in_spawn_local_drop_ct() { let local_set = LocalSet::new(); let jh = local_set.spawn_local(BlockInPlaceOnDrop); local_set.run_until(tokio::task::yield_now()).await; drop(local_set); complete(jh).await; } } #[cfg(not(target_os = "wasi"))] // Wasi doesn't support threads #[tokio::test(flavor = "multi_thread")] async fn local_threadpool_blocking_run() { thread_local! { static ON_RT_THREAD: Cell<bool> = const { Cell::new(false) }; } ON_RT_THREAD.with(|cell| cell.set(true)); LocalSet::new() .run_until(async { assert!(ON_RT_THREAD.with(|cell| cell.get())); let join = task::spawn_local(async move { assert!(ON_RT_THREAD.with(|cell| cell.get())); task::spawn_blocking(|| { assert!( !ON_RT_THREAD.with(|cell| cell.get()), "blocking must not run on the local task set's thread" ); }) .await .unwrap(); assert!(ON_RT_THREAD.with(|cell| cell.get())); }); join.await.unwrap(); }) .await; } #[cfg(not(target_os = "wasi"))] // Wasi doesn't support threads #[tokio::test(flavor = "multi_thread")] async fn all_spawns_are_local() { use futures::future; thread_local! { static ON_RT_THREAD: Cell<bool> = const { Cell::new(false) }; } ON_RT_THREAD.with(|cell| cell.set(true)); LocalSet::new() .run_until(async { assert!(ON_RT_THREAD.with(|cell| cell.get())); let handles = (0..128) .map(|_| { task::spawn_local(async { assert!(ON_RT_THREAD.with(|cell| cell.get())); }) }) .collect::<Vec<_>>(); for joined in future::join_all(handles).await { joined.unwrap(); } }) .await; } #[cfg(not(target_os = "wasi"))] // Wasi doesn't support threads #[tokio::test(flavor = "multi_thread")] async fn nested_spawn_is_local() { thread_local! { static ON_RT_THREAD: Cell<bool> = const { Cell::new(false) }; } ON_RT_THREAD.with(|cell| cell.set(true)); LocalSet::new() .run_until(async { assert!(ON_RT_THREAD.with(|cell| cell.get())); task::spawn_local(async { assert!(ON_RT_THREAD.with(|cell| cell.get())); task::spawn_local(async { assert!(ON_RT_THREAD.with(|cell| cell.get())); task::spawn_local(async { assert!(ON_RT_THREAD.with(|cell| cell.get())); task::spawn_local(async { assert!(ON_RT_THREAD.with(|cell| cell.get())); }) .await .unwrap(); }) .await .unwrap(); }) .await .unwrap(); }) .await .unwrap(); }) .await; } #[cfg(not(target_os = "wasi"))] // Wasi doesn't support threads #[test] fn join_local_future_elsewhere() { thread_local! { static ON_RT_THREAD: Cell<bool> = const { Cell::new(false) }; } ON_RT_THREAD.with(|cell| cell.set(true)); let rt = runtime::Runtime::new().unwrap(); let local = LocalSet::new(); local.block_on(&rt, async move { let (tx, rx) = oneshot::channel(); let join = task::spawn_local(async move { assert!( ON_RT_THREAD.with(|cell| cell.get()), "local task must run on local thread, no matter where it is awaited" ); rx.await.unwrap(); "hello world" }); let join2 = task::spawn(async move { assert!( !ON_RT_THREAD.with(|cell| cell.get()), "spawned task should be on a worker" ); tx.send(()).expect("task shouldn't have ended yet"); join.await.expect("task should complete successfully"); }); join2.await.unwrap() }); } // Tests for <https://github.com/tokio-rs/tokio/issues/4973> #[cfg(not(target_os = "wasi"))] // Wasi doesn't support threads #[tokio::test(flavor = "multi_thread")] async fn localset_in_thread_local() { thread_local! { static LOCAL_SET: LocalSet = LocalSet::new(); } // holds runtime thread until end of main fn. let (_tx, rx) = oneshot::channel::<()>(); let handle = tokio::runtime::Handle::current(); std::thread::spawn(move || { LOCAL_SET.with(|local_set| { handle.block_on(local_set.run_until(async move { let _ = rx.await; })) }); }); } #[test] fn drop_cancels_tasks() { use std::rc::Rc; // This test reproduces issue #1842 let rt = rt(); let rc1 = Rc::new(()); let rc2 = rc1.clone(); let (started_tx, started_rx) = oneshot::channel(); let local = LocalSet::new(); local.spawn_local(async move { // Move this in let _rc2 = rc2; started_tx.send(()).unwrap(); futures::future::pending::<()>().await; }); local.block_on(&rt, async { started_rx.await.unwrap(); }); drop(local); drop(rt); assert_eq!(1, Rc::strong_count(&rc1)); } /// Runs a test function in a separate thread, and panics if the test does not /// complete within the specified timeout, or if the test function panics. /// /// This is intended for running tests whose failure mode is a hang or infinite /// loop that cannot be detected otherwise. fn with_timeout(timeout: Duration, f: impl FnOnce() + Send + 'static) { use std::sync::mpsc::RecvTimeoutError; let (done_tx, done_rx) = std::sync::mpsc::channel(); let thread = std::thread::spawn(move || { f(); // Send a message on the channel so that the test thread can // determine if we have entered an infinite loop: done_tx.send(()).unwrap(); }); // Since the failure mode of this test is an infinite loop, rather than // something we can easily make assertions about, we'll run it in a // thread. When the test thread finishes, it will send a message on a // channel to this thread. We'll wait for that message with a fairly // generous timeout, and if we don't receive it, we assume the test // thread has hung. // // Note that it should definitely complete in under a minute, but just // in case CI is slow, we'll give it a long timeout. match done_rx.recv_timeout(timeout) { Err(RecvTimeoutError::Timeout) => panic!( "test did not complete within {timeout:?} seconds, \ we have (probably) entered an infinite loop!", ), // Did the test thread panic? We'll find out for sure when we `join` // with it. Err(RecvTimeoutError::Disconnected) => {} // Test completed successfully! Ok(()) => {} } thread.join().expect("test thread should not panic!") } #[cfg_attr( target_os = "wasi", ignore = "`unwrap()` in `with_timeout()` panics on Wasi" )] #[test] fn drop_cancels_remote_tasks() { // This test reproduces issue #1885. with_timeout(Duration::from_secs(60), || { let (tx, mut rx) = mpsc::channel::<()>(1024); let rt = rt(); let local = LocalSet::new(); local.spawn_local(async move { while rx.recv().await.is_some() {} }); local.block_on(&rt, async { time::sleep(Duration::from_millis(1)).await; }); drop(tx); // This enters an infinite loop if the remote notified tasks are not // properly cancelled. drop(local); }); } #[cfg_attr( target_os = "wasi", ignore = "FIXME: `task::spawn_local().await.unwrap()` panics on Wasi" )] #[test] fn local_tasks_wake_join_all() { // This test reproduces issue #2460. with_timeout(Duration::from_secs(60), || { use futures::future::join_all; use tokio::task::LocalSet; let rt = rt(); let set = LocalSet::new(); let mut handles = Vec::new(); for _ in 1..=128 { handles.push(set.spawn_local(async move { tokio::task::spawn_local(async move {}).await.unwrap(); })); } rt.block_on(set.run_until(join_all(handles))); }); } #[cfg(not(target_os = "wasi"))] // Wasi doesn't support panic recovery #[test] fn local_tasks_are_polled_after_tick() { // This test depends on timing, so we run it up to five times. for _ in 0..4 { let res = std::panic::catch_unwind(local_tasks_are_polled_after_tick_inner); if res.is_ok() { // success return; } } // Test failed 4 times. Try one more time without catching panics. If it // fails again, the test fails. local_tasks_are_polled_after_tick_inner(); } #[cfg(not(target_os = "wasi"))] // Wasi doesn't support panic recovery #[tokio::main(flavor = "current_thread")] async fn local_tasks_are_polled_after_tick_inner() { // Reproduces issues #1899 and #1900 static RX1: AtomicUsize = AtomicUsize::new(0); static RX2: AtomicUsize = AtomicUsize::new(0); const EXPECTED: usize = 500; RX1.store(0, SeqCst); RX2.store(0, SeqCst); let (tx, mut rx) = mpsc::unbounded_channel(); let local = LocalSet::new(); local .run_until(async { let task2 = task::spawn(async move { // Wait a bit time::sleep(Duration::from_millis(10)).await; let mut oneshots = Vec::with_capacity(EXPECTED); // Send values for _ in 0..EXPECTED { let (oneshot_tx, oneshot_rx) = oneshot::channel(); oneshots.push(oneshot_tx); tx.send(oneshot_rx).unwrap(); } time::sleep(Duration::from_millis(10)).await; for tx in oneshots.drain(..) { tx.send(()).unwrap(); } loop { time::sleep(Duration::from_millis(20)).await; let rx1 = RX1.load(SeqCst); let rx2 = RX2.load(SeqCst); if rx1 == EXPECTED && rx2 == EXPECTED { break; } } }); while let Some(oneshot) = rx.recv().await { RX1.fetch_add(1, SeqCst); task::spawn_local(async move { oneshot.await.unwrap(); RX2.fetch_add(1, SeqCst); }); } task2.await.unwrap(); }) .await; } #[tokio::test] async fn acquire_mutex_in_drop() { use futures::future::pending; let (tx1, rx1) = oneshot::channel(); let (tx2, rx2) = oneshot::channel(); let local = LocalSet::new(); local.spawn_local(async move { let _ = rx2.await; unreachable!(); }); local.spawn_local(async move { let _ = rx1.await; tx2.send(()).unwrap(); unreachable!(); }); // Spawn a task that will never notify local.spawn_local(async move { pending::<()>().await; tx1.send(()).unwrap(); }); // Tick the loop local .run_until(async { task::yield_now().await; }) .await; // Drop the LocalSet drop(local); } #[tokio::test] async fn spawn_wakes_localset() { let local = LocalSet::new(); futures::select! { _ = local.run_until(pending::<()>()).fuse() => unreachable!(), ret = async { local.spawn_local(ready(())).await.unwrap()}.fuse() => ret } } /// Checks that the task wakes up with `enter`. /// Reproduces <https://github.com/tokio-rs/tokio/issues/5020>. #[tokio::test] async fn sleep_with_local_enter_guard() { let local = LocalSet::new(); let _guard = local.enter(); let (tx, rx) = oneshot::channel(); local .run_until(async move { tokio::task::spawn_local(async move { time::sleep(Duration::ZERO).await; tx.send(()).expect("failed to send"); }); assert_eq!(rx.await, Ok(())); }) .await; } #[test] fn store_local_set_in_thread_local_with_runtime() { use tokio::runtime::Runtime; thread_local! { static CURRENT: RtAndLocalSet = RtAndLocalSet::new(); } struct RtAndLocalSet { rt: Runtime, local: LocalSet, } impl RtAndLocalSet { fn new() -> RtAndLocalSet { RtAndLocalSet { rt: tokio::runtime::Builder::new_current_thread() .enable_all() .build() .unwrap(), local: LocalSet::new(), } } async fn inner_method(&self) { self.local .run_until(async move { tokio::task::spawn_local(async {}); }) .await } fn method(&self) { self.rt.block_on(self.inner_method()); } } CURRENT.with(|f| { f.method(); }); } #[cfg(tokio_unstable)] mod unstable { use tokio::runtime::UnhandledPanic; use tokio::task::LocalSet; #[tokio::test] #[should_panic( expected = "a spawned task panicked and the LocalSet is configured to shutdown on unhandled panic" )] async fn shutdown_on_panic() { LocalSet::new() .unhandled_panic(UnhandledPanic::ShutdownRuntime) .run_until(async { tokio::task::spawn_local(async { panic!("boom"); }); futures::future::pending::<()>().await; }) .await; } // This test compares that, when the task driving `run_until` has already // consumed budget, the `run_until` future has less budget than a "spawned" // task. // // "Budget" is a fuzzy metric as the Tokio runtime is able to change values // internally. This is why the test uses indirection to test this. #[tokio::test] async fn run_until_does_not_get_own_budget() { // Consume some budget tokio::task::consume_budget().await; LocalSet::new() .run_until(async { let spawned = tokio::spawn(async { let mut spawned_n = 0; { let mut spawned = tokio_test::task::spawn(async { loop { spawned_n += 1; tokio::task::consume_budget().await; } }); // Poll once assert!(!spawned.poll().is_ready()); } spawned_n }); let mut run_until_n = 0; { let mut run_until = tokio_test::task::spawn(async { loop { run_until_n += 1; tokio::task::consume_budget().await; } }); // Poll once assert!(!run_until.poll().is_ready()); } let spawned_n = spawned.await.unwrap(); assert_ne!(spawned_n, 0); assert_ne!(run_until_n, 0); assert!(spawned_n > run_until_n); }) .await } } fn rt() -> runtime::Runtime { tokio::runtime::Builder::new_current_thread() .enable_all() .build() .unwrap() }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_driver.rs
tokio/tests/io_driver.rs
#![warn(rust_2018_idioms)] // Wasi does not support panic recovery or threading #![cfg(all(feature = "full", not(target_os = "wasi")))] use tokio::net::TcpListener; use tokio::runtime; use tokio_test::{assert_ok, assert_pending}; use futures::task::{waker_ref, ArcWake}; use std::future::Future; use std::net::TcpStream; use std::pin::Pin; use std::sync::{mpsc, Arc, Mutex}; use std::task::Context; struct Task<T> { future: Mutex<Pin<Box<T>>>, } impl<T: Send> ArcWake for Task<T> { fn wake_by_ref(_: &Arc<Self>) { // Do nothing... } } impl<T> Task<T> { fn new(future: T) -> Task<T> { Task { future: Mutex::new(Box::pin(future)), } } } #[test] #[cfg_attr(miri, ignore)] // No `socket` in miri. fn test_drop_on_notify() { // When the reactor receives a kernel notification, it notifies the // task that holds the associated socket. If this notification results in // the task being dropped, the socket will also be dropped. // // Previously, there was a deadlock scenario where the reactor, while // notifying, held a lock and the task being dropped attempted to acquire // that same lock in order to clean up state. // // To simulate this case, we create a fake executor that does nothing when // the task is notified. This simulates an executor in the process of // shutting down. Then, when the task handle is dropped, the task itself is // dropped. let rt = runtime::Builder::new_current_thread() .enable_all() .build() .unwrap(); let (addr_tx, addr_rx) = mpsc::channel(); // Define a task that just drains the listener let task = Arc::new(Task::new(async move { // Create a listener let listener = assert_ok!(TcpListener::bind("127.0.0.1:0").await); // Send the address let addr = listener.local_addr().unwrap(); addr_tx.send(addr).unwrap(); loop { let _ = listener.accept().await; } })); { let _enter = rt.enter(); let waker = waker_ref(&task); let mut cx = Context::from_waker(&waker); assert_pending!(task.future.lock().unwrap().as_mut().poll(&mut cx)); } // Get the address let addr = addr_rx.recv().unwrap(); drop(task); // Establish a connection to the acceptor let _s = TcpStream::connect(addr).unwrap(); // Force the reactor to turn rt.block_on(async {}); } #[test] #[should_panic( expected = "A Tokio 1.x context was found, but IO is disabled. Call `enable_io` on the runtime builder to enable IO." )] #[cfg_attr(miri, ignore)] // No `socket` in miri. fn panics_when_io_disabled() { let rt = runtime::Builder::new_current_thread().build().unwrap(); rt.block_on(async { let listener = std::net::TcpListener::bind("127.0.0.1:0").unwrap(); listener.set_nonblocking(true).unwrap(); let _ = tokio::net::TcpListener::from_std(listener); }); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_async_fd.rs
tokio/tests/io_async_fd.rs
#![warn(rust_2018_idioms)] #![cfg(all(unix, feature = "full"))] use std::os::unix::io::{AsRawFd, RawFd}; use std::sync::{ atomic::{AtomicBool, Ordering}, Arc, }; use std::time::Duration; use std::{ future::Future, io::{self, ErrorKind, Read, Write}, task::{Context, Waker}, }; use nix::unistd::{read, write}; use futures::poll; use tokio::io::unix::{AsyncFd, AsyncFdReadyGuard}; use tokio::io::Interest; use tokio_test::{assert_err, assert_pending}; struct TestWaker { inner: Arc<TestWakerInner>, waker: Waker, } #[derive(Default)] struct TestWakerInner { awoken: AtomicBool, } impl futures::task::ArcWake for TestWakerInner { fn wake_by_ref(arc_self: &Arc<Self>) { arc_self.awoken.store(true, Ordering::SeqCst); } } impl TestWaker { fn new() -> Self { let inner: Arc<TestWakerInner> = Default::default(); Self { inner: inner.clone(), waker: futures::task::waker(inner), } } fn awoken(&self) -> bool { self.inner.awoken.swap(false, Ordering::SeqCst) } fn context(&self) -> Context<'_> { Context::from_waker(&self.waker) } } #[derive(Debug)] struct FileDescriptor { fd: std::os::fd::OwnedFd, } impl AsRawFd for FileDescriptor { fn as_raw_fd(&self) -> RawFd { self.fd.as_raw_fd() } } impl Read for &FileDescriptor { fn read(&mut self, buf: &mut [u8]) -> io::Result<usize> { read(self.fd.as_raw_fd(), buf).map_err(io::Error::from) } } impl Read for FileDescriptor { fn read(&mut self, buf: &mut [u8]) -> io::Result<usize> { (self as &Self).read(buf) } } impl Write for &FileDescriptor { fn write(&mut self, buf: &[u8]) -> io::Result<usize> { write(&self.fd, buf).map_err(io::Error::from) } fn flush(&mut self) -> io::Result<()> { Ok(()) } } impl Write for FileDescriptor { fn write(&mut self, buf: &[u8]) -> io::Result<usize> { (self as &Self).write(buf) } fn flush(&mut self) -> io::Result<()> { (self as &Self).flush() } } fn set_nonblocking(fd: RawFd) { use nix::fcntl::{OFlag, F_GETFL, F_SETFL}; let flags = nix::fcntl::fcntl(fd, F_GETFL).expect("fcntl(F_GETFD)"); if flags < 0 { panic!( "bad return value from fcntl(F_GETFL): {} ({:?})", flags, nix::Error::last() ); } let flags = OFlag::from_bits_truncate(flags) | OFlag::O_NONBLOCK; nix::fcntl::fcntl(fd, F_SETFL(flags)).expect("fcntl(F_SETFD)"); } fn socketpair() -> (FileDescriptor, FileDescriptor) { use nix::sys::socket::{self, AddressFamily, SockFlag, SockType}; let (fd_a, fd_b) = socket::socketpair( AddressFamily::Unix, SockType::Stream, None, SockFlag::empty(), ) .expect("socketpair"); let fds = (FileDescriptor { fd: fd_a }, FileDescriptor { fd: fd_b }); set_nonblocking(fds.0.fd.as_raw_fd()); set_nonblocking(fds.1.fd.as_raw_fd()); fds } fn drain(mut fd: &FileDescriptor, mut amt: usize) { let mut buf = [0u8; 512]; while amt > 0 { match fd.read(&mut buf[..]) { Err(e) if e.kind() == ErrorKind::WouldBlock => {} Ok(0) => panic!("unexpected EOF"), Err(e) => panic!("unexpected error: {e:?}"), Ok(x) => amt -= x, } } } #[tokio::test] async fn initially_writable() { let (a, b) = socketpair(); let afd_a = AsyncFd::new(a).unwrap(); let afd_b = AsyncFd::new(b).unwrap(); afd_a.writable().await.unwrap().clear_ready(); afd_b.writable().await.unwrap().clear_ready(); tokio::select! { biased; _ = tokio::time::sleep(Duration::from_millis(10)) => {}, _ = afd_a.readable() => panic!("Unexpected readable state"), _ = afd_b.readable() => panic!("Unexpected readable state"), } } #[tokio::test] async fn reset_readable() { let (a, mut b) = socketpair(); let afd_a = AsyncFd::new(a).unwrap(); let readable = afd_a.readable(); tokio::pin!(readable); tokio::select! { _ = readable.as_mut() => panic!(), _ = tokio::time::sleep(Duration::from_millis(10)) => {} } b.write_all(b"0").unwrap(); let mut guard = readable.await.unwrap(); guard .try_io(|_| afd_a.get_ref().read(&mut [0])) .unwrap() .unwrap(); // `a` is not readable, but the reactor still thinks it is // (because we have not observed a not-ready error yet) afd_a.readable().await.unwrap().retain_ready(); // Explicitly clear the ready state guard.clear_ready(); let readable = afd_a.readable(); tokio::pin!(readable); tokio::select! { _ = readable.as_mut() => panic!(), _ = tokio::time::sleep(Duration::from_millis(10)) => {} } b.write_all(b"0").unwrap(); // We can observe the new readable event afd_a.readable().await.unwrap().clear_ready(); } #[tokio::test] async fn reset_writable() { let (a, b) = socketpair(); let afd_a = AsyncFd::new(a).unwrap(); let mut guard = afd_a.writable().await.unwrap(); // Write until we get a WouldBlock. This also clears the ready state. let mut bytes = 0; while let Ok(Ok(amt)) = guard.try_io(|_| afd_a.get_ref().write(&[0; 512][..])) { bytes += amt; } // Writable state should be cleared now. let writable = afd_a.writable(); tokio::pin!(writable); tokio::select! { _ = writable.as_mut() => panic!(), _ = tokio::time::sleep(Duration::from_millis(10)) => {} } // Read from the other side; we should become writable now. drain(&b, bytes); let _ = writable.await.unwrap(); } #[derive(Debug)] struct ArcFd<T>(Arc<T>); impl<T: AsRawFd> AsRawFd for ArcFd<T> { fn as_raw_fd(&self) -> RawFd { self.0.as_raw_fd() } } #[tokio::test] async fn drop_closes() { let (a, mut b) = socketpair(); let afd_a = AsyncFd::new(a).unwrap(); assert_eq!( ErrorKind::WouldBlock, b.read(&mut [0]).err().unwrap().kind() ); std::mem::drop(afd_a); assert_eq!(0, b.read(&mut [0]).unwrap()); // into_inner does not close the fd let (a, mut b) = socketpair(); let afd_a = AsyncFd::new(a).unwrap(); let _a: FileDescriptor = afd_a.into_inner(); assert_eq!( ErrorKind::WouldBlock, b.read(&mut [0]).err().unwrap().kind() ); // Drop closure behavior is delegated to the inner object let (a, mut b) = socketpair(); let arc_fd = Arc::new(a); let afd_a = AsyncFd::new(ArcFd(arc_fd.clone())).unwrap(); std::mem::drop(afd_a); assert_eq!( ErrorKind::WouldBlock, b.read(&mut [0]).err().unwrap().kind() ); std::mem::drop(arc_fd); // suppress unnecessary clone clippy warning } #[tokio::test] async fn reregister() { let (a, _b) = socketpair(); let afd_a = AsyncFd::new(a).unwrap(); let a = afd_a.into_inner(); AsyncFd::new(a).unwrap(); } #[tokio::test] async fn guard_try_io() { let (a, mut b) = socketpair(); b.write_all(b"0").unwrap(); let afd_a = AsyncFd::new(a).unwrap(); let mut guard = afd_a.readable().await.unwrap(); afd_a.get_ref().read_exact(&mut [0]).unwrap(); // Should not clear the readable state let _ = guard.try_io(|_| Ok(())); // Still readable... let _ = afd_a.readable().await.unwrap(); // Should clear the readable state let _ = guard.try_io(|_| io::Result::<()>::Err(ErrorKind::WouldBlock.into())); // Assert not readable let readable = afd_a.readable(); tokio::pin!(readable); tokio::select! { _ = readable.as_mut() => panic!(), _ = tokio::time::sleep(Duration::from_millis(10)) => {} } // Write something down b again and make sure we're reawoken b.write_all(b"0").unwrap(); let _ = readable.await.unwrap(); } #[tokio::test] async fn try_io_readable() { let (a, mut b) = socketpair(); let mut afd_a = AsyncFd::new(a).unwrap(); // Give the runtime some time to update bookkeeping. tokio::task::yield_now().await; { let mut called = false; let _ = afd_a.try_io_mut(Interest::READABLE, |_| { called = true; Ok(()) }); assert!( !called, "closure should not have been called, since socket should not be readable" ); } // Make `a` readable by writing to `b`. // Give the runtime some time to update bookkeeping. b.write_all(&[0]).unwrap(); tokio::task::yield_now().await; { let mut called = false; let _ = afd_a.try_io(Interest::READABLE, |_| { called = true; Ok(()) }); assert!( called, "closure should have been called, since socket should have data available to read" ); } { let mut called = false; let _ = afd_a.try_io(Interest::READABLE, |_| { called = true; io::Result::<()>::Err(ErrorKind::WouldBlock.into()) }); assert!( called, "closure should have been called, since socket should have data available to read" ); } { let mut called = false; let _ = afd_a.try_io(Interest::READABLE, |_| { called = true; Ok(()) }); assert!(!called, "closure should not have been called, since socket readable state should have been cleared"); } } #[tokio::test] async fn try_io_writable() { let (a, _b) = socketpair(); let afd_a = AsyncFd::new(a).unwrap(); // Give the runtime some time to update bookkeeping. tokio::task::yield_now().await; { let mut called = false; let _ = afd_a.try_io(Interest::WRITABLE, |_| { called = true; Ok(()) }); assert!( called, "closure should have been called, since socket should still be marked as writable" ); } { let mut called = false; let _ = afd_a.try_io(Interest::WRITABLE, |_| { called = true; io::Result::<()>::Err(ErrorKind::WouldBlock.into()) }); assert!( called, "closure should have been called, since socket should still be marked as writable" ); } { let mut called = false; let _ = afd_a.try_io(Interest::WRITABLE, |_| { called = true; Ok(()) }); assert!(!called, "closure should not have been called, since socket writable state should have been cleared"); } } #[tokio::test] async fn multiple_waiters() { let (a, mut b) = socketpair(); let afd_a = Arc::new(AsyncFd::new(a).unwrap()); let barrier = Arc::new(tokio::sync::Barrier::new(11)); let mut tasks = Vec::new(); for _ in 0..10 { let afd_a = afd_a.clone(); let barrier = barrier.clone(); let f = async move { let notify_barrier = async { barrier.wait().await; futures::future::pending::<()>().await; }; tokio::select! { biased; guard = afd_a.readable() => { tokio::task::yield_now().await; guard.unwrap().clear_ready() }, _ = notify_barrier => unreachable!(), } std::mem::drop(afd_a); }; tasks.push(tokio::spawn(f)); } let mut all_tasks = futures::future::try_join_all(tasks); tokio::select! { r = std::pin::Pin::new(&mut all_tasks) => { r.unwrap(); // propagate panic panic!("Tasks exited unexpectedly") }, _ = barrier.wait() => {} } b.write_all(b"0").unwrap(); all_tasks.await.unwrap(); } #[tokio::test] async fn poll_fns() { let (a, b) = socketpair(); let afd_a = Arc::new(AsyncFd::new(a).unwrap()); let afd_b = Arc::new(AsyncFd::new(b).unwrap()); // Fill up the write side of A let mut bytes = 0; while let Ok(amt) = afd_a.get_ref().write(&[0; 512]) { bytes += amt; } let waker = TestWaker::new(); assert_pending!(afd_a.as_ref().poll_read_ready(&mut waker.context())); let afd_a_2 = afd_a.clone(); let r_barrier = Arc::new(tokio::sync::Barrier::new(2)); let barrier_clone = r_barrier.clone(); let read_fut = tokio::spawn(async move { // Move waker onto this task first assert_pending!(poll!(std::future::poll_fn(|cx| afd_a_2 .as_ref() .poll_read_ready(cx)))); barrier_clone.wait().await; let _ = std::future::poll_fn(|cx| afd_a_2.as_ref().poll_read_ready(cx)).await; }); let afd_a_2 = afd_a.clone(); let w_barrier = Arc::new(tokio::sync::Barrier::new(2)); let barrier_clone = w_barrier.clone(); let mut write_fut = tokio::spawn(async move { // Move waker onto this task first assert_pending!(poll!(std::future::poll_fn(|cx| afd_a_2 .as_ref() .poll_write_ready(cx)))); barrier_clone.wait().await; let _ = std::future::poll_fn(|cx| afd_a_2.as_ref().poll_write_ready(cx)).await; }); r_barrier.wait().await; w_barrier.wait().await; let readable = afd_a.readable(); tokio::pin!(readable); tokio::select! { _ = &mut readable => unreachable!(), _ = tokio::task::yield_now() => {} } // Make A readable. We expect that 'readable' and 'read_fut' will both complete quickly afd_b.get_ref().write_all(b"0").unwrap(); let _ = tokio::join!(readable, read_fut); // Our original waker should _not_ be awoken (poll_read_ready retains only the last context) assert!(!waker.awoken()); // The writable side should not be awoken tokio::select! { _ = &mut write_fut => unreachable!(), _ = tokio::time::sleep(Duration::from_millis(5)) => {} } // Make it writable now drain(afd_b.get_ref(), bytes); // now we should be writable (ie - the waker for poll_write should still be registered after we wake the read side) let _ = write_fut.await; } fn assert_pending<T: std::fmt::Debug, F: Future<Output = T>>(f: F) -> std::pin::Pin<Box<F>> { let mut pinned = Box::pin(f); assert_pending!(pinned .as_mut() .poll(&mut Context::from_waker(futures::task::noop_waker_ref()))); pinned } fn rt() -> tokio::runtime::Runtime { tokio::runtime::Builder::new_current_thread() .enable_all() .build() .unwrap() } #[test] fn driver_shutdown_wakes_currently_pending() { let rt = rt(); let (a, _b) = socketpair(); let afd_a = { let _enter = rt.enter(); AsyncFd::new(a).unwrap() }; let readable = assert_pending(afd_a.readable()); std::mem::drop(rt); // The future was initialized **before** dropping the rt assert_err!(futures::executor::block_on(readable)); // The future is initialized **after** dropping the rt. assert_err!(futures::executor::block_on(afd_a.readable())); } #[test] fn driver_shutdown_wakes_future_pending() { let rt = rt(); let (a, _b) = socketpair(); let afd_a = { let _enter = rt.enter(); AsyncFd::new(a).unwrap() }; std::mem::drop(rt); assert_err!(futures::executor::block_on(afd_a.readable())); } #[test] fn driver_shutdown_wakes_pending_race() { // TODO: make this a loom test for _ in 0..100 { let rt = rt(); let (a, _b) = socketpair(); let afd_a = { let _enter = rt.enter(); AsyncFd::new(a).unwrap() }; let _ = std::thread::spawn(move || std::mem::drop(rt)); // This may or may not return an error (but will be awoken) let _ = futures::executor::block_on(afd_a.readable()); // However retrying will always return an error assert_err!(futures::executor::block_on(afd_a.readable())); } } async fn poll_readable<T: AsRawFd>(fd: &AsyncFd<T>) -> std::io::Result<AsyncFdReadyGuard<'_, T>> { std::future::poll_fn(|cx| fd.poll_read_ready(cx)).await } async fn poll_writable<T: AsRawFd>(fd: &AsyncFd<T>) -> std::io::Result<AsyncFdReadyGuard<'_, T>> { std::future::poll_fn(|cx| fd.poll_write_ready(cx)).await } #[test] fn driver_shutdown_wakes_currently_pending_polls() { let rt = rt(); let (a, _b) = socketpair(); let afd_a = { let _enter = rt.enter(); AsyncFd::new(a).unwrap() }; while afd_a.get_ref().write(&[0; 512]).is_ok() {} // make not writable let readable = assert_pending(poll_readable(&afd_a)); let writable = assert_pending(poll_writable(&afd_a)); std::mem::drop(rt); // Attempting to poll readiness when the rt is dropped is an error assert_err!(futures::executor::block_on(readable)); assert_err!(futures::executor::block_on(writable)); } #[test] fn driver_shutdown_wakes_poll() { let rt = rt(); let (a, _b) = socketpair(); let afd_a = { let _enter = rt.enter(); AsyncFd::new(a).unwrap() }; std::mem::drop(rt); assert_err!(futures::executor::block_on(poll_readable(&afd_a))); assert_err!(futures::executor::block_on(poll_writable(&afd_a))); } #[test] fn driver_shutdown_then_clear_readiness() { let rt = rt(); let (a, _b) = socketpair(); let afd_a = { let _enter = rt.enter(); AsyncFd::new(a).unwrap() }; let mut write_ready = rt.block_on(afd_a.writable()).unwrap(); std::mem::drop(rt); write_ready.clear_ready(); } #[test] fn driver_shutdown_wakes_poll_race() { // TODO: make this a loom test for _ in 0..100 { let rt = rt(); let (a, _b) = socketpair(); let afd_a = { let _enter = rt.enter(); AsyncFd::new(a).unwrap() }; while afd_a.get_ref().write(&[0; 512]).is_ok() {} // make not writable let _ = std::thread::spawn(move || std::mem::drop(rt)); // The poll variants will always return an error in this case assert_err!(futures::executor::block_on(poll_readable(&afd_a))); assert_err!(futures::executor::block_on(poll_writable(&afd_a))); } } #[tokio::test] #[cfg_attr(miri, ignore)] // No socket in miri. #[cfg(any(target_os = "linux", target_os = "android"))] async fn priority_event_on_oob_data() { use std::net::SocketAddr; use tokio::io::Interest; let addr: SocketAddr = "127.0.0.1:0".parse().unwrap(); let listener = std::net::TcpListener::bind(addr).unwrap(); let addr = listener.local_addr().unwrap(); let client = std::net::TcpStream::connect(addr).unwrap(); let client = AsyncFd::with_interest(client, Interest::PRIORITY).unwrap(); let (stream, _) = listener.accept().unwrap(); // Sending out of band data should trigger priority event. send_oob_data(&stream, b"hello").unwrap(); let _ = client.ready(Interest::PRIORITY).await.unwrap(); } #[cfg(any(target_os = "linux", target_os = "android"))] fn send_oob_data<S: AsRawFd>(stream: &S, data: &[u8]) -> io::Result<usize> { unsafe { let res = libc::send( stream.as_raw_fd(), data.as_ptr().cast(), data.len(), libc::MSG_OOB, ); if res == -1 { Err(io::Error::last_os_error()) } else { Ok(res as usize) } } } #[tokio::test] async fn clear_ready_matching_clears_ready() { use tokio::io::{Interest, Ready}; let (a, mut b) = socketpair(); let afd_a = AsyncFd::new(a).unwrap(); b.write_all(b"0").unwrap(); let mut guard = afd_a .ready(Interest::READABLE | Interest::WRITABLE) .await .unwrap(); assert_eq!(guard.ready(), Ready::READABLE | Ready::WRITABLE); guard.clear_ready_matching(Ready::READABLE); assert_eq!(guard.ready(), Ready::WRITABLE); guard.clear_ready_matching(Ready::WRITABLE); assert_eq!(guard.ready(), Ready::EMPTY); } #[tokio::test] async fn clear_ready_matching_clears_ready_mut() { use tokio::io::{Interest, Ready}; let (a, mut b) = socketpair(); let mut afd_a = AsyncFd::new(a).unwrap(); b.write_all(b"0").unwrap(); let mut guard = afd_a .ready_mut(Interest::READABLE | Interest::WRITABLE) .await .unwrap(); assert_eq!(guard.ready(), Ready::READABLE | Ready::WRITABLE); guard.clear_ready_matching(Ready::READABLE); assert_eq!(guard.ready(), Ready::WRITABLE); guard.clear_ready_matching(Ready::WRITABLE); assert_eq!(guard.ready(), Ready::EMPTY); } #[tokio::test] #[cfg_attr(miri, ignore)] // No socket in miri. #[cfg(target_os = "linux")] async fn await_error_readiness_timestamping() { use std::net::{Ipv4Addr, SocketAddr}; use tokio::io::{Interest, Ready}; let address_a = SocketAddr::from((Ipv4Addr::LOCALHOST, 0)); let address_b = SocketAddr::from((Ipv4Addr::LOCALHOST, 0)); let socket = std::net::UdpSocket::bind(address_a).unwrap(); socket.set_nonblocking(true).unwrap(); // configure send timestamps configure_timestamping_socket(&socket).unwrap(); socket.connect(address_b).unwrap(); let fd = AsyncFd::new(socket).unwrap(); tokio::select! { _ = fd.ready(Interest::ERROR) => panic!(), _ = tokio::time::sleep(Duration::from_millis(10)) => {} } let buf = b"hello there"; fd.get_ref().send(buf).unwrap(); // the send timestamp should now be in the error queue let guard = fd.ready(Interest::ERROR).await.unwrap(); assert_eq!(guard.ready(), Ready::ERROR); } #[cfg(target_os = "linux")] fn configure_timestamping_socket(udp_socket: &std::net::UdpSocket) -> std::io::Result<libc::c_int> { // enable software timestamping, and specifically software send timestamping let options = libc::SOF_TIMESTAMPING_SOFTWARE | libc::SOF_TIMESTAMPING_TX_SOFTWARE; let res = unsafe { libc::setsockopt( udp_socket.as_raw_fd(), libc::SOL_SOCKET, libc::SO_TIMESTAMP, &options as *const _ as *const libc::c_void, std::mem::size_of_val(&options) as libc::socklen_t, ) }; if res == -1 { Err(std::io::Error::last_os_error()) } else { Ok(res) } } #[tokio::test] #[cfg_attr(miri, ignore)] // No socket in miri. #[cfg(target_os = "linux")] async fn await_error_readiness_invalid_address() { use std::net::{Ipv4Addr, SocketAddr}; use tokio::io::{Interest, Ready}; let socket_addr = SocketAddr::from((Ipv4Addr::LOCALHOST, 0)); let socket = std::net::UdpSocket::bind(socket_addr).unwrap(); let socket_fd = socket.as_raw_fd(); // Enable IP_RECVERR option to receive error messages // https://man7.org/linux/man-pages/man7/ip.7.html has some extra information let recv_err: libc::c_int = 1; unsafe { let res = libc::setsockopt( socket.as_raw_fd(), libc::SOL_IP, libc::IP_RECVERR, &recv_err as *const _ as *const libc::c_void, std::mem::size_of_val(&recv_err) as libc::socklen_t, ); if res == -1 { panic!("{:?}", std::io::Error::last_os_error()); } } // Spawn a separate thread for sending messages tokio::spawn(async move { // Set the destination address. This address is invalid in this context. the OS will notice // that nobody is listening on port this port. Normally this is ignored (UDP is "fire and forget"), // but because IP_RECVERR is enabled, the error will actually be reported to the sending socket let mut dest_addr = unsafe { std::mem::MaybeUninit::<libc::sockaddr_in>::zeroed().assume_init() }; dest_addr.sin_family = libc::AF_INET as _; // based on https://en.wikipedia.org/wiki/Ephemeral_port, we should pick a port number // below 1024 to guarantee that other tests don't select this port by accident when they // use port 0 to select an ephemeral port. dest_addr.sin_port = 512u16.to_be(); // Destination port // Prepare the message data let message = "Hello, Socket!"; // Prepare the message structure for sendmsg let mut iov = libc::iovec { iov_base: message.as_ptr() as *mut libc::c_void, iov_len: message.len(), }; // Prepare the destination address for the sendmsg call let dest_sockaddr: *const libc::sockaddr = &dest_addr as *const _ as *const libc::sockaddr; let dest_addrlen: libc::socklen_t = std::mem::size_of_val(&dest_addr) as libc::socklen_t; let mut msg: libc::msghdr = unsafe { std::mem::MaybeUninit::zeroed().assume_init() }; msg.msg_name = dest_sockaddr as *mut libc::c_void; msg.msg_namelen = dest_addrlen; msg.msg_iov = &mut iov; msg.msg_iovlen = 1; if unsafe { libc::sendmsg(socket_fd, &msg, 0) } == -1 { panic!("{:?}", std::io::Error::last_os_error()) } }); let fd = AsyncFd::new(socket).unwrap(); let guard = fd.ready(Interest::ERROR).await.unwrap(); assert_eq!(guard.ready(), Ready::ERROR); } #[derive(Debug, PartialEq, Eq)] struct InvalidSource; impl AsRawFd for InvalidSource { fn as_raw_fd(&self) -> RawFd { -1 } } #[tokio::test] async fn try_new() { let original = Arc::new(InvalidSource); let error = AsyncFd::try_new(original.clone()).unwrap_err(); let (returned, _cause) = error.into_parts(); assert!(Arc::ptr_eq(&original, &returned)); } #[tokio::test] async fn try_with_interest() { let original = Arc::new(InvalidSource); let error = AsyncFd::try_with_interest(original.clone(), Interest::READABLE).unwrap_err(); let (returned, _cause) = error.into_parts(); assert!(Arc::ptr_eq(&original, &returned)); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/fs_open_options_windows.rs
tokio/tests/fs_open_options_windows.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // WASI does not support all fs operations #![cfg(windows)] use tokio::fs::OpenOptions; use windows_sys::Win32::Storage::FileSystem; #[tokio::test] #[cfg(windows)] async fn open_options_windows_access_mode() { // TESTING HACK: use Debug output to check the stored data assert!(format!("{:?}", OpenOptions::new().access_mode(0)).contains("access_mode: Some(0)")); } #[tokio::test] #[cfg(windows)] async fn open_options_windows_share_mode() { // TESTING HACK: use Debug output to check the stored data assert!(format!("{:?}", OpenOptions::new().share_mode(0)).contains("share_mode: 0,")); } #[tokio::test] #[cfg(windows)] async fn open_options_windows_custom_flags() { // TESTING HACK: use Debug output to check the stored data assert!(format!( "{:?}", OpenOptions::new().custom_flags(FileSystem::FILE_FLAG_DELETE_ON_CLOSE) ) .contains("custom_flags: 67108864,")); } #[tokio::test] #[cfg(windows)] async fn open_options_windows_attributes() { assert!(format!( "{:?}", OpenOptions::new().attributes(FileSystem::FILE_ATTRIBUTE_HIDDEN) ) .contains("attributes: 2,")); } #[tokio::test] #[cfg(windows)] async fn open_options_windows_security_qos_flags() { assert!(format!( "{:?}", OpenOptions::new().security_qos_flags(FileSystem::SECURITY_IDENTIFICATION) ) .contains("security_qos_flags: 1114112,")); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/sync_broadcast.rs
tokio/tests/sync_broadcast.rs
#![allow(clippy::cognitive_complexity)] #![warn(rust_2018_idioms)] #![cfg(feature = "sync")] #[cfg(all(target_family = "wasm", not(target_os = "wasi")))] use wasm_bindgen_test::wasm_bindgen_test as test; use tokio::sync::broadcast; use tokio_test::task; use tokio_test::{ assert_err, assert_ok, assert_pending, assert_ready, assert_ready_err, assert_ready_ok, }; use std::sync::Arc; macro_rules! assert_recv { ($e:expr) => { match $e.try_recv() { Ok(value) => value, Err(e) => panic!("expected recv; got = {:?}", e), } }; } macro_rules! assert_empty { ($e:expr) => { match $e.try_recv() { Ok(value) => panic!("expected empty; got = {:?}", value), Err(broadcast::error::TryRecvError::Empty) => {} Err(e) => panic!("expected empty; got = {:?}", e), } }; } macro_rules! assert_lagged { ($e:expr, $n:expr) => { match assert_err!($e) { broadcast::error::TryRecvError::Lagged(n) => { assert_eq!(n, $n); } _ => panic!("did not lag"), } }; } macro_rules! assert_closed { ($e:expr) => { match assert_err!($e) { broadcast::error::TryRecvError::Closed => {} _ => panic!("is not closed"), } }; } #[allow(unused)] trait AssertSend: Send + Sync {} impl AssertSend for broadcast::Sender<i32> {} impl AssertSend for broadcast::Receiver<i32> {} impl AssertSend for broadcast::WeakSender<i32> {} #[test] fn send_try_recv_bounded() { let (tx, mut rx) = broadcast::channel(16); assert_empty!(rx); let n = assert_ok!(tx.send("hello")); assert_eq!(n, 1); let val = assert_recv!(rx); assert_eq!(val, "hello"); assert_empty!(rx); } #[test] fn send_two_recv() { let (tx, mut rx1) = broadcast::channel(16); let mut rx2 = tx.subscribe(); assert_empty!(rx1); assert_empty!(rx2); let n = assert_ok!(tx.send("hello")); assert_eq!(n, 2); let val = assert_recv!(rx1); assert_eq!(val, "hello"); let val = assert_recv!(rx2); assert_eq!(val, "hello"); assert_empty!(rx1); assert_empty!(rx2); } #[test] fn send_recv_bounded() { let (tx, mut rx) = broadcast::channel(16); let mut recv = task::spawn(rx.recv()); assert_pending!(recv.poll()); assert_ok!(tx.send("hello")); assert!(recv.is_woken()); let val = assert_ready_ok!(recv.poll()); assert_eq!(val, "hello"); } #[test] fn send_two_recv_bounded() { let (tx, mut rx1) = broadcast::channel(16); let mut rx2 = tx.subscribe(); let mut recv1 = task::spawn(rx1.recv()); let mut recv2 = task::spawn(rx2.recv()); assert_pending!(recv1.poll()); assert_pending!(recv2.poll()); assert_ok!(tx.send("hello")); assert!(recv1.is_woken()); assert!(recv2.is_woken()); let val1 = assert_ready_ok!(recv1.poll()); let val2 = assert_ready_ok!(recv2.poll()); assert_eq!(val1, "hello"); assert_eq!(val2, "hello"); drop((recv1, recv2)); let mut recv1 = task::spawn(rx1.recv()); let mut recv2 = task::spawn(rx2.recv()); assert_pending!(recv1.poll()); assert_ok!(tx.send("world")); assert!(recv1.is_woken()); assert!(!recv2.is_woken()); let val1 = assert_ready_ok!(recv1.poll()); let val2 = assert_ready_ok!(recv2.poll()); assert_eq!(val1, "world"); assert_eq!(val2, "world"); } #[test] fn change_tasks() { let (tx, mut rx) = broadcast::channel(1); let mut recv = Box::pin(rx.recv()); let mut task1 = task::spawn(&mut recv); assert_pending!(task1.poll()); let mut task2 = task::spawn(&mut recv); assert_pending!(task2.poll()); tx.send("hello").unwrap(); assert!(task2.is_woken()); } #[test] fn send_slow_rx() { let (tx, mut rx1) = broadcast::channel(16); let mut rx2 = tx.subscribe(); { let mut recv2 = task::spawn(rx2.recv()); { let mut recv1 = task::spawn(rx1.recv()); assert_pending!(recv1.poll()); assert_pending!(recv2.poll()); assert_ok!(tx.send("one")); assert!(recv1.is_woken()); assert!(recv2.is_woken()); assert_ok!(tx.send("two")); let val = assert_ready_ok!(recv1.poll()); assert_eq!(val, "one"); } let val = assert_ready_ok!(task::spawn(rx1.recv()).poll()); assert_eq!(val, "two"); let mut recv1 = task::spawn(rx1.recv()); assert_pending!(recv1.poll()); assert_ok!(tx.send("three")); assert!(recv1.is_woken()); let val = assert_ready_ok!(recv1.poll()); assert_eq!(val, "three"); let val = assert_ready_ok!(recv2.poll()); assert_eq!(val, "one"); } let val = assert_recv!(rx2); assert_eq!(val, "two"); let val = assert_recv!(rx2); assert_eq!(val, "three"); } #[test] fn drop_rx_while_values_remain() { let (tx, mut rx1) = broadcast::channel(16); let mut rx2 = tx.subscribe(); assert_ok!(tx.send("one")); assert_ok!(tx.send("two")); assert_recv!(rx1); assert_recv!(rx2); drop(rx2); drop(rx1); } #[test] fn lagging_rx() { let (tx, mut rx1) = broadcast::channel(2); let mut rx2 = tx.subscribe(); assert_ok!(tx.send("one")); assert_ok!(tx.send("two")); assert_eq!("one", assert_recv!(rx1)); assert_ok!(tx.send("three")); // Lagged too far let x = dbg!(rx2.try_recv()); assert_lagged!(x, 1); // Calling again gets the next value assert_eq!("two", assert_recv!(rx2)); assert_eq!("two", assert_recv!(rx1)); assert_eq!("three", assert_recv!(rx1)); assert_ok!(tx.send("four")); assert_ok!(tx.send("five")); assert_lagged!(rx2.try_recv(), 1); assert_ok!(tx.send("six")); assert_lagged!(rx2.try_recv(), 1); } #[test] fn send_no_rx() { let (tx, _) = broadcast::channel(16); assert_err!(tx.send("hello")); let mut rx = tx.subscribe(); assert_ok!(tx.send("world")); let val = assert_recv!(rx); assert_eq!("world", val); } #[test] #[should_panic] #[cfg(not(target_family = "wasm"))] // wasm currently doesn't support unwinding fn zero_capacity() { broadcast::channel::<()>(0); } #[test] #[should_panic] #[cfg(not(target_family = "wasm"))] // wasm currently doesn't support unwinding fn capacity_too_big() { broadcast::channel::<()>(1 + (usize::MAX >> 1)); } #[test] #[cfg(panic = "unwind")] #[cfg(not(target_family = "wasm"))] // wasm currently doesn't support unwinding fn panic_in_clone() { use std::panic::{self, AssertUnwindSafe}; #[derive(Eq, PartialEq, Debug)] struct MyVal(usize); impl Clone for MyVal { fn clone(&self) -> MyVal { assert_ne!(0, self.0); MyVal(self.0) } } let (tx, mut rx) = broadcast::channel(16); assert_ok!(tx.send(MyVal(0))); assert_ok!(tx.send(MyVal(1))); let res = panic::catch_unwind(AssertUnwindSafe(|| { let _ = rx.try_recv(); })); assert_err!(res); let val = assert_recv!(rx); assert_eq!(val, MyVal(1)); } #[test] fn dropping_tx_notifies_rx() { let (tx, mut rx1) = broadcast::channel::<()>(16); let mut rx2 = tx.subscribe(); let tx2 = tx.clone(); let mut recv1 = task::spawn(rx1.recv()); let mut recv2 = task::spawn(rx2.recv()); assert_pending!(recv1.poll()); assert_pending!(recv2.poll()); drop(tx); assert_pending!(recv1.poll()); assert_pending!(recv2.poll()); drop(tx2); assert!(recv1.is_woken()); assert!(recv2.is_woken()); let err = assert_ready_err!(recv1.poll()); assert!(is_closed(err)); let err = assert_ready_err!(recv2.poll()); assert!(is_closed(err)); } #[test] fn unconsumed_messages_are_dropped() { let (tx, rx) = broadcast::channel(16); let msg = Arc::new(()); assert_ok!(tx.send(msg.clone())); assert_eq!(2, Arc::strong_count(&msg)); drop(rx); assert_eq!(1, Arc::strong_count(&msg)); } #[test] fn single_capacity_recvs() { let (tx, mut rx) = broadcast::channel(1); assert_ok!(tx.send(1)); assert_eq!(assert_recv!(rx), 1); assert_empty!(rx); } #[test] fn single_capacity_recvs_after_drop_1() { let (tx, mut rx) = broadcast::channel(1); assert_ok!(tx.send(1)); drop(tx); assert_eq!(assert_recv!(rx), 1); assert_closed!(rx.try_recv()); } #[test] fn single_capacity_recvs_after_drop_2() { let (tx, mut rx) = broadcast::channel(1); assert_ok!(tx.send(1)); assert_ok!(tx.send(2)); drop(tx); assert_lagged!(rx.try_recv(), 1); assert_eq!(assert_recv!(rx), 2); assert_closed!(rx.try_recv()); } #[test] fn dropping_sender_does_not_overwrite() { let (tx, mut rx) = broadcast::channel(2); assert_ok!(tx.send(1)); assert_ok!(tx.send(2)); drop(tx); assert_eq!(assert_recv!(rx), 1); assert_eq!(assert_recv!(rx), 2); assert_closed!(rx.try_recv()); } #[test] fn lagging_receiver_recovers_after_wrap_closed_1() { let (tx, mut rx) = broadcast::channel(2); assert_ok!(tx.send(1)); assert_ok!(tx.send(2)); assert_ok!(tx.send(3)); drop(tx); assert_lagged!(rx.try_recv(), 1); assert_eq!(assert_recv!(rx), 2); assert_eq!(assert_recv!(rx), 3); assert_closed!(rx.try_recv()); } #[test] fn lagging_receiver_recovers_after_wrap_closed_2() { let (tx, mut rx) = broadcast::channel(2); assert_ok!(tx.send(1)); assert_ok!(tx.send(2)); assert_ok!(tx.send(3)); assert_ok!(tx.send(4)); drop(tx); assert_lagged!(rx.try_recv(), 2); assert_eq!(assert_recv!(rx), 3); assert_eq!(assert_recv!(rx), 4); assert_closed!(rx.try_recv()); } #[test] fn lagging_receiver_recovers_after_wrap_open() { let (tx, mut rx) = broadcast::channel(2); assert_ok!(tx.send(1)); assert_ok!(tx.send(2)); assert_ok!(tx.send(3)); assert_lagged!(rx.try_recv(), 1); assert_eq!(assert_recv!(rx), 2); assert_eq!(assert_recv!(rx), 3); assert_empty!(rx); } #[test] fn receiver_len_with_lagged() { let (tx, mut rx) = broadcast::channel(3); tx.send(10).unwrap(); tx.send(20).unwrap(); tx.send(30).unwrap(); tx.send(40).unwrap(); assert_eq!(rx.len(), 4); assert_eq!(assert_recv!(rx), 10); tx.send(50).unwrap(); tx.send(60).unwrap(); assert_eq!(rx.len(), 5); assert_lagged!(rx.try_recv(), 1); } fn is_closed(err: broadcast::error::RecvError) -> bool { matches!(err, broadcast::error::RecvError::Closed) } #[test] fn resubscribe_points_to_tail() { let (tx, mut rx) = broadcast::channel(3); tx.send(1).unwrap(); let mut rx_resub = rx.resubscribe(); // verify we're one behind at the start assert_empty!(rx_resub); assert_eq!(assert_recv!(rx), 1); // verify we do not affect rx tx.send(2).unwrap(); assert_eq!(assert_recv!(rx_resub), 2); tx.send(3).unwrap(); assert_eq!(assert_recv!(rx), 2); assert_eq!(assert_recv!(rx), 3); assert_empty!(rx); assert_eq!(assert_recv!(rx_resub), 3); assert_empty!(rx_resub); } #[test] fn resubscribe_lagged() { let (tx, mut rx) = broadcast::channel(1); tx.send(1).unwrap(); tx.send(2).unwrap(); let mut rx_resub = rx.resubscribe(); assert_lagged!(rx.try_recv(), 1); assert_empty!(rx_resub); assert_eq!(assert_recv!(rx), 2); assert_empty!(rx); assert_empty!(rx_resub); } #[test] fn resubscribe_to_closed_channel() { let (tx, rx) = tokio::sync::broadcast::channel::<u32>(2); drop(tx); let mut rx_resub = rx.resubscribe(); assert_closed!(rx_resub.try_recv()); } #[test] fn sender_len() { let (tx, mut rx1) = broadcast::channel(4); let mut rx2 = tx.subscribe(); assert_eq!(tx.len(), 0); assert!(tx.is_empty()); tx.send(1).unwrap(); tx.send(2).unwrap(); tx.send(3).unwrap(); assert_eq!(tx.len(), 3); assert!(!tx.is_empty()); assert_recv!(rx1); assert_recv!(rx1); assert_eq!(tx.len(), 3); assert!(!tx.is_empty()); assert_recv!(rx2); assert_eq!(tx.len(), 2); assert!(!tx.is_empty()); tx.send(4).unwrap(); tx.send(5).unwrap(); tx.send(6).unwrap(); assert_eq!(tx.len(), 4); assert!(!tx.is_empty()); } #[test] #[cfg(not(all(target_family = "wasm", not(target_os = "wasi"))))] fn sender_len_random() { let (tx, mut rx1) = broadcast::channel(16); let mut rx2 = tx.subscribe(); for _ in 0..1000 { match rand::random_range(0..4) { 0 => { let _ = rx1.try_recv(); } 1 => { let _ = rx2.try_recv(); } _ => { tx.send(0).unwrap(); } } let expected_len = usize::min(usize::max(rx1.len(), rx2.len()), 16); assert_eq!(tx.len(), expected_len); } } #[test] fn send_in_waker_drop() { use futures::task::ArcWake; use std::future::Future; use std::task::Context; struct SendOnDrop(broadcast::Sender<()>); impl Drop for SendOnDrop { fn drop(&mut self) { let _ = self.0.send(()); } } impl ArcWake for SendOnDrop { fn wake_by_ref(_arc_self: &Arc<Self>) {} } // Test if there is no deadlock when replacing the old waker. let (tx, mut rx) = broadcast::channel(16); let mut fut = Box::pin(async { let _ = rx.recv().await; }); // Store our special waker in the receiving future. let waker = futures::task::waker(Arc::new(SendOnDrop(tx))); let mut cx = Context::from_waker(&waker); assert!(fut.as_mut().poll(&mut cx).is_pending()); drop(waker); // Second poll shouldn't deadlock. let mut cx = Context::from_waker(futures::task::noop_waker_ref()); let _ = fut.as_mut().poll(&mut cx); // Test if there is no deadlock when calling waker.wake(). let (tx, mut rx) = broadcast::channel(16); let mut fut = Box::pin(async { let _ = rx.recv().await; }); // Store our special waker in the receiving future. let waker = futures::task::waker(Arc::new(SendOnDrop(tx.clone()))); let mut cx = Context::from_waker(&waker); assert!(fut.as_mut().poll(&mut cx).is_pending()); drop(waker); // Shouldn't deadlock. let _ = tx.send(()); } #[tokio::test] async fn receiver_recv_is_cooperative() { let (tx, mut rx) = broadcast::channel(8); tokio::select! { biased; _ = async { loop { assert!(tx.send(()).is_ok()); assert!(rx.recv().await.is_ok()); } } => {}, _ = tokio::task::yield_now() => {}, } } #[test] fn broadcast_sender_closed() { let (tx, rx) = broadcast::channel::<()>(1); let rx2 = tx.subscribe(); let mut task = task::spawn(tx.closed()); assert_pending!(task.poll()); drop(rx); assert!(!task.is_woken()); assert_pending!(task.poll()); drop(rx2); assert!(task.is_woken()); assert_ready!(task.poll()); } #[test] fn broadcast_sender_closed_with_extra_subscribe() { let (tx, rx) = broadcast::channel::<()>(1); let rx2 = tx.subscribe(); let mut task = task::spawn(tx.closed()); assert_pending!(task.poll()); drop(rx); assert!(!task.is_woken()); assert_pending!(task.poll()); drop(rx2); assert!(task.is_woken()); let rx3 = tx.subscribe(); assert_pending!(task.poll()); drop(rx3); assert!(task.is_woken()); assert_ready!(task.poll()); let mut task2 = task::spawn(tx.closed()); assert_ready!(task2.poll()); let rx4 = tx.subscribe(); let mut task3 = task::spawn(tx.closed()); assert_pending!(task3.poll()); drop(rx4); assert!(task3.is_woken()); assert_ready!(task3.poll()); } #[tokio::test] async fn broadcast_sender_new_must_be_closed() { let capacity = 1; let tx: broadcast::Sender<()> = broadcast::Sender::new(capacity); let mut task = task::spawn(tx.closed()); assert_ready!(task.poll()); let _rx = tx.subscribe(); let mut task2 = task::spawn(tx.closed()); assert_pending!(task2.poll()); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/time_panic.rs
tokio/tests/time_panic.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // Wasi doesn't support panic recovery #![cfg(panic = "unwind")] use futures::future; use std::error::Error; use std::time::Duration; use tokio::runtime::{Builder, Runtime}; use tokio::time::{self, interval, interval_at, timeout, Instant}; mod support { pub mod panic; } use support::panic::test_panic; fn rt_combinations() -> Vec<Runtime> { let mut rts = vec![]; let rt = tokio::runtime::Builder::new_current_thread() .enable_all() .build() .unwrap(); rts.push(rt); let rt = tokio::runtime::Builder::new_multi_thread() .worker_threads(1) .enable_all() .build() .unwrap(); rts.push(rt); let rt = tokio::runtime::Builder::new_multi_thread() .worker_threads(4) .enable_all() .build() .unwrap(); rts.push(rt); #[cfg(tokio_unstable)] { let rt = tokio::runtime::Builder::new_multi_thread() .worker_threads(1) .enable_alt_timer() .enable_all() .build() .unwrap(); rts.push(rt); let rt = tokio::runtime::Builder::new_multi_thread() .worker_threads(4) .enable_alt_timer() .enable_all() .build() .unwrap(); rts.push(rt); } rts } #[test] fn pause_panic_caller() -> Result<(), Box<dyn Error>> { for rt in rt_combinations() { let panic_location_file = test_panic(|| { rt.block_on(async { time::pause(); time::pause(); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); } Ok(()) } #[test] fn resume_panic_caller() -> Result<(), Box<dyn Error>> { let panic_location_file = test_panic(|| { let rt = current_thread(); rt.block_on(async { time::resume(); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] fn interval_panic_caller() -> Result<(), Box<dyn Error>> { let panic_location_file = test_panic(|| { let _ = interval(Duration::from_millis(0)); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] fn interval_at_panic_caller() -> Result<(), Box<dyn Error>> { let panic_location_file = test_panic(|| { let _ = interval_at(Instant::now(), Duration::from_millis(0)); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] fn timeout_panic_caller() -> Result<(), Box<dyn Error>> { let panic_location_file = test_panic(|| { // Runtime without `enable_time` so it has no current timer set. let rt = Builder::new_current_thread().build().unwrap(); rt.block_on(async { let _timeout = timeout(Duration::from_millis(5), future::pending::<()>()); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } fn current_thread() -> Runtime { tokio::runtime::Builder::new_current_thread() .enable_all() .build() .unwrap() }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/time_sleep.rs
tokio/tests/time_sleep.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] #![cfg(not(miri))] // Too slow on Miri. use std::future::Future; use std::task::Context; use futures::task::noop_waker_ref; use tokio::time::{self, Duration, Instant}; use tokio_test::{assert_elapsed, assert_pending, assert_ready, task}; #[tokio::test] async fn immediate_sleep() { time::pause(); let now = Instant::now(); // Ready! time::sleep_until(now).await; assert_elapsed!(now, ms(1)); } #[tokio::test] async fn is_elapsed() { time::pause(); let sleep = time::sleep(Duration::from_millis(10)); tokio::pin!(sleep); assert!(!sleep.is_elapsed()); assert!(futures::poll!(sleep.as_mut()).is_pending()); assert!(!sleep.is_elapsed()); sleep.as_mut().await; assert!(sleep.is_elapsed()); } #[tokio::test] async fn delayed_sleep_level_0() { time::pause(); for &i in &[1, 10, 60] { let now = Instant::now(); let dur = ms(i); time::sleep_until(now + dur).await; assert_elapsed!(now, dur); } } #[tokio::test] async fn sub_ms_delayed_sleep() { time::pause(); for _ in 0..5 { let now = Instant::now(); let deadline = now + ms(1) + Duration::new(0, 1); time::sleep_until(deadline).await; assert_elapsed!(now, ms(1)); } } #[tokio::test] async fn delayed_sleep_wrapping_level_0() { time::pause(); time::sleep(ms(5)).await; let now = Instant::now(); time::sleep_until(now + ms(60)).await; assert_elapsed!(now, ms(60)); } #[tokio::test] async fn reset_future_sleep_before_fire() { time::pause(); let now = Instant::now(); let mut sleep = task::spawn(Box::pin(time::sleep_until(now + ms(100)))); assert_pending!(sleep.poll()); let mut sleep = sleep.into_inner(); sleep.as_mut().reset(Instant::now() + ms(200)); sleep.await; assert_elapsed!(now, ms(200)); } #[tokio::test] async fn reset_past_sleep_before_turn() { time::pause(); let now = Instant::now(); let mut sleep = task::spawn(Box::pin(time::sleep_until(now + ms(100)))); assert_pending!(sleep.poll()); let mut sleep = sleep.into_inner(); sleep.as_mut().reset(now + ms(80)); sleep.await; assert_elapsed!(now, ms(80)); } #[tokio::test] async fn reset_past_sleep_before_fire() { time::pause(); let now = Instant::now(); let mut sleep = task::spawn(Box::pin(time::sleep_until(now + ms(100)))); assert_pending!(sleep.poll()); let mut sleep = sleep.into_inner(); time::sleep(ms(10)).await; sleep.as_mut().reset(now + ms(80)); sleep.await; assert_elapsed!(now, ms(80)); } #[tokio::test] async fn reset_future_sleep_after_fire() { time::pause(); let now = Instant::now(); let mut sleep = Box::pin(time::sleep_until(now + ms(100))); sleep.as_mut().await; assert_elapsed!(now, ms(100)); sleep.as_mut().reset(now + ms(110)); sleep.await; assert_elapsed!(now, ms(110)); } #[tokio::test] async fn reset_sleep_to_past() { time::pause(); let now = Instant::now(); let mut sleep = task::spawn(Box::pin(time::sleep_until(now + ms(100)))); assert_pending!(sleep.poll()); time::sleep(ms(50)).await; assert!(!sleep.is_woken()); sleep.as_mut().reset(now + ms(40)); // TODO: is this required? //assert!(sleep.is_woken()); assert_ready!(sleep.poll()); } #[cfg(not(target_os = "wasi"))] // Wasi doesn't support panic recovery #[test] #[should_panic] fn creating_sleep_outside_of_context() { let now = Instant::now(); // This creates a delay outside of the context of a mock timer. This tests // that it will panic. let _fut = time::sleep_until(now + ms(500)); } #[tokio::test] async fn greater_than_max() { const YR_5: u64 = 5 * 365 * 24 * 60 * 60 * 1000; time::pause(); time::sleep_until(Instant::now() + ms(YR_5)).await; } #[tokio::test] async fn short_sleeps() { for _ in 0..1000 { tokio::time::sleep(std::time::Duration::from_millis(0)).await; } } #[tokio::test] async fn multi_long_sleeps() { tokio::time::pause(); for _ in 0..5u32 { tokio::time::sleep(Duration::from_secs( // about a year 365 * 24 * 3600, )) .await; } let deadline = tokio::time::Instant::now() + Duration::from_secs( // about 10 years 10 * 365 * 24 * 3600, ); tokio::time::sleep_until(deadline).await; assert!(tokio::time::Instant::now() >= deadline); } #[tokio::test] async fn long_sleeps() { tokio::time::pause(); let deadline = tokio::time::Instant::now() + Duration::from_secs( // about 10 years 10 * 365 * 24 * 3600, ); tokio::time::sleep_until(deadline).await; assert!(tokio::time::Instant::now() >= deadline); assert!(tokio::time::Instant::now() <= deadline + Duration::from_millis(1)); } #[tokio::test] async fn reset_after_firing() { let timer = tokio::time::sleep(std::time::Duration::from_millis(1)); tokio::pin!(timer); let deadline = timer.deadline(); timer.as_mut().await; assert_ready!(timer .as_mut() .poll(&mut Context::from_waker(noop_waker_ref()))); timer .as_mut() .reset(tokio::time::Instant::now() + std::time::Duration::from_secs(600)); assert_ne!(deadline, timer.deadline()); assert_pending!(timer .as_mut() .poll(&mut Context::from_waker(noop_waker_ref()))); assert_pending!(timer .as_mut() .poll(&mut Context::from_waker(noop_waker_ref()))); } #[tokio::test] async fn exactly_max() { time::pause(); time::sleep(Duration::MAX).await; } #[tokio::test] async fn issue_5183() { time::pause(); let big = std::time::Duration::from_secs(u64::MAX / 10); // This is a workaround since awaiting sleep(big) will never finish. #[rustfmt::skip] tokio::select! { biased; _ = tokio::time::sleep(big) => {} _ = tokio::time::sleep(std::time::Duration::from_nanos(1)) => {} } } #[tokio::test] async fn no_out_of_bounds_close_to_max() { time::pause(); time::sleep(Duration::MAX - Duration::from_millis(1)).await; } fn ms(n: u64) -> Duration { Duration::from_millis(n) } #[tokio::test] async fn drop_after_reschedule_at_new_scheduled_time() { use futures::poll; tokio::time::pause(); let start = tokio::time::Instant::now(); let mut a = Box::pin(tokio::time::sleep(Duration::from_millis(5))); let mut b = Box::pin(tokio::time::sleep(Duration::from_millis(5))); let mut c = Box::pin(tokio::time::sleep(Duration::from_millis(10))); let _ = poll!(&mut a); let _ = poll!(&mut b); let _ = poll!(&mut c); b.as_mut().reset(start + Duration::from_millis(10)); a.await; drop(b); } #[tokio::test] async fn drop_from_wake() { use std::future::Future; use std::pin::Pin; use std::sync::atomic::{AtomicBool, Ordering}; use std::sync::{Arc, Mutex}; use std::task::Context; let panicked = Arc::new(AtomicBool::new(false)); let list: Arc<Mutex<Vec<Pin<Box<tokio::time::Sleep>>>>> = Arc::new(Mutex::new(Vec::new())); let arc_wake = Arc::new(DropWaker(panicked.clone(), list.clone())); let arc_wake = futures::task::waker(arc_wake); tokio::time::pause(); { let mut lock = list.lock().unwrap(); for _ in 0..100 { let mut timer = Box::pin(tokio::time::sleep(Duration::from_millis(10))); let _ = timer.as_mut().poll(&mut Context::from_waker(&arc_wake)); lock.push(timer); } } tokio::time::sleep(Duration::from_millis(11)).await; assert!( !panicked.load(Ordering::SeqCst), "panicked when dropping timers" ); #[derive(Clone)] struct DropWaker( Arc<AtomicBool>, Arc<Mutex<Vec<Pin<Box<tokio::time::Sleep>>>>>, ); impl futures::task::ArcWake for DropWaker { fn wake_by_ref(arc_self: &Arc<Self>) { let result = std::panic::catch_unwind(std::panic::AssertUnwindSafe(|| { *arc_self.1.lock().expect("panic in lock") = Vec::new() })); if result.is_err() { arc_self.0.store(true, Ordering::SeqCst); } } } }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/fs_uring.rs
tokio/tests/fs_uring.rs
//! Uring file operations tests. #![cfg(all( tokio_unstable, feature = "io-uring", feature = "rt", feature = "fs", target_os = "linux" ))] use futures::future::FutureExt; use std::sync::mpsc; use std::task::Poll; use std::time::Duration; use std::{future::poll_fn, path::PathBuf}; use tempfile::NamedTempFile; use tokio::{ fs::OpenOptions, runtime::{Builder, Runtime}, }; use tokio_util::task::TaskTracker; fn multi_rt(n: usize) -> Box<dyn Fn() -> Runtime> { Box::new(move || { Builder::new_multi_thread() .worker_threads(n) .enable_all() .build() .unwrap() }) } fn current_rt() -> Box<dyn Fn() -> Runtime> { Box::new(|| Builder::new_current_thread().enable_all().build().unwrap()) } fn rt_combinations() -> Vec<Box<dyn Fn() -> Runtime>> { vec![ current_rt(), multi_rt(1), multi_rt(2), multi_rt(8), multi_rt(64), multi_rt(256), ] } #[test] fn shutdown_runtime_while_performing_io_uring_ops() { fn run(rt: Runtime) { let (tx, rx) = mpsc::channel(); let (done_tx, done_rx) = mpsc::channel(); let (_tmp, path) = create_tmp_files(1); rt.spawn(async move { let path = path[0].clone(); // spawning a bunch of uring operations. loop { let path = path.clone(); tokio::spawn(async move { let mut opt = OpenOptions::new(); opt.read(true); opt.open(&path).await.unwrap(); }); // Avoid busy looping. tokio::task::yield_now().await; } }); std::thread::spawn(move || { let rt: Runtime = rx.recv().unwrap(); rt.shutdown_timeout(Duration::from_millis(300)); done_tx.send(()).unwrap(); }); tx.send(rt).unwrap(); done_rx.recv().unwrap(); } for rt in rt_combinations() { run(rt()); } } #[test] fn open_many_files() { fn run(rt: Runtime) { const NUM_FILES: usize = 512; let (_tmp_files, paths): (Vec<NamedTempFile>, Vec<PathBuf>) = create_tmp_files(NUM_FILES); rt.block_on(async move { let tracker = TaskTracker::new(); for i in 0..10_000 { let path = paths.get(i % NUM_FILES).unwrap().clone(); tracker.spawn(async move { let _file = OpenOptions::new().read(true).open(path).await.unwrap(); }); } tracker.close(); tracker.wait().await; }); } for rt in rt_combinations() { run(rt()); } } #[tokio::test] async fn cancel_op_future() { let (_tmp_file, path): (Vec<NamedTempFile>, Vec<PathBuf>) = create_tmp_files(1); let (tx, mut rx) = tokio::sync::mpsc::unbounded_channel(); let handle = tokio::spawn(async move { poll_fn(|cx| { let opt = { let mut opt = tokio::fs::OpenOptions::new(); opt.read(true); opt }; let fut = opt.open(&path[0]); // If io_uring is enabled (and not falling back to the thread pool), // the first poll should return Pending. let _pending = Box::pin(fut).poll_unpin(cx); tx.send(()).unwrap(); Poll::<()>::Pending }) .await; }); // Wait for the first poll rx.recv().await.unwrap(); handle.abort(); let res = handle.await.unwrap_err(); assert!(res.is_cancelled()); } fn create_tmp_files(num_files: usize) -> (Vec<NamedTempFile>, Vec<PathBuf>) { let mut files = Vec::with_capacity(num_files); for _ in 0..num_files { let tmp = NamedTempFile::new().unwrap(); let path = tmp.path().to_path_buf(); files.push((tmp, path)); } files.into_iter().unzip() }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/rt_unstable_metrics.rs
tokio/tests/rt_unstable_metrics.rs
#![warn(rust_2018_idioms)] #![cfg(all( feature = "full", tokio_unstable, not(target_os = "wasi"), target_has_atomic = "64" ))] use std::future::Future; use std::sync::{mpsc, Arc, Mutex}; use std::task::Poll; use std::thread; use tokio::macros::support::poll_fn; use tokio::runtime::{HistogramConfiguration, HistogramScale, LogHistogram, Runtime}; use tokio::task::consume_budget; use tokio::time::{self, Duration}; #[test] fn num_workers() { let rt = current_thread(); assert_eq!(1, rt.metrics().num_workers()); let rt = threaded(); assert_eq!(2, rt.metrics().num_workers()); } #[test] fn num_blocking_threads() { let rt = current_thread(); assert_eq!(0, rt.metrics().num_blocking_threads()); let _ = rt.block_on(rt.spawn_blocking(move || {})); assert_eq!(1, rt.metrics().num_blocking_threads()); let rt = threaded(); assert_eq!(0, rt.metrics().num_blocking_threads()); let _ = rt.block_on(rt.spawn_blocking(move || {})); assert_eq!(1, rt.metrics().num_blocking_threads()); } #[test] fn num_idle_blocking_threads() { let rt = current_thread(); assert_eq!(0, rt.metrics().num_idle_blocking_threads()); let _ = rt.block_on(rt.spawn_blocking(move || {})); rt.block_on(async { time::sleep(Duration::from_millis(5)).await; }); // We need to wait until the blocking thread has become idle. Usually 5ms is // enough for this to happen, but not always. When it isn't enough, sleep // for another second. We don't always wait for a whole second since we want // the test suite to finish quickly. // // Note that the timeout for idle threads to be killed is 10 seconds. if 0 == rt.metrics().num_idle_blocking_threads() { rt.block_on(async { time::sleep(Duration::from_secs(1)).await; }); } assert_eq!(1, rt.metrics().num_idle_blocking_threads()); } #[test] fn blocking_queue_depth() { let rt = tokio::runtime::Builder::new_current_thread() .enable_all() .max_blocking_threads(1) .build() .unwrap(); assert_eq!(0, rt.metrics().blocking_queue_depth()); let ready = Arc::new(Mutex::new(())); let guard = ready.lock().unwrap(); let ready_cloned = ready.clone(); let wait_until_ready = move || { let _unused = ready_cloned.lock().unwrap(); }; let h1 = rt.spawn_blocking(wait_until_ready.clone()); let h2 = rt.spawn_blocking(wait_until_ready); assert!(rt.metrics().blocking_queue_depth() > 0); drop(guard); let _ = rt.block_on(h1); let _ = rt.block_on(h2); assert_eq!(0, rt.metrics().blocking_queue_depth()); } #[test] fn spawned_tasks_count() { let rt = current_thread(); let metrics = rt.metrics(); assert_eq!(0, metrics.spawned_tasks_count()); rt.block_on(rt.spawn(async move { assert_eq!(1, metrics.spawned_tasks_count()); })) .unwrap(); assert_eq!(1, rt.metrics().spawned_tasks_count()); let rt = threaded(); let metrics = rt.metrics(); assert_eq!(0, metrics.spawned_tasks_count()); rt.block_on(rt.spawn(async move { assert_eq!(1, metrics.spawned_tasks_count()); })) .unwrap(); assert_eq!(1, rt.metrics().spawned_tasks_count()); } #[test] fn remote_schedule_count() { use std::thread; let rt = current_thread(); let handle = rt.handle().clone(); let task = thread::spawn(move || { handle.spawn(async { // DO nothing }) }) .join() .unwrap(); rt.block_on(task).unwrap(); assert_eq!(1, rt.metrics().remote_schedule_count()); let rt = threaded(); let handle = rt.handle().clone(); let task = thread::spawn(move || { handle.spawn(async { // DO nothing }) }) .join() .unwrap(); rt.block_on(task).unwrap(); assert_eq!(1, rt.metrics().remote_schedule_count()); } #[test] fn worker_thread_id_current_thread() { let rt = current_thread(); let metrics = rt.metrics(); // Check that runtime is on this thread. rt.block_on(async {}); assert_eq!(Some(thread::current().id()), metrics.worker_thread_id(0)); // Move runtime to another thread. let thread_id = std::thread::scope(|scope| { let join_handle = scope.spawn(|| { rt.block_on(async {}); }); join_handle.thread().id() }); assert_eq!(Some(thread_id), metrics.worker_thread_id(0)); // Move runtime back to this thread. rt.block_on(async {}); assert_eq!(Some(thread::current().id()), metrics.worker_thread_id(0)); } #[test] fn worker_thread_id_threaded() { let rt = threaded(); let metrics = rt.metrics(); rt.block_on(rt.spawn(async move { // Check that we are running on a worker thread and determine // the index of our worker. let thread_id = std::thread::current().id(); let this_worker = (0..2) .position(|w| metrics.worker_thread_id(w) == Some(thread_id)) .expect("task not running on any worker thread"); // Force worker to another thread. let moved_thread_id = tokio::task::block_in_place(|| { assert_eq!(thread_id, std::thread::current().id()); // Wait for worker to move to another thread. for _ in 0..100 { let new_id = metrics.worker_thread_id(this_worker).unwrap(); if thread_id != new_id { return new_id; } std::thread::sleep(Duration::from_millis(100)); } panic!("worker did not move to new thread"); }); // After blocking task worker either stays on new thread or // is moved back to current thread. assert!( metrics.worker_thread_id(this_worker) == Some(moved_thread_id) || metrics.worker_thread_id(this_worker) == Some(thread_id) ); })) .unwrap() } #[test] fn worker_noop_count() { // There isn't really a great way to generate no-op parks as they happen as // false-positive events under concurrency. let rt = current_thread(); let metrics = rt.metrics(); rt.block_on(async { time::sleep(Duration::from_millis(1)).await; }); drop(rt); assert!(0 < metrics.worker_noop_count(0)); let rt = threaded(); let metrics = rt.metrics(); rt.block_on(async { time::sleep(Duration::from_millis(1)).await; }); drop(rt); assert!(0 < metrics.worker_noop_count(0)); assert!(0 < metrics.worker_noop_count(1)); } #[test] fn worker_steal_count() { // This metric only applies to the multi-threaded runtime. for _ in 0..10 { let rt = threaded_no_lifo(); let metrics = rt.metrics(); let successfully_spawned_stealable_task = rt.block_on(async { // The call to `try_spawn_stealable_task` may time out, which means // that the sending task couldn't be scheduled due to a deadlock in // the runtime. // This is expected behaviour, we just retry until we succeed or // exhaust all tries, the latter causing this test to fail. try_spawn_stealable_task().await.is_ok() }); drop(rt); if successfully_spawned_stealable_task { let n: u64 = (0..metrics.num_workers()) .map(|i| metrics.worker_steal_count(i)) .sum(); assert_eq!(1, n); return; } } panic!("exhausted every try to schedule the stealable task"); } #[test] fn worker_poll_count_and_time() { const N: u64 = 5; async fn task() { // Sync sleep std::thread::sleep(std::time::Duration::from_micros(10)); } let rt = current_thread(); let metrics = rt.metrics(); rt.block_on(async { for _ in 0..N { tokio::spawn(task()).await.unwrap(); } }); drop(rt); assert_eq!(N, metrics.worker_poll_count(0)); // Not currently supported for current-thread runtime assert_eq!(Duration::default(), metrics.worker_mean_poll_time(0)); // Does not populate the histogram assert!(!metrics.poll_time_histogram_enabled()); for i in 0..10 { assert_eq!(0, metrics.poll_time_histogram_bucket_count(0, i)); } let rt = threaded(); let metrics = rt.metrics(); rt.block_on(async { for _ in 0..N { tokio::spawn(task()).await.unwrap(); } }); drop(rt); // Account for the `block_on` task let n = (0..metrics.num_workers()) .map(|i| metrics.worker_poll_count(i)) .sum(); assert_eq!(N, n); let n: Duration = (0..metrics.num_workers()) .map(|i| metrics.worker_mean_poll_time(i)) .sum(); assert!(n > Duration::default()); // Does not populate the histogram assert!(!metrics.poll_time_histogram_enabled()); for n in 0..metrics.num_workers() { for i in 0..10 { assert_eq!(0, metrics.poll_time_histogram_bucket_count(n, i)); } } } #[test] fn log_histogram() { const N: u64 = 50; let rt = tokio::runtime::Builder::new_current_thread() .enable_all() .enable_metrics_poll_time_histogram() .metrics_poll_time_histogram_configuration(HistogramConfiguration::log( LogHistogram::builder() .max_value(Duration::from_secs(60)) .min_value(Duration::from_nanos(100)) .max_error(0.25), )) .build() .unwrap(); let metrics = rt.metrics(); let num_buckets = rt.metrics().poll_time_histogram_num_buckets(); assert_eq!(num_buckets, 119); rt.block_on(async { for _ in 0..N { tokio::spawn(async {}).await.unwrap(); } }); drop(rt); assert_eq!( metrics.poll_time_histogram_bucket_range(0), Duration::from_nanos(0)..Duration::from_nanos(96) ); assert_eq!( metrics.poll_time_histogram_bucket_range(1), Duration::from_nanos(96)..Duration::from_nanos(96 + 2_u64.pow(4)) ); assert_eq!( metrics.poll_time_histogram_bucket_range(118).end, Duration::from_nanos(u64::MAX) ); let n = (0..metrics.num_workers()) .flat_map(|i| (0..num_buckets).map(move |j| (i, j))) .map(|(worker, bucket)| metrics.poll_time_histogram_bucket_count(worker, bucket)) .sum(); assert_eq!(N, n); } #[test] fn minimal_log_histogram() { let rt = tokio::runtime::Builder::new_current_thread() .enable_all() .enable_metrics_poll_time_histogram() .metrics_poll_time_histogram_configuration(HistogramConfiguration::log( LogHistogram::builder() .max_value(Duration::from_millis(4)) .min_value(Duration::from_micros(20)) .precision_exact(0), )) .build() .unwrap(); let metrics = rt.metrics(); let num_buckets = rt.metrics().poll_time_histogram_num_buckets(); for b in 1..num_buckets - 1 { let range = metrics.poll_time_histogram_bucket_range(b); let size = range.end - range.start; // Assert the buckets continue doubling in size assert_eq!( size, Duration::from_nanos((1 << (b - 1)) * 16384), "incorrect range for {b}" ); } assert_eq!(num_buckets, 10); } #[test] #[allow(deprecated)] fn legacy_log_histogram() { let rt = tokio::runtime::Builder::new_multi_thread() .enable_all() .enable_metrics_poll_time_histogram() .metrics_poll_count_histogram_scale(HistogramScale::Log) .metrics_poll_count_histogram_resolution(Duration::from_micros(50)) .metrics_poll_count_histogram_buckets(20) .build() .unwrap(); let num_buckets = rt.metrics().poll_time_histogram_num_buckets(); assert_eq!(num_buckets, 20); } #[test] fn log_histogram_default_configuration() { let rt = tokio::runtime::Builder::new_current_thread() .enable_all() .enable_metrics_poll_time_histogram() .metrics_poll_time_histogram_configuration(HistogramConfiguration::log( LogHistogram::default(), )) .build() .unwrap(); let num_buckets = rt.metrics().poll_time_histogram_num_buckets(); assert_eq!(num_buckets, 119); } #[test] fn worker_poll_count_histogram() { const N: u64 = 5; let rts = [ tokio::runtime::Builder::new_current_thread() .enable_all() .enable_metrics_poll_time_histogram() .metrics_poll_time_histogram_configuration(HistogramConfiguration::linear( Duration::from_millis(50), 3, )) .build() .unwrap(), tokio::runtime::Builder::new_multi_thread() .worker_threads(2) .enable_all() .enable_metrics_poll_time_histogram() .metrics_poll_time_histogram_configuration(HistogramConfiguration::linear( Duration::from_millis(50), 3, )) .build() .unwrap(), ]; for rt in rts { let metrics = rt.metrics(); rt.block_on(async { for _ in 0..N { tokio::spawn(async {}).await.unwrap(); } }); drop(rt); let num_workers = metrics.num_workers(); let num_buckets = metrics.poll_time_histogram_num_buckets(); assert!(metrics.poll_time_histogram_enabled()); assert_eq!(num_buckets, 3); let n = (0..num_workers) .flat_map(|i| (0..num_buckets).map(move |j| (i, j))) .map(|(worker, bucket)| metrics.poll_time_histogram_bucket_count(worker, bucket)) .sum(); assert_eq!(N, n); } } #[test] fn worker_poll_count_histogram_range() { let max = Duration::from_nanos(u64::MAX); let rt = tokio::runtime::Builder::new_current_thread() .enable_all() .enable_metrics_poll_time_histogram() .metrics_poll_time_histogram_configuration(HistogramConfiguration::linear(us(50), 3)) .build() .unwrap(); let metrics = rt.metrics(); assert_eq!(metrics.poll_time_histogram_bucket_range(0), us(0)..us(50)); assert_eq!(metrics.poll_time_histogram_bucket_range(1), us(50)..us(100)); assert_eq!(metrics.poll_time_histogram_bucket_range(2), us(100)..max); // ensure the old methods work too #[allow(deprecated)] let rt = tokio::runtime::Builder::new_current_thread() .enable_all() .enable_metrics_poll_time_histogram() .metrics_poll_count_histogram_scale(tokio::runtime::HistogramScale::Log) .metrics_poll_count_histogram_buckets(3) .metrics_poll_count_histogram_resolution(us(50)) .build() .unwrap(); let metrics = rt.metrics(); let a = Duration::from_nanos(50000_u64.next_power_of_two()); let b = a * 2; assert_eq!(metrics.poll_time_histogram_bucket_range(0), us(0)..a); assert_eq!(metrics.poll_time_histogram_bucket_range(1), a..b); assert_eq!(metrics.poll_time_histogram_bucket_range(2), b..max); } #[test] fn worker_poll_count_histogram_disabled_without_explicit_enable() { let rts = [ tokio::runtime::Builder::new_current_thread() .enable_all() .metrics_poll_time_histogram_configuration(HistogramConfiguration::linear( Duration::from_millis(50), 3, )) .build() .unwrap(), tokio::runtime::Builder::new_multi_thread() .worker_threads(2) .enable_all() .metrics_poll_time_histogram_configuration(HistogramConfiguration::linear( Duration::from_millis(50), 3, )) .build() .unwrap(), ]; for rt in rts { let metrics = rt.metrics(); assert!(!metrics.poll_time_histogram_enabled()); } } #[test] fn worker_local_schedule_count() { let rt = current_thread(); let metrics = rt.metrics(); rt.block_on(async { tokio::spawn(async {}).await.unwrap(); }); drop(rt); assert_eq!(1, metrics.worker_local_schedule_count(0)); assert_eq!(0, metrics.remote_schedule_count()); let rt = threaded(); let metrics = rt.metrics(); rt.block_on(async { // Move to the runtime tokio::spawn(async { tokio::spawn(async {}).await.unwrap(); }) .await .unwrap(); }); drop(rt); let n: u64 = (0..metrics.num_workers()) .map(|i| metrics.worker_local_schedule_count(i)) .sum(); assert_eq!(2, n); assert_eq!(1, metrics.remote_schedule_count()); } #[test] fn worker_overflow_count() { // Only applies to the threaded worker let rt = threaded(); let metrics = rt.metrics(); rt.block_on(async { // Move to the runtime tokio::spawn(async { let (tx1, rx1) = std::sync::mpsc::channel(); let (tx2, rx2) = std::sync::mpsc::channel(); // First, we need to block the other worker until all tasks have // been spawned. // // We spawn from outside the runtime to ensure that the other worker // will pick it up: // <https://github.com/tokio-rs/tokio/issues/4730> tokio::task::spawn_blocking(|| { tokio::spawn(async move { tx1.send(()).unwrap(); rx2.recv().unwrap(); }); }); rx1.recv().unwrap(); // Spawn many tasks for _ in 0..300 { tokio::spawn(async {}); } tx2.send(()).unwrap(); }) .await .unwrap(); }); drop(rt); let n: u64 = (0..metrics.num_workers()) .map(|i| metrics.worker_overflow_count(i)) .sum(); assert_eq!(1, n); } #[test] fn worker_local_queue_depth() { const N: usize = 100; let rt = current_thread(); let metrics = rt.metrics(); rt.block_on(async { for _ in 0..N { tokio::spawn(async {}); } assert_eq!(N, metrics.worker_local_queue_depth(0)); }); let rt = threaded(); let metrics = rt.metrics(); rt.block_on(async move { // Move to the runtime tokio::spawn(async move { let (tx1, rx1) = std::sync::mpsc::channel(); let (tx2, rx2) = std::sync::mpsc::channel(); // First, we need to block the other worker until all tasks have // been spawned. tokio::spawn(async move { tx1.send(()).unwrap(); rx2.recv().unwrap(); }); // Bump the next-run spawn tokio::spawn(async {}); rx1.recv().unwrap(); // Spawn some tasks for _ in 0..100 { tokio::spawn(async {}); } let n: usize = (0..metrics.num_workers()) .map(|i| metrics.worker_local_queue_depth(i)) .sum(); assert_eq!(n, N); tx2.send(()).unwrap(); }) .await .unwrap(); }); } #[test] fn budget_exhaustion_yield() { let rt = current_thread(); let metrics = rt.metrics(); assert_eq!(0, metrics.budget_forced_yield_count()); let mut did_yield = false; // block on a task which consumes budget until it yields rt.block_on(poll_fn(|cx| loop { if did_yield { return Poll::Ready(()); } let fut = consume_budget(); tokio::pin!(fut); if fut.poll(cx).is_pending() { did_yield = true; return Poll::Pending; } })); assert_eq!(1, rt.metrics().budget_forced_yield_count()); } #[test] fn budget_exhaustion_yield_with_joins() { let rt = current_thread(); let metrics = rt.metrics(); assert_eq!(0, metrics.budget_forced_yield_count()); let mut did_yield_1 = false; let mut did_yield_2 = false; // block on a task which consumes budget until it yields rt.block_on(async { tokio::join!( poll_fn(|cx| loop { if did_yield_1 { return Poll::Ready(()); } let fut = consume_budget(); tokio::pin!(fut); if fut.poll(cx).is_pending() { did_yield_1 = true; return Poll::Pending; } }), poll_fn(|cx| loop { if did_yield_2 { return Poll::Ready(()); } let fut = consume_budget(); tokio::pin!(fut); if fut.poll(cx).is_pending() { did_yield_2 = true; return Poll::Pending; } }) ) }); assert_eq!(1, rt.metrics().budget_forced_yield_count()); } #[cfg(any(target_os = "linux", target_os = "macos"))] #[test] fn io_driver_fd_count() { let rt = current_thread(); let metrics = rt.metrics(); assert_eq!(metrics.io_driver_fd_registered_count(), 0); let stream = tokio::net::TcpStream::connect("google.com:80"); let stream = rt.block_on(async move { stream.await.unwrap() }); assert_eq!(metrics.io_driver_fd_registered_count(), 1); assert_eq!(metrics.io_driver_fd_deregistered_count(), 0); drop(stream); assert_eq!(metrics.io_driver_fd_deregistered_count(), 1); assert_eq!(metrics.io_driver_fd_registered_count(), 1); } #[cfg(any(target_os = "linux", target_os = "macos"))] #[test] fn io_driver_ready_count() { let rt = current_thread(); let metrics = rt.metrics(); let stream = tokio::net::TcpStream::connect("google.com:80"); let _stream = rt.block_on(async move { stream.await.unwrap() }); assert_eq!(metrics.io_driver_ready_count(), 1); } async fn try_spawn_stealable_task() -> Result<(), mpsc::RecvTimeoutError> { // We use a blocking channel to synchronize the tasks. let (tx, rx) = mpsc::channel(); // Make sure we are in the context of the runtime. tokio::spawn(async move { // Spawn the task that sends to the channel. // // Note that the runtime needs to have the lifo slot disabled to make // this task stealable. tokio::spawn(async move { tx.send(()).unwrap(); }); // Blocking receive on the channel, timing out if the sending task // wasn't scheduled in time. rx.recv_timeout(Duration::from_secs(1)) }) .await .unwrap()?; Ok(()) } fn current_thread() -> Runtime { tokio::runtime::Builder::new_current_thread() .enable_all() .build() .unwrap() } fn threaded() -> Runtime { tokio::runtime::Builder::new_multi_thread() .worker_threads(2) .enable_all() .build() .unwrap() } fn threaded_no_lifo() -> Runtime { tokio::runtime::Builder::new_multi_thread() .worker_threads(2) .disable_lifo_slot() .enable_all() .build() .unwrap() } fn us(n: u64) -> Duration { Duration::from_micros(n) }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_copy_bidirectional.rs
tokio/tests/io_copy_bidirectional.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // Wasi does not support bind() use std::time::Duration; use tokio::io::{self, copy_bidirectional, AsyncReadExt, AsyncWriteExt}; use tokio::net::TcpStream; use tokio::task::JoinHandle; async fn make_socketpair() -> (TcpStream, TcpStream) { let listener = tokio::net::TcpListener::bind("127.0.0.1:0").await.unwrap(); let addr = listener.local_addr().unwrap(); let connector = TcpStream::connect(addr); let acceptor = listener.accept(); let (c1, c2) = tokio::join!(connector, acceptor); (c1.unwrap(), c2.unwrap().0) } async fn block_write(s: &mut TcpStream) -> usize { static BUF: [u8; 2048] = [0; 2048]; let mut copied = 0; loop { tokio::select! { result = s.write(&BUF) => { copied += result.expect("write error") }, _ = tokio::time::sleep(Duration::from_millis(10)) => { break; } } } copied } async fn symmetric<F, Fut>(mut cb: F) where F: FnMut(JoinHandle<io::Result<(u64, u64)>>, TcpStream, TcpStream) -> Fut, Fut: std::future::Future<Output = ()>, { // We run the test twice, with streams passed to copy_bidirectional in // different orders, in order to ensure that the two arguments are // interchangeable. let (a, mut a1) = make_socketpair().await; let (b, mut b1) = make_socketpair().await; let handle = tokio::spawn(async move { copy_bidirectional(&mut a1, &mut b1).await }); cb(handle, a, b).await; let (a, mut a1) = make_socketpair().await; let (b, mut b1) = make_socketpair().await; let handle = tokio::spawn(async move { copy_bidirectional(&mut b1, &mut a1).await }); cb(handle, b, a).await; } #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` in miri. async fn test_basic_transfer() { symmetric(|_handle, mut a, mut b| async move { a.write_all(b"test").await.unwrap(); let mut tmp = [0; 4]; b.read_exact(&mut tmp).await.unwrap(); assert_eq!(&tmp[..], b"test"); }) .await } #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` in miri. async fn test_transfer_after_close() { symmetric(|handle, mut a, mut b| async move { AsyncWriteExt::shutdown(&mut a).await.unwrap(); b.read_to_end(&mut Vec::new()).await.unwrap(); b.write_all(b"quux").await.unwrap(); let mut tmp = [0; 4]; a.read_exact(&mut tmp).await.unwrap(); assert_eq!(&tmp[..], b"quux"); // Once both are closed, we should have our handle back drop(b); assert_eq!(handle.await.unwrap().unwrap(), (0, 4)); }) .await } #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` in miri. async fn blocking_one_side_does_not_block_other() { symmetric(|handle, mut a, mut b| async move { block_write(&mut a).await; b.write_all(b"quux").await.unwrap(); let mut tmp = [0; 4]; a.read_exact(&mut tmp).await.unwrap(); assert_eq!(&tmp[..], b"quux"); AsyncWriteExt::shutdown(&mut a).await.unwrap(); let mut buf = Vec::new(); b.read_to_end(&mut buf).await.unwrap(); drop(b); assert_eq!(handle.await.unwrap().unwrap(), (buf.len() as u64, 4)); }) .await } #[tokio::test] async fn immediate_exit_on_write_error() { let payload = b"here, take this"; let error = || io::Error::new(io::ErrorKind::Other, "no thanks!"); let mut a = tokio_test::io::Builder::new() .read(payload) .write_error(error()) .build(); let mut b = tokio_test::io::Builder::new() .read(payload) .write_error(error()) .build(); assert!(copy_bidirectional(&mut a, &mut b).await.is_err()); } #[tokio::test] async fn immediate_exit_on_read_error() { let error = || io::Error::new(io::ErrorKind::Other, "got nothing!"); let mut a = tokio_test::io::Builder::new().read_error(error()).build(); let mut b = tokio_test::io::Builder::new().read_error(error()).build(); assert!(copy_bidirectional(&mut a, &mut b).await.is_err()); } #[tokio::test] async fn copy_bidirectional_is_cooperative() { tokio::select! { biased; _ = async { loop { let payload = b"here, take this"; let mut a = tokio_test::io::Builder::new() .read(payload) .write(payload) .build(); let mut b = tokio_test::io::Builder::new() .read(payload) .write(payload) .build(); let _ = copy_bidirectional(&mut a, &mut b).await; } } => {}, _ = tokio::task::yield_now() => {} } }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/task_yield_now.rs
tokio/tests/task_yield_now.rs
#![cfg(all(feature = "full", not(target_os = "wasi"), tokio_unstable))] use tokio::task; use tokio_test::task::spawn; // `yield_now` is tested within the runtime in `rt_common`. #[test] fn yield_now_outside_of_runtime() { let mut task = spawn(async { task::yield_now().await; }); assert!(task.poll().is_pending()); assert!(task.is_woken()); assert!(task.poll().is_ready()); } #[tokio::test(flavor = "multi_thread")] async fn yield_now_external_executor_and_block_in_place() { let j = tokio::spawn(async { task::block_in_place(|| futures::executor::block_on(task::yield_now())); }); j.await.unwrap(); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/process_raw_handle.rs
tokio/tests/process_raw_handle.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] #![cfg(windows)] #![cfg(not(miri))] use tokio::process::Command; use windows_sys::Win32::System::Threading::GetProcessId; #[tokio::test] async fn obtain_raw_handle() { let mut cmd = Command::new("cmd"); cmd.kill_on_drop(true); cmd.arg("/c"); cmd.arg("pause"); let child = cmd.spawn().unwrap(); let orig_id = child.id().expect("missing id"); assert!(orig_id > 0); let handle = child.raw_handle().expect("process stopped"); let handled_id = unsafe { GetProcessId(handle as _) }; assert_eq!(handled_id, orig_id); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/process_change_of_runtime.rs
tokio/tests/process_change_of_runtime.rs
#![cfg(feature = "process")] #![warn(rust_2018_idioms)] // This tests test the behavior of `process::Command::spawn` when it is used // outside runtime, and when `process::Child::wait ` is used in a different // runtime from which `process::Command::spawn` is used. #![cfg(all(unix, not(target_os = "freebsd"), not(miri)))] // Miri cannot run system commands use std::process::Stdio; use tokio::{process::Command, runtime::Runtime}; #[test] fn process_spawned_and_wait_in_different_runtime() { let mut child = Runtime::new().unwrap().block_on(async { Command::new("true") .stdin(Stdio::piped()) .stdout(Stdio::null()) .spawn() .unwrap() }); Runtime::new().unwrap().block_on(async { let _ = child.wait().await.unwrap(); }); } #[test] #[should_panic( expected = "there is no reactor running, must be called from the context of a Tokio 1.x runtime" )] fn process_spawned_outside_runtime() { let _ = Command::new("true") .stdin(Stdio::piped()) .stdout(Stdio::null()) .spawn(); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_util_empty.rs
tokio/tests/io_util_empty.rs
#![cfg(feature = "full")] use tokio::io::{AsyncBufReadExt, AsyncReadExt, AsyncSeekExt}; #[tokio::test] async fn empty_read_is_cooperative() { tokio::select! { biased; _ = async { loop { let mut buf = [0u8; 4096]; let _ = tokio::io::empty().read(&mut buf).await; } } => {}, _ = tokio::task::yield_now() => {} } } #[tokio::test] async fn empty_buf_reads_are_cooperative() { tokio::select! { biased; _ = async { loop { let mut buf = String::new(); let _ = tokio::io::empty().read_line(&mut buf).await; } } => {}, _ = tokio::task::yield_now() => {} } } #[tokio::test] async fn empty_seek() { use std::io::SeekFrom; let mut empty = tokio::io::empty(); assert!(matches!(empty.seek(SeekFrom::Start(0)).await, Ok(0))); assert!(matches!(empty.seek(SeekFrom::Start(1)).await, Ok(0))); assert!(matches!(empty.seek(SeekFrom::Start(u64::MAX)).await, Ok(0))); assert!(matches!(empty.seek(SeekFrom::End(i64::MIN)).await, Ok(0))); assert!(matches!(empty.seek(SeekFrom::End(-1)).await, Ok(0))); assert!(matches!(empty.seek(SeekFrom::End(0)).await, Ok(0))); assert!(matches!(empty.seek(SeekFrom::End(1)).await, Ok(0))); assert!(matches!(empty.seek(SeekFrom::End(i64::MAX)).await, Ok(0))); assert!(matches!( empty.seek(SeekFrom::Current(i64::MIN)).await, Ok(0) )); assert!(matches!(empty.seek(SeekFrom::Current(-1)).await, Ok(0))); assert!(matches!(empty.seek(SeekFrom::Current(0)).await, Ok(0))); assert!(matches!(empty.seek(SeekFrom::Current(1)).await, Ok(0))); assert!(matches!( empty.seek(SeekFrom::Current(i64::MAX)).await, Ok(0) )); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/rt_metrics.rs
tokio/tests/rt_metrics.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi"), target_has_atomic = "64"))] use std::sync::mpsc; use std::time::Duration; use tokio::runtime::Runtime; use tokio::time; #[test] fn num_workers() { let rt = current_thread(); assert_eq!(1, rt.metrics().num_workers()); let rt = threaded(); assert_eq!(2, rt.metrics().num_workers()); } #[test] fn num_alive_tasks() { let rt = current_thread(); let metrics = rt.metrics(); assert_eq!(0, metrics.num_alive_tasks()); rt.block_on(rt.spawn(async move { assert_eq!(1, metrics.num_alive_tasks()); })) .unwrap(); assert_eq!(0, rt.metrics().num_alive_tasks()); let rt = threaded(); let metrics = rt.metrics(); assert_eq!(0, metrics.num_alive_tasks()); rt.block_on(rt.spawn(async move { assert_eq!(1, metrics.num_alive_tasks()); })) .unwrap(); // try for 10 seconds to see if this eventually succeeds. // wake_join() is called before the task is released, so in multithreaded // code, this means we sometimes exit the block_on before the counter decrements. for _ in 0..100 { if rt.metrics().num_alive_tasks() == 0 { break; } std::thread::sleep(std::time::Duration::from_millis(100)); } assert_eq!(0, rt.metrics().num_alive_tasks()); } #[test] fn global_queue_depth_current_thread() { use std::thread; let rt = current_thread(); let handle = rt.handle().clone(); let metrics = rt.metrics(); thread::spawn(move || { handle.spawn(async {}); }) .join() .unwrap(); assert_eq!(1, metrics.global_queue_depth()); } #[test] fn global_queue_depth_multi_thread() { for _ in 0..10 { let rt = threaded(); let metrics = rt.metrics(); if let Ok(_blocking_tasks) = try_block_threaded(&rt) { for i in 0..10 { assert_eq!(i, metrics.global_queue_depth()); rt.spawn(async {}); } return; } } panic!("exhausted every try to block the runtime"); } #[test] fn worker_total_busy_duration() { const N: usize = 5; let zero = Duration::from_millis(0); let rt = current_thread(); let metrics = rt.metrics(); rt.block_on(async { for _ in 0..N { tokio::spawn(async { tokio::task::yield_now().await; }) .await .unwrap(); } }); drop(rt); assert!(zero < metrics.worker_total_busy_duration(0)); let rt = threaded(); let metrics = rt.metrics(); rt.block_on(async { for _ in 0..N { tokio::spawn(async { tokio::task::yield_now().await; }) .await .unwrap(); } }); drop(rt); for i in 0..metrics.num_workers() { assert!(zero < metrics.worker_total_busy_duration(i)); } } #[test] fn worker_park_count() { let rt = current_thread(); let metrics = rt.metrics(); rt.block_on(async { time::sleep(Duration::from_millis(1)).await; }); drop(rt); assert!(1 <= metrics.worker_park_count(0)); let rt = threaded(); let metrics = rt.metrics(); rt.block_on(async { time::sleep(Duration::from_millis(1)).await; }); drop(rt); assert!(1 <= metrics.worker_park_count(0)); assert!(1 <= metrics.worker_park_count(1)); } #[test] fn worker_park_unpark_count() { let rt = current_thread(); let metrics = rt.metrics(); rt.block_on(rt.spawn(async {})).unwrap(); drop(rt); assert!(2 <= metrics.worker_park_unpark_count(0)); let rt = threaded(); let metrics = rt.metrics(); // Wait for workers to be parked after runtime startup. for _ in 0..100 { if 1 <= metrics.worker_park_unpark_count(0) && 1 <= metrics.worker_park_unpark_count(1) { break; } std::thread::sleep(std::time::Duration::from_millis(100)); } assert_eq!(1, metrics.worker_park_unpark_count(0)); assert_eq!(1, metrics.worker_park_unpark_count(1)); // Spawn a task to unpark and then park a worker. rt.block_on(rt.spawn(async {})).unwrap(); for _ in 0..100 { if 3 <= metrics.worker_park_unpark_count(0) || 3 <= metrics.worker_park_unpark_count(1) { break; } std::thread::sleep(std::time::Duration::from_millis(100)); } assert!(3 <= metrics.worker_park_unpark_count(0) || 3 <= metrics.worker_park_unpark_count(1)); // Both threads unpark for runtime shutdown. drop(rt); assert_eq!(0, metrics.worker_park_unpark_count(0) % 2); assert_eq!(0, metrics.worker_park_unpark_count(1) % 2); assert!(4 <= metrics.worker_park_unpark_count(0) || 4 <= metrics.worker_park_unpark_count(1)); } fn try_block_threaded(rt: &Runtime) -> Result<Vec<mpsc::Sender<()>>, mpsc::RecvTimeoutError> { let (tx, rx) = mpsc::channel(); let blocking_tasks = (0..rt.metrics().num_workers()) .map(|_| { let tx = tx.clone(); let (task, barrier) = mpsc::channel(); // Spawn a task per runtime worker to block it. rt.spawn(async move { tx.send(()).ok(); barrier.recv().ok(); }); task }) .collect(); // Make sure the previously spawned tasks are blocking the runtime by // receiving a message from each blocking task. // // If this times out we were unsuccessful in blocking the runtime and hit // a deadlock instead (which might happen and is expected behaviour). for _ in 0..rt.metrics().num_workers() { rx.recv_timeout(Duration::from_secs(1))?; } // Return senders of the mpsc channels used for blocking the runtime as a // surrogate handle for the tasks. Sending a message or dropping the senders // will unblock the runtime. Ok(blocking_tasks) } fn current_thread() -> Runtime { tokio::runtime::Builder::new_current_thread() .enable_all() .build() .unwrap() } fn threaded() -> Runtime { tokio::runtime::Builder::new_multi_thread() .worker_threads(2) .enable_all() .build() .unwrap() }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/net_panic.rs
tokio/tests/net_panic.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] #![cfg(panic = "unwind")] use std::error::Error; use tokio::net::{TcpListener, TcpStream}; use tokio::runtime::{Builder, Runtime}; mod support { pub mod panic; } use support::panic::test_panic; #[test] #[cfg_attr(miri, ignore)] // No `socket` in miri. fn udp_socket_from_std_panic_caller() -> Result<(), Box<dyn Error>> { use std::net::SocketAddr; use tokio::net::UdpSocket; let addr = "127.0.0.1:0".parse::<SocketAddr>().unwrap(); let std_sock = std::net::UdpSocket::bind(addr).unwrap(); std_sock.set_nonblocking(true).unwrap(); let panic_location_file = test_panic(|| { let rt = runtime_without_io(); rt.block_on(async { let _sock = UdpSocket::from_std(std_sock); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] #[cfg_attr(miri, ignore)] // No `socket` in miri. fn tcp_listener_from_std_panic_caller() -> Result<(), Box<dyn Error>> { let std_listener = std::net::TcpListener::bind("127.0.0.1:0").unwrap(); std_listener.set_nonblocking(true).unwrap(); let panic_location_file = test_panic(|| { let rt = runtime_without_io(); rt.block_on(async { let _ = TcpListener::from_std(std_listener); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] #[cfg_attr(miri, ignore)] // No `socket` in miri. fn tcp_stream_from_std_panic_caller() -> Result<(), Box<dyn Error>> { let std_listener = std::net::TcpListener::bind("127.0.0.1:0").unwrap(); let std_stream = std::net::TcpStream::connect(std_listener.local_addr().unwrap()).unwrap(); std_stream.set_nonblocking(true).unwrap(); let panic_location_file = test_panic(|| { let rt = runtime_without_io(); rt.block_on(async { let _ = TcpStream::from_std(std_stream); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] #[cfg(unix)] #[cfg_attr(miri, ignore)] // No `socket` in miri. fn unix_listener_bind_panic_caller() -> Result<(), Box<dyn Error>> { use tokio::net::UnixListener; let dir = tempfile::tempdir().unwrap(); let sock_path = dir.path().join("socket"); let panic_location_file = test_panic(|| { let rt = runtime_without_io(); rt.block_on(async { let _ = UnixListener::bind(&sock_path); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] #[cfg(unix)] #[cfg_attr(miri, ignore)] // No `socket` in miri. fn unix_listener_from_std_panic_caller() -> Result<(), Box<dyn Error>> { use tokio::net::UnixListener; let dir = tempfile::tempdir().unwrap(); let sock_path = dir.path().join("socket"); let std_listener = std::os::unix::net::UnixListener::bind(sock_path).unwrap(); let panic_location_file = test_panic(|| { let rt = runtime_without_io(); rt.block_on(async { let _ = UnixListener::from_std(std_listener); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] #[cfg(unix)] #[cfg_attr(miri, ignore)] // No `socket` in miri. fn unix_stream_from_std_panic_caller() -> Result<(), Box<dyn Error>> { use tokio::net::UnixStream; let dir = tempfile::tempdir().unwrap(); let sock_path = dir.path().join("socket"); let _std_listener = std::os::unix::net::UnixListener::bind(&sock_path).unwrap(); let std_stream = std::os::unix::net::UnixStream::connect(&sock_path).unwrap(); let panic_location_file = test_panic(|| { let rt = runtime_without_io(); rt.block_on(async { let _ = UnixStream::from_std(std_stream); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] #[cfg(unix)] #[cfg_attr(miri, ignore)] // No `socket` in miri. fn unix_datagram_from_std_panic_caller() -> Result<(), Box<dyn Error>> { use std::os::unix::net::UnixDatagram as StdUDS; use tokio::net::UnixDatagram; let dir = tempfile::tempdir().unwrap(); let sock_path = dir.path().join("socket"); // Bind the socket to a filesystem path // /let socket_path = tmp.path().join("socket"); let std_socket = StdUDS::bind(sock_path).unwrap(); std_socket.set_nonblocking(true).unwrap(); let panic_location_file = test_panic(move || { let rt = runtime_without_io(); rt.block_on(async { let _ = UnixDatagram::from_std(std_socket); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] #[cfg(windows)] fn server_options_max_instances_panic_caller() -> Result<(), Box<dyn Error>> { use tokio::net::windows::named_pipe::ServerOptions; let panic_location_file = test_panic(move || { let rt = runtime_without_io(); rt.block_on(async { let mut options = ServerOptions::new(); options.max_instances(255); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } // Runtime without `enable_io` so it has no IO driver set. fn runtime_without_io() -> Runtime { Builder::new_current_thread().build().unwrap() }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/net_named_pipe.rs
tokio/tests/net_named_pipe.rs
#![cfg(feature = "full")] #![cfg(windows)] use std::io; use std::time::Duration; use tokio::io::{AsyncReadExt, AsyncWriteExt}; use tokio::net::windows::named_pipe::{ClientOptions, PipeMode, ServerOptions}; use tokio::time; use windows_sys::Win32::Foundation::{ERROR_NO_DATA, ERROR_PIPE_BUSY}; #[tokio::test] async fn test_named_pipe_client_drop() -> io::Result<()> { const PIPE_NAME: &str = r"\\.\pipe\test-named-pipe-client-drop"; let mut server = ServerOptions::new().create(PIPE_NAME)?; let client = ClientOptions::new().open(PIPE_NAME)?; server.connect().await?; drop(client); // instance will be broken because client is gone match server.write_all(b"ping").await { Err(e) if e.raw_os_error() == Some(ERROR_NO_DATA as i32) => (), x => panic!("{:?}", x), } Ok(()) } #[tokio::test] async fn test_named_pipe_single_client() -> io::Result<()> { use tokio::io::{AsyncBufReadExt as _, BufReader}; const PIPE_NAME: &str = r"\\.\pipe\test-named-pipe-single-client"; let server = ServerOptions::new().create(PIPE_NAME)?; let server = tokio::spawn(async move { // Note: we wait for a client to connect. server.connect().await?; let mut server = BufReader::new(server); let mut buf = String::new(); server.read_line(&mut buf).await?; server.write_all(b"pong\n").await?; Ok::<_, io::Error>(buf) }); let client = tokio::spawn(async move { let client = ClientOptions::new().open(PIPE_NAME)?; let mut client = BufReader::new(client); let mut buf = String::new(); client.write_all(b"ping\n").await?; client.read_line(&mut buf).await?; Ok::<_, io::Error>(buf) }); let (server, client) = tokio::try_join!(server, client)?; assert_eq!(server?, "ping\n"); assert_eq!(client?, "pong\n"); Ok(()) } #[tokio::test] async fn test_named_pipe_multi_client() -> io::Result<()> { use tokio::io::{AsyncBufReadExt as _, BufReader}; const PIPE_NAME: &str = r"\\.\pipe\test-named-pipe-multi-client"; const N: usize = 10; // The first server needs to be constructed early so that clients can // be correctly connected. Otherwise calling .wait will cause the client to // error. let mut server = ServerOptions::new().create(PIPE_NAME)?; let server = tokio::spawn(async move { for _ in 0..N { // Wait for client to connect. server.connect().await?; let mut inner = BufReader::new(server); // Construct the next server to be connected before sending the one // we already have of onto a task. This ensures that the server // isn't closed (after it's done in the task) before a new one is // available. Otherwise the client might error with // `io::ErrorKind::NotFound`. server = ServerOptions::new().create(PIPE_NAME)?; tokio::spawn(async move { let mut buf = String::new(); inner.read_line(&mut buf).await?; inner.write_all(b"pong\n").await?; inner.flush().await?; Ok::<_, io::Error>(()) }); } Ok::<_, io::Error>(()) }); let mut clients = Vec::new(); for _ in 0..N { clients.push(tokio::spawn(async move { // This showcases a generic connect loop. // // We immediately try to create a client, if it's not found or the // pipe is busy we use the specialized wait function on the client // builder. let client = loop { match ClientOptions::new().open(PIPE_NAME) { Ok(client) => break client, Err(e) if e.raw_os_error() == Some(ERROR_PIPE_BUSY as i32) => (), Err(e) if e.kind() == io::ErrorKind::NotFound => (), Err(e) => return Err(e), } // Wait for a named pipe to become available. time::sleep(Duration::from_millis(10)).await; }; let mut client = BufReader::new(client); let mut buf = String::new(); client.write_all(b"ping\n").await?; client.flush().await?; client.read_line(&mut buf).await?; Ok::<_, io::Error>(buf) })); } for client in clients { let result = client.await?; assert_eq!(result?, "pong\n"); } server.await??; Ok(()) } #[tokio::test] async fn test_named_pipe_multi_client_ready() -> io::Result<()> { use tokio::io::Interest; const PIPE_NAME: &str = r"\\.\pipe\test-named-pipe-multi-client-ready"; const N: usize = 10; // The first server needs to be constructed early so that clients can // be correctly connected. Otherwise calling .wait will cause the client to // error. let mut server = ServerOptions::new().create(PIPE_NAME)?; let server = tokio::spawn(async move { for _ in 0..N { // Wait for client to connect. server.connect().await?; let inner_server = server; // Construct the next server to be connected before sending the one // we already have of onto a task. This ensures that the server // isn't closed (after it's done in the task) before a new one is // available. Otherwise the client might error with // `io::ErrorKind::NotFound`. server = ServerOptions::new().create(PIPE_NAME)?; tokio::spawn(async move { let server = inner_server; { let mut read_buf = [0u8; 5]; let mut read_buf_cursor = 0; loop { server.readable().await?; let buf = &mut read_buf[read_buf_cursor..]; match server.try_read(buf) { Ok(n) => { read_buf_cursor += n; if read_buf_cursor == read_buf.len() { break; } } Err(e) if e.kind() == io::ErrorKind::WouldBlock => { continue; } Err(e) => { return Err(e); } } } }; { let write_buf = b"pong\n"; let mut write_buf_cursor = 0; loop { server.writable().await?; let buf = &write_buf[write_buf_cursor..]; match server.try_write(buf) { Ok(n) => { write_buf_cursor += n; if write_buf_cursor == write_buf.len() { break; } } Err(e) if e.kind() == io::ErrorKind::WouldBlock => { continue; } Err(e) => { return Err(e); } } } } Ok::<_, io::Error>(()) }); } Ok::<_, io::Error>(()) }); let mut clients = Vec::new(); for _ in 0..N { clients.push(tokio::spawn(async move { // This showcases a generic connect loop. // // We immediately try to create a client, if it's not found or the // pipe is busy we use the specialized wait function on the client // builder. let client = loop { match ClientOptions::new().open(PIPE_NAME) { Ok(client) => break client, Err(e) if e.raw_os_error() == Some(ERROR_PIPE_BUSY as i32) => (), Err(e) if e.kind() == io::ErrorKind::NotFound => (), Err(e) => return Err(e), } // Wait for a named pipe to become available. time::sleep(Duration::from_millis(10)).await; }; let mut read_buf = [0u8; 5]; let mut read_buf_cursor = 0; let write_buf = b"ping\n"; let mut write_buf_cursor = 0; loop { let mut interest = Interest::READABLE; if write_buf_cursor < write_buf.len() { interest |= Interest::WRITABLE; } let ready = client.ready(interest).await?; if ready.is_readable() { let buf = &mut read_buf[read_buf_cursor..]; match client.try_read(buf) { Ok(n) => { read_buf_cursor += n; if read_buf_cursor == read_buf.len() { break; } } Err(e) if e.kind() == io::ErrorKind::WouldBlock => { continue; } Err(e) => { return Err(e); } } } if ready.is_writable() { let buf = &write_buf[write_buf_cursor..]; if buf.is_empty() { continue; } match client.try_write(buf) { Ok(n) => { write_buf_cursor += n; } Err(e) if e.kind() == io::ErrorKind::WouldBlock => { continue; } Err(e) => { return Err(e); } } } } let buf = String::from_utf8_lossy(&read_buf).into_owned(); Ok::<_, io::Error>(buf) })); } for client in clients { let result = client.await?; assert_eq!(result?, "pong\n"); } server.await??; Ok(()) } // This tests that message mode works as expected. #[tokio::test] async fn test_named_pipe_mode_message() -> io::Result<()> { // it's easy to accidentally get a seemingly working test here because byte pipes // often return contents at write boundaries. to make sure we're doing the right thing we // explicitly test that it doesn't work in byte mode. _named_pipe_mode_message(PipeMode::Message).await?; _named_pipe_mode_message(PipeMode::Byte).await } async fn _named_pipe_mode_message(mode: PipeMode) -> io::Result<()> { let pipe_name = format!( r"\\.\pipe\test-named-pipe-mode-message-{}", matches!(mode, PipeMode::Message) ); let mut buf = [0u8; 32]; let mut server = ServerOptions::new() .first_pipe_instance(true) .pipe_mode(mode) .create(&pipe_name)?; let mut client = ClientOptions::new().pipe_mode(mode).open(&pipe_name)?; server.connect().await?; // this needs a few iterations, presumably Windows waits for a few calls before merging buffers for _ in 0..10 { client.write_all(b"hello").await?; server.write_all(b"world").await?; } for _ in 0..10 { let n = server.read(&mut buf).await?; if buf[..n] != b"hello"[..] { assert!(matches!(mode, PipeMode::Byte)); return Ok(()); } let n = client.read(&mut buf).await?; if buf[..n] != b"world"[..] { assert!(matches!(mode, PipeMode::Byte)); return Ok(()); } } // byte mode should have errored before. assert!(matches!(mode, PipeMode::Message)); Ok(()) } // This tests `NamedPipeServer::connect` with various access settings. #[tokio::test] async fn test_named_pipe_access() -> io::Result<()> { const PIPE_NAME: &str = r"\\.\pipe\test-named-pipe-access"; for (inb, outb) in [(true, true), (true, false), (false, true)] { let (tx, rx) = tokio::sync::oneshot::channel(); let server = tokio::spawn(async move { let s = ServerOptions::new() .access_inbound(inb) .access_outbound(outb) .create(PIPE_NAME)?; let mut connect_fut = tokio_test::task::spawn(s.connect()); assert!(connect_fut.poll().is_pending()); tx.send(()).unwrap(); connect_fut.await }); // Wait for the server to call connect. rx.await.unwrap(); let _ = ClientOptions::new().read(outb).write(inb).open(PIPE_NAME)?; server.await??; } Ok(()) }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/sync_mpsc.rs
tokio/tests/sync_mpsc.rs
#![allow(clippy::redundant_clone)] #![warn(rust_2018_idioms)] #![cfg(feature = "sync")] #[cfg(all(target_family = "wasm", not(target_os = "wasi")))] use wasm_bindgen_test::wasm_bindgen_test as test; #[cfg(all(target_family = "wasm", not(target_os = "wasi")))] use wasm_bindgen_test::wasm_bindgen_test as maybe_tokio_test; #[cfg(not(all(target_family = "wasm", not(target_os = "wasi"))))] use tokio::test as maybe_tokio_test; use std::fmt; use std::panic; use std::sync::Arc; use tokio::sync::mpsc; use tokio::sync::mpsc::error::{TryRecvError, TrySendError}; use tokio_test::*; #[cfg(not(target_family = "wasm"))] mod support { pub(crate) mod mpsc_stream; } #[allow(unused)] trait AssertRefUnwindSafe: panic::RefUnwindSafe {} impl<T> AssertRefUnwindSafe for mpsc::OwnedPermit<T> {} impl<'a, T> AssertRefUnwindSafe for mpsc::Permit<'a, T> {} impl<'a, T> AssertRefUnwindSafe for mpsc::PermitIterator<'a, T> {} impl<T> AssertRefUnwindSafe for mpsc::Receiver<T> {} impl<T> AssertRefUnwindSafe for mpsc::Sender<T> {} impl<T> AssertRefUnwindSafe for mpsc::UnboundedReceiver<T> {} impl<T> AssertRefUnwindSafe for mpsc::UnboundedSender<T> {} impl<T> AssertRefUnwindSafe for mpsc::WeakSender<T> {} impl<T> AssertRefUnwindSafe for mpsc::WeakUnboundedSender<T> {} #[allow(unused)] trait AssertUnwindSafe: panic::UnwindSafe {} impl<T> AssertUnwindSafe for mpsc::OwnedPermit<T> {} impl<'a, T> AssertUnwindSafe for mpsc::Permit<'a, T> {} impl<'a, T> AssertUnwindSafe for mpsc::PermitIterator<'a, T> {} impl<T> AssertUnwindSafe for mpsc::Receiver<T> {} impl<T> AssertUnwindSafe for mpsc::Sender<T> {} impl<T> AssertUnwindSafe for mpsc::UnboundedReceiver<T> {} impl<T> AssertUnwindSafe for mpsc::UnboundedSender<T> {} impl<T> AssertUnwindSafe for mpsc::WeakSender<T> {} impl<T> AssertUnwindSafe for mpsc::WeakUnboundedSender<T> {} #[maybe_tokio_test] async fn send_recv_with_buffer() { let (tx, mut rx) = mpsc::channel::<i32>(16); // Using poll_ready / try_send // let permit assert_ready_ok!(tx.reserve()); let permit = tx.reserve().await.unwrap(); permit.send(1); // Without poll_ready tx.try_send(2).unwrap(); drop(tx); let val = rx.recv().await; assert_eq!(val, Some(1)); let val = rx.recv().await; assert_eq!(val, Some(2)); let val = rx.recv().await; assert!(val.is_none()); } #[tokio::test] #[cfg(feature = "full")] async fn reserve_disarm() { let (tx, mut rx) = mpsc::channel::<i32>(2); let tx1 = tx.clone(); let tx2 = tx.clone(); let tx3 = tx.clone(); let tx4 = tx; // We should be able to `poll_ready` two handles without problem let permit1 = assert_ok!(tx1.reserve().await); let permit2 = assert_ok!(tx2.reserve().await); // But a third should not be ready let mut r3 = tokio_test::task::spawn(tx3.reserve()); assert_pending!(r3.poll()); let mut r4 = tokio_test::task::spawn(tx4.reserve()); assert_pending!(r4.poll()); // Using one of the reserved slots should allow a new handle to become ready permit1.send(1); // We also need to receive for the slot to be free assert!(!r3.is_woken()); rx.recv().await.unwrap(); // Now there's a free slot! assert!(r3.is_woken()); assert!(!r4.is_woken()); // Dropping a permit should also open up a slot drop(permit2); assert!(r4.is_woken()); let mut r1 = tokio_test::task::spawn(tx1.reserve()); assert_pending!(r1.poll()); } #[tokio::test] #[cfg(all(feature = "full", not(target_os = "wasi")))] // Wasi doesn't support threads async fn send_recv_stream_with_buffer() { use tokio_stream::StreamExt; let (tx, rx) = support::mpsc_stream::channel_stream::<i32>(16); let mut rx = Box::pin(rx); tokio::spawn(async move { assert_ok!(tx.send(1).await); assert_ok!(tx.send(2).await); }); assert_eq!(Some(1), rx.next().await); assert_eq!(Some(2), rx.next().await); assert_eq!(None, rx.next().await); } #[tokio::test] #[cfg(feature = "full")] async fn async_send_recv_with_buffer() { let (tx, mut rx) = mpsc::channel(16); tokio::spawn(async move { assert_ok!(tx.send(1).await); assert_ok!(tx.send(2).await); }); assert_eq!(Some(1), rx.recv().await); assert_eq!(Some(2), rx.recv().await); assert_eq!(None, rx.recv().await); } #[tokio::test] #[cfg(feature = "full")] async fn async_send_recv_many_with_buffer() { let (tx, mut rx) = mpsc::channel(2); let mut buffer = Vec::<i32>::with_capacity(3); // With `limit=0` does not sleep, returns immediately assert_eq!(0, rx.recv_many(&mut buffer, 0).await); let handle = tokio::spawn(async move { assert_ok!(tx.send(1).await); assert_ok!(tx.send(2).await); assert_ok!(tx.send(7).await); assert_ok!(tx.send(0).await); }); let limit = 3; let mut recv_count = 0usize; while recv_count < 4 { recv_count += rx.recv_many(&mut buffer, limit).await; assert_eq!(buffer.len(), recv_count); } assert_eq!(vec![1, 2, 7, 0], buffer); assert_eq!(0, rx.recv_many(&mut buffer, limit).await); handle.await.unwrap(); } #[tokio::test] #[cfg(feature = "full")] async fn start_send_past_cap() { use std::future::Future; let mut t1 = tokio_test::task::spawn(()); let (tx1, mut rx) = mpsc::channel(1); let tx2 = tx1.clone(); assert_ok!(tx1.try_send(())); let mut r1 = Box::pin(tx1.reserve()); t1.enter(|cx, _| assert_pending!(r1.as_mut().poll(cx))); { let mut r2 = tokio_test::task::spawn(tx2.reserve()); assert_pending!(r2.poll()); drop(r1); assert!(rx.recv().await.is_some()); assert!(r2.is_woken()); assert!(!t1.is_woken()); } drop(tx1); drop(tx2); assert!(rx.recv().await.is_none()); } #[test] #[should_panic] #[cfg(not(target_family = "wasm"))] // wasm currently doesn't support unwinding fn buffer_gteq_one() { mpsc::channel::<i32>(0); } #[maybe_tokio_test] async fn send_recv_unbounded() { let (tx, mut rx) = mpsc::unbounded_channel::<i32>(); // Using `try_send` assert_ok!(tx.send(1)); assert_ok!(tx.send(2)); assert_eq!(rx.recv().await, Some(1)); assert_eq!(rx.recv().await, Some(2)); drop(tx); assert!(rx.recv().await.is_none()); } #[maybe_tokio_test] async fn send_recv_many_unbounded() { let (tx, mut rx) = mpsc::unbounded_channel::<i32>(); let mut buffer: Vec<i32> = Vec::new(); // With `limit=0` does not sleep, returns immediately rx.recv_many(&mut buffer, 0).await; assert_eq!(0, buffer.len()); assert_ok!(tx.send(7)); assert_ok!(tx.send(13)); assert_ok!(tx.send(100)); assert_ok!(tx.send(1002)); rx.recv_many(&mut buffer, 0).await; assert_eq!(0, buffer.len()); let mut count = 0; while count < 4 { count += rx.recv_many(&mut buffer, 1).await; } assert_eq!(count, 4); assert_eq!(vec![7, 13, 100, 1002], buffer); let final_capacity = buffer.capacity(); assert!(final_capacity > 0); buffer.clear(); assert_ok!(tx.send(5)); assert_ok!(tx.send(6)); assert_ok!(tx.send(7)); assert_ok!(tx.send(2)); // Re-use existing capacity count = rx.recv_many(&mut buffer, 32).await; assert_eq!(final_capacity, buffer.capacity()); assert_eq!(count, 4); assert_eq!(vec![5, 6, 7, 2], buffer); drop(tx); // recv_many will immediately return zero if the channel // is closed and no more messages are waiting assert_eq!(0, rx.recv_many(&mut buffer, 4).await); assert!(rx.recv().await.is_none()); } #[tokio::test] #[cfg(feature = "full")] async fn send_recv_many_bounded_capacity() { let mut buffer: Vec<String> = Vec::with_capacity(9); let limit = buffer.capacity(); let (tx, mut rx) = mpsc::channel(100); let mut expected: Vec<String> = (0..limit) .map(|x: usize| format!("{x}")) .collect::<Vec<_>>(); for x in expected.clone() { tx.send(x).await.unwrap() } tx.send("one more".to_string()).await.unwrap(); // Here `recv_many` receives all but the last value; // the initial capacity is adequate, so the buffer does // not increase in side. assert_eq!(buffer.capacity(), rx.recv_many(&mut buffer, limit).await); assert_eq!(expected, buffer); assert_eq!(limit, buffer.capacity()); // Receive up more values: assert_eq!(1, rx.recv_many(&mut buffer, limit).await); assert!(buffer.capacity() > limit); expected.push("one more".to_string()); assert_eq!(expected, buffer); tokio::spawn(async move { tx.send("final".to_string()).await.unwrap(); }); // 'tx' is dropped, but `recv_many` is guaranteed not // to return 0 as the channel has outstanding permits assert_eq!(1, rx.recv_many(&mut buffer, limit).await); expected.push("final".to_string()); assert_eq!(expected, buffer); // The channel is now closed and `recv_many` returns 0. assert_eq!(0, rx.recv_many(&mut buffer, limit).await); assert_eq!(expected, buffer); } #[tokio::test] #[cfg(feature = "full")] async fn send_recv_many_unbounded_capacity() { let mut buffer: Vec<String> = Vec::with_capacity(9); // capacity >= 9 let limit = buffer.capacity(); let (tx, mut rx) = mpsc::unbounded_channel(); let mut expected: Vec<String> = (0..limit) .map(|x: usize| format!("{x}")) .collect::<Vec<_>>(); for x in expected.clone() { tx.send(x).unwrap() } tx.send("one more".to_string()).unwrap(); // Here `recv_many` receives all but the last value; // the initial capacity is adequate, so the buffer does // not increase in side. assert_eq!(buffer.capacity(), rx.recv_many(&mut buffer, limit).await); assert_eq!(expected, buffer); assert_eq!(limit, buffer.capacity()); // Receive up more values: assert_eq!(1, rx.recv_many(&mut buffer, limit).await); assert!(buffer.capacity() > limit); expected.push("one more".to_string()); assert_eq!(expected, buffer); tokio::spawn(async move { tx.send("final".to_string()).unwrap(); }); // 'tx' is dropped, but `recv_many` is guaranteed not // to return 0 as the channel has outstanding permits assert_eq!(1, rx.recv_many(&mut buffer, limit).await); expected.push("final".to_string()); assert_eq!(expected, buffer); // The channel is now closed and `recv_many` returns 0. assert_eq!(0, rx.recv_many(&mut buffer, limit).await); assert_eq!(expected, buffer); } #[tokio::test] #[cfg(feature = "full")] async fn async_send_recv_unbounded() { let (tx, mut rx) = mpsc::unbounded_channel(); tokio::spawn(async move { assert_ok!(tx.send(1)); assert_ok!(tx.send(2)); }); assert_eq!(Some(1), rx.recv().await); assert_eq!(Some(2), rx.recv().await); assert_eq!(None, rx.recv().await); } #[tokio::test] #[cfg(all(feature = "full", not(target_os = "wasi")))] // Wasi doesn't support threads async fn send_recv_stream_unbounded() { use tokio_stream::StreamExt; let (tx, rx) = support::mpsc_stream::unbounded_channel_stream::<i32>(); let mut rx = Box::pin(rx); tokio::spawn(async move { assert_ok!(tx.send(1)); assert_ok!(tx.send(2)); }); assert_eq!(Some(1), rx.next().await); assert_eq!(Some(2), rx.next().await); assert_eq!(None, rx.next().await); } #[maybe_tokio_test] async fn no_t_bounds_buffer() { struct NoImpls; let (tx, mut rx) = mpsc::channel(100); // sender should be Debug even though T isn't Debug is_debug(&tx); // same with Receiver is_debug(&rx); // and sender should be Clone even though T isn't Clone assert!(tx.clone().try_send(NoImpls).is_ok()); assert!(rx.recv().await.is_some()); } #[maybe_tokio_test] async fn no_t_bounds_unbounded() { struct NoImpls; let (tx, mut rx) = mpsc::unbounded_channel(); // sender should be Debug even though T isn't Debug is_debug(&tx); // same with Receiver is_debug(&rx); // and sender should be Clone even though T isn't Clone assert!(tx.clone().send(NoImpls).is_ok()); assert!(rx.recv().await.is_some()); } #[tokio::test] #[cfg(feature = "full")] async fn send_recv_buffer_limited() { let (tx, mut rx) = mpsc::channel::<i32>(1); // Reserve capacity let p1 = assert_ok!(tx.reserve().await); // Send first message p1.send(1); // Not ready let mut p2 = tokio_test::task::spawn(tx.reserve()); assert_pending!(p2.poll()); // Take the value assert!(rx.recv().await.is_some()); // Notified assert!(p2.is_woken()); // Trying to send fails assert_err!(tx.try_send(1337)); // Send second let permit = assert_ready_ok!(p2.poll()); permit.send(2); assert!(rx.recv().await.is_some()); } #[maybe_tokio_test] async fn recv_close_gets_none_idle() { let (tx, mut rx) = mpsc::channel::<i32>(10); rx.close(); assert!(rx.recv().await.is_none()); assert_err!(tx.send(1).await); } #[tokio::test] #[cfg(feature = "full")] async fn recv_close_gets_none_reserved() { let (tx1, mut rx) = mpsc::channel::<i32>(1); let tx2 = tx1.clone(); let permit1 = assert_ok!(tx1.reserve().await); let mut permit2 = tokio_test::task::spawn(tx2.reserve()); assert_pending!(permit2.poll()); rx.close(); assert!(permit2.is_woken()); assert_ready_err!(permit2.poll()); { let mut recv = tokio_test::task::spawn(rx.recv()); assert_pending!(recv.poll()); permit1.send(123); assert!(recv.is_woken()); let v = assert_ready!(recv.poll()); assert_eq!(v, Some(123)); } assert!(rx.recv().await.is_none()); } #[maybe_tokio_test] async fn tx_close_gets_none() { let (_, mut rx) = mpsc::channel::<i32>(10); assert!(rx.recv().await.is_none()); } #[maybe_tokio_test] async fn try_send_fail() { let (tx, mut rx) = mpsc::channel(1); tx.try_send("hello").unwrap(); // This should fail match assert_err!(tx.try_send("fail")) { TrySendError::Full(..) => {} _ => panic!(), } assert_eq!(rx.recv().await, Some("hello")); assert_ok!(tx.try_send("goodbye")); drop(tx); assert_eq!(rx.recv().await, Some("goodbye")); assert!(rx.recv().await.is_none()); } #[maybe_tokio_test] async fn try_send_fail_with_try_recv() { let (tx, mut rx) = mpsc::channel(1); tx.try_send("hello").unwrap(); // This should fail match assert_err!(tx.try_send("fail")) { TrySendError::Full(..) => {} _ => panic!(), } assert_eq!(rx.try_recv(), Ok("hello")); assert_ok!(tx.try_send("goodbye")); drop(tx); assert_eq!(rx.try_recv(), Ok("goodbye")); assert_eq!(rx.try_recv(), Err(TryRecvError::Disconnected)); } #[maybe_tokio_test] async fn reserve_many_above_cap() { const MAX_PERMITS: usize = tokio::sync::Semaphore::MAX_PERMITS; let (tx, _rx) = mpsc::channel::<()>(1); assert_err!(tx.reserve_many(2).await); assert_err!(tx.reserve_many(MAX_PERMITS + 1).await); assert_err!(tx.reserve_many(usize::MAX).await); } #[test] fn try_reserve_many_zero() { let (tx, rx) = mpsc::channel::<()>(1); // Succeeds when not closed. assert!(assert_ok!(tx.try_reserve_many(0)).next().is_none()); // Even when channel is full. tx.try_send(()).unwrap(); assert!(assert_ok!(tx.try_reserve_many(0)).next().is_none()); drop(rx); // Closed error when closed. assert_eq!( assert_err!(tx.try_reserve_many(0)), TrySendError::Closed(()) ); } #[maybe_tokio_test] async fn reserve_many_zero() { let (tx, rx) = mpsc::channel::<()>(1); // Succeeds when not closed. assert!(assert_ok!(tx.reserve_many(0).await).next().is_none()); // Even when channel is full. tx.send(()).await.unwrap(); assert!(assert_ok!(tx.reserve_many(0).await).next().is_none()); drop(rx); // Closed error when closed. assert_err!(tx.reserve_many(0).await); } #[maybe_tokio_test] async fn try_reserve_many_edge_cases() { const MAX_PERMITS: usize = tokio::sync::Semaphore::MAX_PERMITS; let (tx, rx) = mpsc::channel::<()>(1); let mut permit = assert_ok!(tx.try_reserve_many(0)); assert!(permit.next().is_none()); let permit = tx.try_reserve_many(MAX_PERMITS + 1); match assert_err!(permit) { TrySendError::Full(..) => {} _ => panic!(), } let permit = tx.try_reserve_many(usize::MAX); match assert_err!(permit) { TrySendError::Full(..) => {} _ => panic!(), } // Dropping the receiver should close the channel drop(rx); assert_err!(tx.reserve_many(0).await); } #[maybe_tokio_test] async fn try_reserve_fails() { let (tx, mut rx) = mpsc::channel(1); let permit = tx.try_reserve().unwrap(); // This should fail match assert_err!(tx.try_reserve()) { TrySendError::Full(()) => {} _ => panic!(), } permit.send("foo"); assert_eq!(rx.recv().await, Some("foo")); // Dropping permit releases the slot. let permit = tx.try_reserve().unwrap(); drop(permit); let _permit = tx.try_reserve().unwrap(); } #[maybe_tokio_test] async fn reserve_many_and_send() { let (tx, mut rx) = mpsc::channel(100); for i in 0..100 { for permit in assert_ok!(tx.reserve_many(i).await) { permit.send("foo"); assert_eq!(rx.recv().await, Some("foo")); } assert_eq!(rx.try_recv(), Err(TryRecvError::Empty)); } } #[maybe_tokio_test] async fn try_reserve_many_and_send() { let (tx, mut rx) = mpsc::channel(100); for i in 0..100 { for permit in assert_ok!(tx.try_reserve_many(i)) { permit.send("foo"); assert_eq!(rx.recv().await, Some("foo")); } assert_eq!(rx.try_recv(), Err(TryRecvError::Empty)); } } #[maybe_tokio_test] async fn reserve_many_on_closed_channel() { let (tx, rx) = mpsc::channel::<()>(100); drop(rx); assert_err!(tx.reserve_many(10).await); } #[maybe_tokio_test] async fn try_reserve_many_on_closed_channel() { let (tx, rx) = mpsc::channel::<usize>(100); drop(rx); match assert_err!(tx.try_reserve_many(10)) { TrySendError::Closed(()) => {} _ => panic!(), }; } #[maybe_tokio_test] #[cfg_attr(miri, ignore)] // Too slow on miri. async fn try_reserve_many_full() { // Reserve n capacity and send k messages for n in 1..100 { for k in 0..n { let (tx, mut rx) = mpsc::channel::<usize>(n); let permits = assert_ok!(tx.try_reserve_many(n)); assert_eq!(permits.len(), n); assert_eq!(tx.capacity(), 0); match assert_err!(tx.try_reserve_many(1)) { TrySendError::Full(..) => {} _ => panic!(), }; for permit in permits.take(k) { permit.send(0); } // We only used k permits on the n reserved assert_eq!(tx.capacity(), n - k); // We can reserve more permits assert_ok!(tx.try_reserve_many(1)); // But not more than the current capacity match assert_err!(tx.try_reserve_many(n - k + 1)) { TrySendError::Full(..) => {} _ => panic!(), }; for _i in 0..k { assert_eq!(rx.recv().await, Some(0)); } // Now that we've received everything, capacity should be back to n assert_eq!(tx.capacity(), n); } } } #[tokio::test] #[cfg(feature = "full")] async fn drop_permit_releases_permit() { // poll_ready reserves capacity, ensure that the capacity is released if tx // is dropped w/o sending a value. let (tx1, _rx) = mpsc::channel::<i32>(1); let tx2 = tx1.clone(); let permit = assert_ok!(tx1.reserve().await); let mut reserve2 = tokio_test::task::spawn(tx2.reserve()); assert_pending!(reserve2.poll()); drop(permit); assert!(reserve2.is_woken()); assert_ready_ok!(reserve2.poll()); } #[maybe_tokio_test] async fn drop_permit_iterator_releases_permits() { // poll_ready reserves capacity, ensure that the capacity is released if tx // is dropped w/o sending a value. for n in 1..100 { let (tx1, _rx) = mpsc::channel::<i32>(n); let tx2 = tx1.clone(); let permits = assert_ok!(tx1.reserve_many(n).await); let mut reserve2 = tokio_test::task::spawn(tx2.reserve_many(n)); assert_pending!(reserve2.poll()); drop(permits); assert!(reserve2.is_woken()); let permits = assert_ready_ok!(reserve2.poll()); drop(permits); assert_eq!(tx1.capacity(), n); } } #[maybe_tokio_test] async fn dropping_rx_closes_channel() { let (tx, rx) = mpsc::channel(100); let msg = Arc::new(()); assert_ok!(tx.try_send(msg.clone())); drop(rx); assert_err!(tx.reserve().await); assert_err!(tx.reserve_many(10).await); assert_eq!(1, Arc::strong_count(&msg)); } #[test] fn dropping_rx_closes_channel_for_try() { let (tx, rx) = mpsc::channel(100); let msg = Arc::new(()); tx.try_send(msg.clone()).unwrap(); drop(rx); assert!(matches!( tx.try_send(msg.clone()), Err(TrySendError::Closed(_)) )); assert!(matches!(tx.try_reserve(), Err(TrySendError::Closed(_)))); assert!(matches!( tx.try_reserve_owned(), Err(TrySendError::Closed(_)) )); assert_eq!(1, Arc::strong_count(&msg)); } #[test] fn unconsumed_messages_are_dropped() { let msg = Arc::new(()); let (tx, rx) = mpsc::channel(100); tx.try_send(msg.clone()).unwrap(); assert_eq!(2, Arc::strong_count(&msg)); drop((tx, rx)); assert_eq!(1, Arc::strong_count(&msg)); } #[test] #[cfg(all(feature = "full", not(target_os = "wasi")))] // Wasi doesn't support threads fn blocking_recv() { let (tx, mut rx) = mpsc::channel::<u8>(1); let sync_code = std::thread::spawn(move || { assert_eq!(Some(10), rx.blocking_recv()); }); tokio::runtime::Runtime::new() .unwrap() .block_on(async move { let _ = tx.send(10).await; }); sync_code.join().unwrap() } #[tokio::test] #[should_panic] #[cfg(not(target_family = "wasm"))] // wasm currently doesn't support unwinding async fn blocking_recv_async() { let (_tx, mut rx) = mpsc::channel::<()>(1); let _ = rx.blocking_recv(); } #[test] #[cfg(all(feature = "full", not(target_os = "wasi")))] // Wasi doesn't support threads fn blocking_send() { let (tx, mut rx) = mpsc::channel::<u8>(1); let sync_code = std::thread::spawn(move || { tx.blocking_send(10).unwrap(); }); tokio::runtime::Runtime::new() .unwrap() .block_on(async move { assert_eq!(Some(10), rx.recv().await); }); sync_code.join().unwrap() } #[tokio::test] #[should_panic] #[cfg(not(target_family = "wasm"))] // wasm currently doesn't support unwinding async fn blocking_send_async() { let (tx, _rx) = mpsc::channel::<()>(1); let _ = tx.blocking_send(()); } #[tokio::test] #[cfg(feature = "full")] async fn ready_close_cancel_bounded() { let (tx, mut rx) = mpsc::channel::<()>(100); let _tx2 = tx.clone(); let permit = assert_ok!(tx.reserve().await); rx.close(); let mut recv = tokio_test::task::spawn(rx.recv()); assert_pending!(recv.poll()); drop(permit); assert!(recv.is_woken()); let val = assert_ready!(recv.poll()); assert!(val.is_none()); } #[tokio::test] #[cfg(feature = "full")] async fn permit_available_not_acquired_close() { let (tx1, mut rx) = mpsc::channel::<()>(1); let tx2 = tx1.clone(); let permit1 = assert_ok!(tx1.reserve().await); let mut permit2 = tokio_test::task::spawn(tx2.reserve()); assert_pending!(permit2.poll()); rx.close(); drop(permit1); assert!(permit2.is_woken()); drop(permit2); assert!(rx.recv().await.is_none()); } #[test] fn try_recv_bounded() { let (tx, mut rx) = mpsc::channel(5); tx.try_send("hello").unwrap(); tx.try_send("hello").unwrap(); tx.try_send("hello").unwrap(); tx.try_send("hello").unwrap(); tx.try_send("hello").unwrap(); assert!(tx.try_send("hello").is_err()); assert_eq!(Ok("hello"), rx.try_recv()); assert_eq!(Ok("hello"), rx.try_recv()); assert_eq!(Ok("hello"), rx.try_recv()); assert_eq!(Ok("hello"), rx.try_recv()); assert_eq!(Ok("hello"), rx.try_recv()); assert_eq!(Err(TryRecvError::Empty), rx.try_recv()); tx.try_send("hello").unwrap(); tx.try_send("hello").unwrap(); tx.try_send("hello").unwrap(); tx.try_send("hello").unwrap(); assert_eq!(Ok("hello"), rx.try_recv()); tx.try_send("hello").unwrap(); tx.try_send("hello").unwrap(); assert!(tx.try_send("hello").is_err()); assert_eq!(Ok("hello"), rx.try_recv()); assert_eq!(Ok("hello"), rx.try_recv()); assert_eq!(Ok("hello"), rx.try_recv()); assert_eq!(Ok("hello"), rx.try_recv()); assert_eq!(Ok("hello"), rx.try_recv()); assert_eq!(Err(TryRecvError::Empty), rx.try_recv()); tx.try_send("hello").unwrap(); tx.try_send("hello").unwrap(); tx.try_send("hello").unwrap(); drop(tx); assert_eq!(Ok("hello"), rx.try_recv()); assert_eq!(Ok("hello"), rx.try_recv()); assert_eq!(Ok("hello"), rx.try_recv()); assert_eq!(Err(TryRecvError::Disconnected), rx.try_recv()); } #[test] fn try_recv_unbounded() { for num in 0..100 { let (tx, mut rx) = mpsc::unbounded_channel(); for i in 0..num { tx.send(i).unwrap(); } for i in 0..num { assert_eq!(rx.try_recv(), Ok(i)); } assert_eq!(rx.try_recv(), Err(TryRecvError::Empty)); drop(tx); assert_eq!(rx.try_recv(), Err(TryRecvError::Disconnected)); } } #[test] fn try_recv_after_receiver_close() { let (_tx, mut rx) = mpsc::channel::<()>(5); assert_eq!(Err(TryRecvError::Empty), rx.try_recv()); rx.close(); assert_eq!(Err(TryRecvError::Disconnected), rx.try_recv()); } #[test] fn try_recv_close_while_empty_bounded() { let (tx, mut rx) = mpsc::channel::<()>(5); assert_eq!(Err(TryRecvError::Empty), rx.try_recv()); drop(tx); assert_eq!(Err(TryRecvError::Disconnected), rx.try_recv()); } #[test] fn try_recv_close_while_empty_unbounded() { let (tx, mut rx) = mpsc::unbounded_channel::<()>(); assert_eq!(Err(TryRecvError::Empty), rx.try_recv()); drop(tx); assert_eq!(Err(TryRecvError::Disconnected), rx.try_recv()); } #[tokio::test(start_paused = true)] #[cfg(feature = "full")] async fn recv_timeout() { use tokio::sync::mpsc::error::SendTimeoutError::{Closed, Timeout}; use tokio::time::Duration; let (tx, rx) = mpsc::channel(5); assert_eq!(tx.send_timeout(10, Duration::from_secs(1)).await, Ok(())); assert_eq!(tx.send_timeout(20, Duration::from_secs(1)).await, Ok(())); assert_eq!(tx.send_timeout(30, Duration::from_secs(1)).await, Ok(())); assert_eq!(tx.send_timeout(40, Duration::from_secs(1)).await, Ok(())); assert_eq!(tx.send_timeout(50, Duration::from_secs(1)).await, Ok(())); assert_eq!( tx.send_timeout(60, Duration::from_secs(1)).await, Err(Timeout(60)) ); drop(rx); assert_eq!( tx.send_timeout(70, Duration::from_secs(1)).await, Err(Closed(70)) ); } #[test] #[should_panic = "there is no reactor running, must be called from the context of a Tokio 1.x runtime"] #[cfg(not(target_family = "wasm"))] // wasm currently doesn't support unwinding fn recv_timeout_panic() { use futures::future::FutureExt; use tokio::time::Duration; let (tx, _rx) = mpsc::channel(5); tx.send_timeout(10, Duration::from_secs(1)).now_or_never(); } // Tests that channel `capacity` changes and `max_capacity` stays the same #[tokio::test] async fn test_tx_capacity() { let (tx, _rx) = mpsc::channel::<()>(10); // both capacities are same before assert_eq!(tx.capacity(), 10); assert_eq!(tx.max_capacity(), 10); let _permit = tx.reserve().await.unwrap(); // after reserve, only capacity should drop by one assert_eq!(tx.capacity(), 9); assert_eq!(tx.max_capacity(), 10); tx.send(()).await.unwrap(); // after send, capacity should drop by one again assert_eq!(tx.capacity(), 8); assert_eq!(tx.max_capacity(), 10); } #[tokio::test] async fn test_rx_is_closed_when_calling_close_with_sender() { // is_closed should return true after calling close but still has a sender let (_tx, mut rx) = mpsc::channel::<()>(10); rx.close(); assert!(rx.is_closed()); } #[tokio::test] async fn test_rx_is_closed_when_dropping_all_senders() { // is_closed should return true after dropping all senders let (tx, rx) = mpsc::channel::<()>(10); let another_tx = tx.clone(); let task = tokio::spawn(async move { drop(another_tx); }); drop(tx); let _ = task.await; assert!(rx.is_closed()); } #[tokio::test] async fn test_rx_is_not_closed_when_there_are_senders() { // is_closed should return false when there is a sender let (_tx, rx) = mpsc::channel::<()>(10); assert!(!rx.is_closed()); } #[tokio::test] async fn test_rx_is_not_closed_when_there_are_senders_and_buffer_filled() { // is_closed should return false when there is a sender, even if enough messages have been sent to fill the channel let (tx, rx) = mpsc::channel(10); for i in 0..10 { assert!(tx.send(i).await.is_ok()); } assert!(!rx.is_closed()); } #[tokio::test] async fn test_rx_is_closed_when_there_are_no_senders_and_there_are_messages() { // is_closed should return true when there are messages in the buffer, but no senders let (tx, rx) = mpsc::channel(10); for i in 0..10 { assert!(tx.send(i).await.is_ok()); } drop(tx); assert!(rx.is_closed()); } #[tokio::test] async fn test_rx_is_closed_when_there_are_messages_and_close_is_called() { // is_closed should return true when there are messages in the buffer, and close is called let (tx, mut rx) = mpsc::channel(10); for i in 0..10 { assert!(tx.send(i).await.is_ok()); } rx.close(); assert!(rx.is_closed()); } #[tokio::test] async fn test_rx_is_not_closed_when_there_are_permits_but_not_senders() { // is_closed should return false when there is a permit (but no senders) let (tx, rx) = mpsc::channel::<()>(10); let _permit = tx.reserve_owned().await.expect("Failed to reserve permit"); assert!(!rx.is_closed()); } #[tokio::test] async fn test_rx_is_empty_when_no_messages_were_sent() { let (_tx, rx) = mpsc::channel::<()>(10); assert!(rx.is_empty()) } #[tokio::test] async fn test_rx_is_not_empty_when_there_are_messages_in_the_buffer() { let (tx, rx) = mpsc::channel::<()>(10); assert!(tx.send(()).await.is_ok()); assert!(!rx.is_empty()) } #[tokio::test] async fn test_rx_is_not_empty_when_the_buffer_is_full() { let (tx, rx) = mpsc::channel(10); for i in 0..10 { assert!(tx.send(i).await.is_ok()); } assert!(!rx.is_empty()) } #[tokio::test] async fn test_rx_is_not_empty_when_all_but_one_messages_are_consumed() { let (tx, mut rx) = mpsc::channel(10); for i in 0..10 { assert!(tx.send(i).await.is_ok()); } for _ in 0..9 { assert!(rx.recv().await.is_some()); } assert!(!rx.is_empty()) } #[tokio::test] async fn test_rx_is_empty_when_all_messages_are_consumed() { let (tx, mut rx) = mpsc::channel(10); for i in 0..10 { assert!(tx.send(i).await.is_ok()); } while rx.try_recv().is_ok() {} assert!(rx.is_empty()) } #[tokio::test] async fn test_rx_is_empty_all_senders_are_dropped_and_messages_consumed() { let (tx, mut rx) = mpsc::channel(10); for i in 0..10 { assert!(tx.send(i).await.is_ok()); } drop(tx); for _ in 0..10 { assert!(rx.recv().await.is_some()); } assert!(rx.is_empty()) } #[tokio::test] async fn test_rx_len_on_empty_channel() { let (_tx, rx) = mpsc::channel::<()>(100); assert_eq!(rx.len(), 0); } #[tokio::test] async fn test_rx_len_on_empty_channel_without_senders() { // when all senders are dropped, a "closed" value is added to the end of the linked list. // here we test that the "closed" value does not change the len of the channel. let (tx, rx) = mpsc::channel::<()>(100); drop(tx); assert_eq!(rx.len(), 0); } #[tokio::test] async fn test_rx_len_on_filled_channel() { let (tx, rx) = mpsc::channel(100);
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
true
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/rt_handle_block_on.rs
tokio/tests/rt_handle_block_on.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(miri)))] // All io tests that deal with shutdown is currently ignored because there are known bugs in with // shutting down the io driver while concurrently registering new resources. See // https://github.com/tokio-rs/tokio/pull/3569#pullrequestreview-612703467 for more details. // // When this has been fixed we want to re-enable these tests. use std::time::Duration; use tokio::runtime::{Handle, Runtime}; use tokio::sync::mpsc; #[cfg(not(target_os = "wasi"))] use tokio::{net, time}; #[cfg(not(target_os = "wasi"))] // Wasi doesn't support threads macro_rules! multi_threaded_rt_test { ($($t:tt)*) => { mod threaded_scheduler_4_threads_only { use super::*; $($t)* fn rt() -> Runtime { tokio::runtime::Builder::new_multi_thread() .worker_threads(4) .enable_all() .build() .unwrap() } } mod threaded_scheduler_1_thread_only { use super::*; $($t)* fn rt() -> Runtime { tokio::runtime::Builder::new_multi_thread() .worker_threads(1) .enable_all() .build() .unwrap() } } } } #[cfg(not(target_os = "wasi"))] macro_rules! rt_test { ($($t:tt)*) => { mod current_thread_scheduler { use super::*; $($t)* fn rt() -> Runtime { tokio::runtime::Builder::new_current_thread() .enable_all() .build() .unwrap() } } mod threaded_scheduler_4_threads { use super::*; $($t)* fn rt() -> Runtime { tokio::runtime::Builder::new_multi_thread() .worker_threads(4) .enable_all() .build() .unwrap() } } mod threaded_scheduler_1_thread { use super::*; $($t)* fn rt() -> Runtime { tokio::runtime::Builder::new_multi_thread() .worker_threads(1) .enable_all() .build() .unwrap() } } } } // ==== runtime independent futures ====== #[test] fn basic() { test_with_runtimes(|| { let one = Handle::current().block_on(async { 1 }); assert_eq!(1, one); }); } #[test] fn bounded_mpsc_channel() { test_with_runtimes(|| { let (tx, mut rx) = mpsc::channel(1024); Handle::current().block_on(tx.send(42)).unwrap(); let value = Handle::current().block_on(rx.recv()).unwrap(); assert_eq!(value, 42); }); } #[test] fn unbounded_mpsc_channel() { test_with_runtimes(|| { let (tx, mut rx) = mpsc::unbounded_channel(); let _ = tx.send(42); let value = Handle::current().block_on(rx.recv()).unwrap(); assert_eq!(value, 42); }) } #[test] fn yield_in_block_in_place() { test_with_runtimes(|| { Handle::current().block_on(async { tokio::task::block_in_place(|| { Handle::current().block_on(tokio::task::yield_now()); }); }); }) } #[cfg(not(target_os = "wasi"))] // Wasi doesn't support file operations or bind rt_test! { use tokio::fs; // ==== spawn blocking futures ====== #[test] fn basic_fs() { let rt = rt(); let _enter = rt.enter(); let contents = Handle::current() .block_on(fs::read_to_string("Cargo.toml")) .unwrap(); assert!(contents.contains("https://tokio.rs")); } #[test] fn fs_shutdown_before_started() { let rt = rt(); let _enter = rt.enter(); rt.shutdown_timeout(Duration::from_secs(1000)); let err: std::io::Error = Handle::current() .block_on(fs::read_to_string("Cargo.toml")) .unwrap_err(); assert_eq!(err.kind(), std::io::ErrorKind::Other); let inner_err = err.get_ref().expect("no inner error"); assert_eq!(inner_err.to_string(), "background task failed"); } #[test] fn basic_spawn_blocking() { use tokio::task::spawn_blocking; let rt = rt(); let _enter = rt.enter(); let answer = Handle::current() .block_on(spawn_blocking(|| { std::thread::sleep(Duration::from_millis(100)); 42 })) .unwrap(); assert_eq!(answer, 42); } #[test] fn spawn_blocking_after_shutdown_fails() { use tokio::task::spawn_blocking; let rt = rt(); let _enter = rt.enter(); rt.shutdown_timeout(Duration::from_secs(1000)); let join_err = Handle::current() .block_on(spawn_blocking(|| { std::thread::sleep(Duration::from_millis(100)); 42 })) .unwrap_err(); assert!(join_err.is_cancelled()); } #[test] fn spawn_blocking_started_before_shutdown_continues() { use tokio::task::spawn_blocking; let rt = rt(); let _enter = rt.enter(); let handle = spawn_blocking(|| { std::thread::sleep(Duration::from_secs(1)); 42 }); rt.shutdown_timeout(Duration::from_secs(1000)); let answer = Handle::current().block_on(handle).unwrap(); assert_eq!(answer, 42); } #[test] fn thread_park_ok() { use core::task::Poll; let rt = rt(); let mut exit = false; rt.handle().block_on(core::future::poll_fn(move |cx| { if exit { return Poll::Ready(()); } cx.waker().wake_by_ref(); // Verify that consuming the park token does not result in a hang. std::thread::current().unpark(); std::thread::park(); exit = true; Poll::Pending })); } // ==== net ====== #[test] #[cfg_attr(miri, ignore)] // No `socket` in miri. fn tcp_listener_bind() { let rt = rt(); let _enter = rt.enter(); Handle::current() .block_on(net::TcpListener::bind("127.0.0.1:0")) .unwrap(); } // All io tests are ignored for now. See above why that is. #[ignore] #[test] fn tcp_listener_connect_after_shutdown() { let rt = rt(); let _enter = rt.enter(); rt.shutdown_timeout(Duration::from_secs(1000)); let err = Handle::current() .block_on(net::TcpListener::bind("127.0.0.1:0")) .unwrap_err(); assert_eq!(err.kind(), std::io::ErrorKind::Other); assert_eq!( err.get_ref().unwrap().to_string(), "A Tokio 1.x context was found, but it is being shutdown.", ); } // All io tests are ignored for now. See above why that is. #[ignore] #[test] fn tcp_listener_connect_before_shutdown() { let rt = rt(); let _enter = rt.enter(); let bind_future = net::TcpListener::bind("127.0.0.1:0"); rt.shutdown_timeout(Duration::from_secs(1000)); let err = Handle::current().block_on(bind_future).unwrap_err(); assert_eq!(err.kind(), std::io::ErrorKind::Other); assert_eq!( err.get_ref().unwrap().to_string(), "A Tokio 1.x context was found, but it is being shutdown.", ); } #[test] #[cfg_attr(miri, ignore)] // No `socket` in miri. fn udp_socket_bind() { let rt = rt(); let _enter = rt.enter(); Handle::current() .block_on(net::UdpSocket::bind("127.0.0.1:0")) .unwrap(); } // All io tests are ignored for now. See above why that is. #[ignore] #[test] fn udp_stream_bind_after_shutdown() { let rt = rt(); let _enter = rt.enter(); rt.shutdown_timeout(Duration::from_secs(1000)); let err = Handle::current() .block_on(net::UdpSocket::bind("127.0.0.1:0")) .unwrap_err(); assert_eq!(err.kind(), std::io::ErrorKind::Other); assert_eq!( err.get_ref().unwrap().to_string(), "A Tokio 1.x context was found, but it is being shutdown.", ); } // All io tests are ignored for now. See above why that is. #[ignore] #[test] fn udp_stream_bind_before_shutdown() { let rt = rt(); let _enter = rt.enter(); let bind_future = net::UdpSocket::bind("127.0.0.1:0"); rt.shutdown_timeout(Duration::from_secs(1000)); let err = Handle::current().block_on(bind_future).unwrap_err(); assert_eq!(err.kind(), std::io::ErrorKind::Other); assert_eq!( err.get_ref().unwrap().to_string(), "A Tokio 1.x context was found, but it is being shutdown.", ); } // All io tests are ignored for now. See above why that is. #[ignore] #[cfg(unix)] #[test] fn unix_listener_bind_after_shutdown() { let rt = rt(); let _enter = rt.enter(); let dir = tempfile::tempdir().unwrap(); let path = dir.path().join("socket"); rt.shutdown_timeout(Duration::from_secs(1000)); let err = net::UnixListener::bind(path).unwrap_err(); assert_eq!(err.kind(), std::io::ErrorKind::Other); assert_eq!( err.get_ref().unwrap().to_string(), "A Tokio 1.x context was found, but it is being shutdown.", ); } // All io tests are ignored for now. See above why that is. #[ignore] #[cfg(unix)] #[test] fn unix_listener_shutdown_after_bind() { let rt = rt(); let _enter = rt.enter(); let dir = tempfile::tempdir().unwrap(); let path = dir.path().join("socket"); let listener = net::UnixListener::bind(path).unwrap(); rt.shutdown_timeout(Duration::from_secs(1000)); // this should not timeout but fail immediately since the runtime has been shutdown let err = Handle::current().block_on(listener.accept()).unwrap_err(); assert_eq!(err.kind(), std::io::ErrorKind::Other); assert_eq!(err.get_ref().unwrap().to_string(), "reactor gone"); } // All io tests are ignored for now. See above why that is. #[ignore] #[cfg(unix)] #[test] fn unix_listener_shutdown_after_accept() { let rt = rt(); let _enter = rt.enter(); let dir = tempfile::tempdir().unwrap(); let path = dir.path().join("socket"); let listener = net::UnixListener::bind(path).unwrap(); let accept_future = listener.accept(); rt.shutdown_timeout(Duration::from_secs(1000)); // this should not timeout but fail immediately since the runtime has been shutdown let err = Handle::current().block_on(accept_future).unwrap_err(); assert_eq!(err.kind(), std::io::ErrorKind::Other); assert_eq!(err.get_ref().unwrap().to_string(), "reactor gone"); } // ==== nesting ====== #[test] #[should_panic( expected = "Cannot start a runtime from within a runtime. This happens because a function (like `block_on`) attempted to block the current thread while the thread is being used to drive asynchronous tasks." )] fn nesting() { fn some_non_async_function() -> i32 { Handle::current().block_on(time::sleep(Duration::from_millis(10))); 1 } let rt = rt(); rt.block_on(async { some_non_async_function() }); } #[test] fn spawn_after_runtime_dropped() { use futures::future::FutureExt; let rt = rt(); let handle = rt.block_on(async move { Handle::current() }); let jh1 = handle.spawn(futures::future::pending::<()>()); drop(rt); let jh2 = handle.spawn(futures::future::pending::<()>()); let err1 = jh1.now_or_never().unwrap().unwrap_err(); let err2 = jh2.now_or_never().unwrap().unwrap_err(); assert!(err1.is_cancelled()); assert!(err2.is_cancelled()); } } #[cfg(not(target_os = "wasi"))] multi_threaded_rt_test! { #[cfg(unix)] #[cfg_attr(miri, ignore)] // No `socket` in miri. #[test] fn unix_listener_bind() { let rt = rt(); let _enter = rt.enter(); let dir = tempfile::tempdir().unwrap(); let path = dir.path().join("socket"); let listener = net::UnixListener::bind(path).unwrap(); // this should timeout and not fail immediately since the runtime has not been shutdown let _: tokio::time::error::Elapsed = Handle::current() .block_on(tokio::time::timeout( Duration::from_millis(10), listener.accept(), )) .unwrap_err(); } // ==== timers ====== // `Handle::block_on` doesn't work with timer futures on a current thread runtime as there is no // one to drive the timers so they will just hang forever. Therefore they are not tested. #[test] fn sleep() { let rt = rt(); let _enter = rt.enter(); Handle::current().block_on(time::sleep(Duration::from_millis(100))); } #[test] #[should_panic(expected = "A Tokio 1.x context was found, but it is being shutdown.")] fn sleep_before_shutdown_panics() { let rt = rt(); let _enter = rt.enter(); let f = time::sleep(Duration::from_millis(100)); rt.shutdown_timeout(Duration::from_secs(1000)); Handle::current().block_on(f); } #[test] #[should_panic(expected = "A Tokio 1.x context was found, but it is being shutdown.")] fn sleep_after_shutdown_panics() { let rt = rt(); let _enter = rt.enter(); rt.shutdown_timeout(Duration::from_secs(1000)); Handle::current().block_on(time::sleep(Duration::from_millis(100))); } } // ==== utils ====== /// Create a new multi threaded runtime #[cfg(not(target_os = "wasi"))] fn new_multi_thread(n: usize) -> Runtime { tokio::runtime::Builder::new_multi_thread() .worker_threads(n) .enable_all() .build() .unwrap() } /// Create a new single threaded runtime fn new_current_thread() -> Runtime { tokio::runtime::Builder::new_current_thread() .enable_all() .build() .unwrap() } /// Utility to test things on both kinds of runtimes both before and after shutting it down. fn test_with_runtimes<F>(f: F) where F: Fn(), { { let rt = new_current_thread(); let _enter = rt.enter(); f(); rt.shutdown_timeout(Duration::from_secs(1000)); f(); } #[cfg(not(target_os = "wasi"))] { let rt = new_multi_thread(1); let _enter = rt.enter(); f(); rt.shutdown_timeout(Duration::from_secs(1000)); f(); } #[cfg(not(target_os = "wasi"))] { let rt = new_multi_thread(4); let _enter = rt.enter(); f(); rt.shutdown_timeout(Duration::from_secs(1000)); f(); } }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/unwindsafe.rs
tokio/tests/unwindsafe.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // Wasi does not support panic recovery use std::panic::{RefUnwindSafe, UnwindSafe}; #[test] fn notify_is_unwind_safe() { is_unwind_safe::<tokio::sync::Notify>(); } #[test] fn join_handle_is_unwind_safe() { is_unwind_safe::<tokio::task::JoinHandle<()>>(); } #[test] fn net_types_are_unwind_safe() { is_unwind_safe::<tokio::net::TcpListener>(); is_unwind_safe::<tokio::net::TcpSocket>(); is_unwind_safe::<tokio::net::TcpStream>(); is_unwind_safe::<tokio::net::UdpSocket>(); } #[test] #[cfg(unix)] fn unix_net_types_are_unwind_safe() { is_unwind_safe::<tokio::net::UnixDatagram>(); is_unwind_safe::<tokio::net::UnixListener>(); is_unwind_safe::<tokio::net::UnixStream>(); } #[test] #[cfg(windows)] fn windows_net_types_are_unwind_safe() { use tokio::net::windows::named_pipe::NamedPipeClient; use tokio::net::windows::named_pipe::NamedPipeServer; is_unwind_safe::<NamedPipeClient>(); is_unwind_safe::<NamedPipeServer>(); } fn is_unwind_safe<T: UnwindSafe + RefUnwindSafe>() {}
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_read_to_end.rs
tokio/tests/io_read_to_end.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use std::pin::Pin; use std::task::{Context, Poll}; use tokio::io::{AsyncRead, AsyncReadExt, ReadBuf}; use tokio_test::assert_ok; use tokio_test::io::Builder; #[tokio::test] async fn read_to_end() { let mut buf = vec![]; let mut rd: &[u8] = b"hello world"; let n = assert_ok!(rd.read_to_end(&mut buf).await); assert_eq!(n, 11); assert_eq!(buf[..], b"hello world"[..]); } #[derive(Copy, Clone, Debug)] enum State { Initializing, JustFilling, Done, } struct UninitTest { num_init: usize, state: State, } impl AsyncRead for UninitTest { fn poll_read( self: Pin<&mut Self>, _cx: &mut Context<'_>, buf: &mut ReadBuf<'_>, ) -> Poll<std::io::Result<()>> { let me = Pin::into_inner(self); let real_num_init = buf.initialized().len() - buf.filled().len(); assert_eq!(real_num_init, me.num_init, "{:?}", me.state); match me.state { State::Initializing => { buf.initialize_unfilled_to(me.num_init + 2); buf.advance(1); me.num_init += 1; if me.num_init == 24 { me.state = State::JustFilling; } } State::JustFilling => { buf.advance(1); me.num_init -= 1; if me.num_init == 15 { // The buffer is resized on next call. me.num_init = 0; me.state = State::Done; } } State::Done => { /* .. do nothing .. */ } } Poll::Ready(Ok(())) } } #[tokio::test] async fn read_to_end_uninit() { let mut buf = Vec::with_capacity(64); let mut test = UninitTest { num_init: 0, state: State::Initializing, }; test.read_to_end(&mut buf).await.unwrap(); assert_eq!(buf.len(), 33); } #[tokio::test] #[cfg_attr(miri, ignore)] // too slow with miri async fn read_to_end_doesnt_grow_with_capacity() { let arr: Vec<u8> = (0..100).collect(); // We only test from 32 since we allocate at least 32 bytes each time for len in 32..100 { let bytes = &arr[..len]; for split in 0..len { for cap in 0..101 { let mut mock = if split == 0 { Builder::new().read(bytes).build() } else { Builder::new() .read(&bytes[..split]) .read(&bytes[split..]) .build() }; let mut buf = Vec::with_capacity(cap); AsyncReadExt::read_to_end(&mut mock, &mut buf) .await .unwrap(); // It has the right data. assert_eq!(buf.as_slice(), bytes); // Unless cap was smaller than length, then we did not reallocate. if cap >= len { assert_eq!(buf.capacity(), cap); } } } } } #[tokio::test] async fn read_to_end_grows_capacity_if_unfit() { let bytes = b"the_vector_startingcap_will_be_smaller"; let mut mock = Builder::new().read(bytes).build(); let initial_capacity = bytes.len() - 4; let mut buf = Vec::with_capacity(initial_capacity); AsyncReadExt::read_to_end(&mut mock, &mut buf) .await .unwrap(); // *4 since it doubles when it doesn't fit and again when reaching EOF assert_eq!(buf.capacity(), initial_capacity * 4); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/task_panic.rs
tokio/tests/task_panic.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] #![cfg(panic = "unwind")] use futures::future; use std::error::Error; use tokio::runtime::Builder; use tokio::task::{self, block_in_place}; mod support { pub mod panic; } use support::panic::test_panic; #[test] fn block_in_place_panic_caller() -> Result<(), Box<dyn Error>> { let panic_location_file = test_panic(|| { let rt = Builder::new_current_thread().enable_all().build().unwrap(); rt.block_on(async { block_in_place(|| {}); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] fn local_set_spawn_local_panic_caller() -> Result<(), Box<dyn Error>> { let panic_location_file = test_panic(|| { let _local = task::LocalSet::new(); task::spawn_local(async {}); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] fn local_set_block_on_panic_caller() -> Result<(), Box<dyn Error>> { let panic_location_file = test_panic(|| { let rt = Builder::new_current_thread().enable_all().build().unwrap(); let local = task::LocalSet::new(); rt.block_on(async { local.block_on(&rt, future::pending::<()>()); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] fn spawn_panic_caller() -> Result<(), Box<dyn Error>> { let panic_location_file = test_panic(|| { tokio::spawn(future::pending::<()>()); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] fn local_key_sync_scope_panic_caller() -> Result<(), Box<dyn Error>> { tokio::task_local! { static NUMBER: u32; } let panic_location_file = test_panic(|| { NUMBER.sync_scope(1, || { NUMBER.with(|_| { NUMBER.sync_scope(1, || {}); }); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] fn local_key_with_panic_caller() -> Result<(), Box<dyn Error>> { tokio::task_local! { static NUMBER: u32; } let panic_location_file = test_panic(|| { NUMBER.with(|_| {}); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] fn local_key_get_panic_caller() -> Result<(), Box<dyn Error>> { tokio::task_local! { static NUMBER: u32; } let panic_location_file = test_panic(|| { NUMBER.get(); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/rt_threaded.rs
tokio/tests/rt_threaded.rs
#![warn(rust_2018_idioms)] // Too slow on miri. #![cfg(all(feature = "full", not(target_os = "wasi"), not(miri)))] use tokio::io::{AsyncReadExt, AsyncWriteExt}; use tokio::net::{TcpListener, TcpStream}; use tokio::runtime; use tokio::sync::oneshot; use tokio_test::{assert_err, assert_ok}; use std::future::{poll_fn, Future}; use std::pin::Pin; use std::sync::atomic::Ordering::Relaxed; use std::sync::atomic::{AtomicUsize, Ordering}; use std::sync::{mpsc, Arc, Mutex}; use std::task::{Context, Poll, Waker}; macro_rules! cfg_metrics { ($($t:tt)*) => { #[cfg(all(tokio_unstable, target_has_atomic = "64"))] { $( $t )* } } } #[test] fn single_thread() { // No panic when starting a runtime w/ a single thread let _ = runtime::Builder::new_multi_thread() .enable_all() .worker_threads(1) .build() .unwrap(); } #[test] fn many_oneshot_futures() { // used for notifying the main thread const NUM: usize = 1_000; for _ in 0..5 { let (tx, rx) = mpsc::channel(); let rt = rt(); let cnt = Arc::new(AtomicUsize::new(0)); for _ in 0..NUM { let cnt = cnt.clone(); let tx = tx.clone(); rt.spawn(async move { let num = cnt.fetch_add(1, Relaxed) + 1; if num == NUM { tx.send(()).unwrap(); } }); } rx.recv().unwrap(); // Wait for the pool to shutdown drop(rt); } } #[test] fn spawn_two() { let rt = rt(); let out = rt.block_on(async { let (tx, rx) = oneshot::channel(); tokio::spawn(async move { tokio::spawn(async move { tx.send("ZOMG").unwrap(); }); }); assert_ok!(rx.await) }); assert_eq!(out, "ZOMG"); cfg_metrics! { let metrics = rt.metrics(); drop(rt); assert_eq!(1, metrics.remote_schedule_count()); let mut local = 0; for i in 0..metrics.num_workers() { local += metrics.worker_local_schedule_count(i); } assert_eq!(1, local); } } #[test] fn many_multishot_futures() { const CHAIN: usize = 200; const CYCLES: usize = 5; const TRACKS: usize = 50; for _ in 0..50 { let rt = rt(); let mut start_txs = Vec::with_capacity(TRACKS); let mut final_rxs = Vec::with_capacity(TRACKS); for _ in 0..TRACKS { let (start_tx, mut chain_rx) = tokio::sync::mpsc::channel(10); for _ in 0..CHAIN { let (next_tx, next_rx) = tokio::sync::mpsc::channel(10); // Forward all the messages rt.spawn(async move { while let Some(v) = chain_rx.recv().await { next_tx.send(v).await.unwrap(); } }); chain_rx = next_rx; } // This final task cycles if needed let (final_tx, final_rx) = tokio::sync::mpsc::channel(10); let cycle_tx = start_tx.clone(); let mut rem = CYCLES; rt.spawn(async move { for _ in 0..CYCLES { let msg = chain_rx.recv().await.unwrap(); rem -= 1; if rem == 0 { final_tx.send(msg).await.unwrap(); } else { cycle_tx.send(msg).await.unwrap(); } } }); start_txs.push(start_tx); final_rxs.push(final_rx); } { rt.block_on(async move { for start_tx in start_txs { start_tx.send("ping").await.unwrap(); } for mut final_rx in final_rxs { final_rx.recv().await.unwrap(); } }); } } } #[test] fn lifo_slot_budget() { async fn my_fn() { spawn_another(); } fn spawn_another() { tokio::spawn(my_fn()); } let rt = runtime::Builder::new_multi_thread() .enable_all() .worker_threads(1) .build() .unwrap(); let (send, recv) = oneshot::channel(); rt.spawn(async move { tokio::spawn(my_fn()); let _ = send.send(()); }); let _ = rt.block_on(recv); } #[test] #[cfg_attr(miri, ignore)] // No `socket` in miri. fn spawn_shutdown() { let rt = rt(); let (tx, rx) = mpsc::channel(); rt.block_on(async { tokio::spawn(client_server(tx.clone())); }); // Use spawner rt.spawn(client_server(tx)); assert_ok!(rx.recv()); assert_ok!(rx.recv()); drop(rt); assert_err!(rx.try_recv()); } async fn client_server(tx: mpsc::Sender<()>) { let server = assert_ok!(TcpListener::bind("127.0.0.1:0").await); // Get the assigned address let addr = assert_ok!(server.local_addr()); // Spawn the server tokio::spawn(async move { // Accept a socket let (mut socket, _) = server.accept().await.unwrap(); // Write some data socket.write_all(b"hello").await.unwrap(); }); let mut client = TcpStream::connect(&addr).await.unwrap(); let mut buf = vec![]; client.read_to_end(&mut buf).await.unwrap(); assert_eq!(buf, b"hello"); tx.send(()).unwrap(); } #[test] fn drop_threadpool_drops_futures() { for _ in 0..1_000 { let num_inc = Arc::new(AtomicUsize::new(0)); let num_dec = Arc::new(AtomicUsize::new(0)); let num_drop = Arc::new(AtomicUsize::new(0)); struct Never(Arc<AtomicUsize>); impl Future for Never { type Output = (); fn poll(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<()> { Poll::Pending } } impl Drop for Never { fn drop(&mut self) { self.0.fetch_add(1, Relaxed); } } let a = num_inc.clone(); let b = num_dec.clone(); let rt = runtime::Builder::new_multi_thread() .enable_all() .on_thread_start(move || { a.fetch_add(1, Relaxed); }) .on_thread_stop(move || { b.fetch_add(1, Relaxed); }) .build() .unwrap(); rt.spawn(Never(num_drop.clone())); // Wait for the pool to shutdown drop(rt); // Assert that only a single thread was spawned. let a = num_inc.load(Relaxed); assert!(a >= 1); // Assert that all threads shutdown let b = num_dec.load(Relaxed); assert_eq!(a, b); // Assert that the future was dropped let c = num_drop.load(Relaxed); assert_eq!(c, 1); } } #[test] fn start_stop_callbacks_called() { use std::sync::atomic::{AtomicUsize, Ordering}; let after_start = Arc::new(AtomicUsize::new(0)); let before_stop = Arc::new(AtomicUsize::new(0)); let after_inner = after_start.clone(); let before_inner = before_stop.clone(); let rt = tokio::runtime::Builder::new_multi_thread() .enable_all() .on_thread_start(move || { after_inner.clone().fetch_add(1, Ordering::Relaxed); }) .on_thread_stop(move || { before_inner.clone().fetch_add(1, Ordering::Relaxed); }) .build() .unwrap(); let (tx, rx) = oneshot::channel(); rt.spawn(async move { assert_ok!(tx.send(())); }); assert_ok!(rt.block_on(rx)); drop(rt); assert!(after_start.load(Ordering::Relaxed) > 0); assert!(before_stop.load(Ordering::Relaxed) > 0); } #[test] // too slow on miri #[cfg_attr(miri, ignore)] fn blocking() { // used for notifying the main thread const NUM: usize = 1_000; for _ in 0..10 { let (tx, rx) = mpsc::channel(); let rt = rt(); let cnt = Arc::new(AtomicUsize::new(0)); // there are four workers in the pool // so, if we run 4 blocking tasks, we know that handoff must have happened let block = Arc::new(std::sync::Barrier::new(5)); for _ in 0..4 { let block = block.clone(); rt.spawn(async move { tokio::task::block_in_place(move || { block.wait(); block.wait(); }) }); } block.wait(); for _ in 0..NUM { let cnt = cnt.clone(); let tx = tx.clone(); rt.spawn(async move { let num = cnt.fetch_add(1, Relaxed) + 1; if num == NUM { tx.send(()).unwrap(); } }); } rx.recv().unwrap(); // Wait for the pool to shutdown block.wait(); } } #[test] fn multi_threadpool() { use tokio::sync::oneshot; let rt1 = rt(); let rt2 = rt(); let (tx, rx) = oneshot::channel(); let (done_tx, done_rx) = mpsc::channel(); rt2.spawn(async move { rx.await.unwrap(); done_tx.send(()).unwrap(); }); rt1.spawn(async move { tx.send(()).unwrap(); }); done_rx.recv().unwrap(); } // When `block_in_place` returns, it attempts to reclaim the yielded runtime // worker. In this case, the remainder of the task is on the runtime worker and // must take part in the cooperative task budgeting system. // // The test ensures that, when this happens, attempting to consume from a // channel yields occasionally even if there are values ready to receive. #[test] fn coop_and_block_in_place() { let rt = tokio::runtime::Builder::new_multi_thread() // Setting max threads to 1 prevents another thread from claiming the // runtime worker yielded as part of `block_in_place` and guarantees the // same thread will reclaim the worker at the end of the // `block_in_place` call. .max_blocking_threads(1) .build() .unwrap(); rt.block_on(async move { let (tx, mut rx) = tokio::sync::mpsc::channel(1024); // Fill the channel for _ in 0..1024 { tx.send(()).await.unwrap(); } drop(tx); tokio::spawn(async move { // Block in place without doing anything tokio::task::block_in_place(|| {}); // Receive all the values, this should trigger a `Pending` as the // coop limit will be reached. poll_fn(|cx| { while let Poll::Ready(v) = { tokio::pin! { let fut = rx.recv(); } Pin::new(&mut fut).poll(cx) } { if v.is_none() { panic!("did not yield"); } } Poll::Ready(()) }) .await }) .await .unwrap(); }); } #[test] fn yield_after_block_in_place() { let rt = tokio::runtime::Builder::new_multi_thread() .worker_threads(1) .build() .unwrap(); rt.block_on(async { tokio::spawn(async move { // Block in place then enter a new runtime tokio::task::block_in_place(|| { let rt = tokio::runtime::Builder::new_current_thread() .build() .unwrap(); rt.block_on(async {}); }); // Yield, then complete tokio::task::yield_now().await; }) .await .unwrap() }); } // Testing this does not panic #[test] fn max_blocking_threads() { let _rt = tokio::runtime::Builder::new_multi_thread() .max_blocking_threads(1) .build() .unwrap(); } #[test] #[should_panic] fn max_blocking_threads_set_to_zero() { let _rt = tokio::runtime::Builder::new_multi_thread() .max_blocking_threads(0) .build() .unwrap(); } /// Regression test for #6445. /// /// After #6445, setting `global_queue_interval` to 1 is now technically valid. /// This test confirms that there is no regression in `multi_thread_runtime` /// when global_queue_interval is set to 1. #[test] fn global_queue_interval_set_to_one() { let rt = tokio::runtime::Builder::new_multi_thread() .global_queue_interval(1) .build() .unwrap(); // Perform a simple work. let cnt = Arc::new(AtomicUsize::new(0)); rt.block_on(async { let mut set = tokio::task::JoinSet::new(); for _ in 0..10 { let cnt = cnt.clone(); set.spawn(async move { cnt.fetch_add(1, Ordering::Relaxed) }); } while let Some(res) = set.join_next().await { res.unwrap(); } }); assert_eq!(cnt.load(Relaxed), 10); } #[tokio::test(flavor = "multi_thread", worker_threads = 2)] async fn hang_on_shutdown() { let (sync_tx, sync_rx) = std::sync::mpsc::channel::<()>(); tokio::spawn(async move { tokio::task::block_in_place(|| sync_rx.recv().ok()); }); tokio::spawn(async { tokio::time::sleep(std::time::Duration::from_secs(2)).await; drop(sync_tx); }); tokio::time::sleep(std::time::Duration::from_secs(1)).await; } /// Demonstrates tokio-rs/tokio#3869 #[test] fn wake_during_shutdown() { struct Shared { waker: Option<Waker>, } struct MyFuture { shared: Arc<Mutex<Shared>>, put_waker: bool, } impl MyFuture { fn new() -> (Self, Self) { let shared = Arc::new(Mutex::new(Shared { waker: None })); let f1 = MyFuture { shared: shared.clone(), put_waker: true, }; let f2 = MyFuture { shared, put_waker: false, }; (f1, f2) } } impl Future for MyFuture { type Output = (); fn poll(self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<()> { let me = Pin::into_inner(self); let mut lock = me.shared.lock().unwrap(); if me.put_waker { lock.waker = Some(cx.waker().clone()); } Poll::Pending } } impl Drop for MyFuture { fn drop(&mut self) { let mut lock = self.shared.lock().unwrap(); if !self.put_waker { lock.waker.take().unwrap().wake(); } drop(lock); } } let rt = tokio::runtime::Builder::new_multi_thread() .worker_threads(1) .enable_all() .build() .unwrap(); let (f1, f2) = MyFuture::new(); rt.spawn(f1); rt.spawn(f2); rt.block_on(async { tokio::time::sleep(tokio::time::Duration::from_millis(20)).await }); } #[should_panic] #[tokio::test] async fn test_block_in_place1() { tokio::task::block_in_place(|| {}); } #[tokio::test(flavor = "multi_thread")] async fn test_block_in_place2() { tokio::task::block_in_place(|| {}); } #[should_panic] #[tokio::main(flavor = "current_thread")] #[test] async fn test_block_in_place3() { tokio::task::block_in_place(|| {}); } #[tokio::main] #[test] async fn test_block_in_place4() { tokio::task::block_in_place(|| {}); } // Repro for tokio-rs/tokio#5239 #[test] fn test_nested_block_in_place_with_block_on_between() { let rt = runtime::Builder::new_multi_thread() .worker_threads(1) // Needs to be more than 0 .max_blocking_threads(1) .build() .unwrap(); // Triggered by a race condition, so run a few times to make sure it is OK. for _ in 0..100 { let h = rt.handle().clone(); rt.block_on(async move { tokio::spawn(async move { tokio::task::block_in_place(|| { h.block_on(async { tokio::task::block_in_place(|| {}); }); }) }) .await .unwrap() }); } } #[test] fn yield_now_in_block_in_place() { let rt = runtime::Builder::new_multi_thread() .worker_threads(1) .build() .unwrap(); rt.block_on(async { tokio::spawn(async { tokio::task::block_in_place(|| { tokio::runtime::Handle::current().block_on(tokio::task::yield_now()); }) }) .await .unwrap() }) } #[test] fn mutex_in_block_in_place() { const BUDGET: usize = 128; let rt = runtime::Builder::new_multi_thread() .worker_threads(1) .build() .unwrap(); rt.block_on(async { let lock = tokio::sync::Mutex::new(0); tokio::spawn(async move { tokio::task::block_in_place(|| { tokio::runtime::Handle::current().block_on(async move { for i in 0..(BUDGET + 1) { let mut guard = lock.lock().await; *guard = i; } }); }) }) .await .unwrap(); }) } // Testing the tuning logic is tricky as it is inherently timing based, and more // of a heuristic than an exact behavior. This test checks that the interval // changes over time based on load factors. There are no assertions, completion // is sufficient. If there is a regression, this test will hang. In theory, we // could add limits, but that would be likely to fail on CI. #[test] #[cfg(not(tokio_no_tuning_tests))] fn test_tuning() { use std::sync::atomic::AtomicBool; use std::time::Duration; let rt = runtime::Builder::new_multi_thread() .worker_threads(1) .build() .unwrap(); fn iter(flag: Arc<AtomicBool>, counter: Arc<AtomicUsize>, stall: bool) { if flag.load(Relaxed) { if stall { std::thread::sleep(Duration::from_micros(5)); } counter.fetch_add(1, Relaxed); tokio::spawn(async move { iter(flag, counter, stall) }); } } let flag = Arc::new(AtomicBool::new(true)); let counter = Arc::new(AtomicUsize::new(61)); let interval = Arc::new(AtomicUsize::new(61)); { let flag = flag.clone(); let counter = counter.clone(); rt.spawn(async move { iter(flag, counter, true) }); } // Now, hammer the injection queue until the interval drops. let mut n = 0; loop { let curr = interval.load(Relaxed); if curr <= 8 { n += 1; } else { n = 0; } // Make sure we get a few good rounds. Jitter in the tuning could result // in one "good" value without being representative of reaching a good // state. if n == 3 { break; } if Arc::strong_count(&interval) < 5_000 { let counter = counter.clone(); let interval = interval.clone(); rt.spawn(async move { let prev = counter.swap(0, Relaxed); interval.store(prev, Relaxed); }); std::thread::yield_now(); } } flag.store(false, Relaxed); let w = Arc::downgrade(&interval); drop(interval); while w.strong_count() > 0 { std::thread::sleep(Duration::from_micros(500)); } // Now, run it again with a faster task let flag = Arc::new(AtomicBool::new(true)); // Set it high, we know it shouldn't ever really be this high let counter = Arc::new(AtomicUsize::new(10_000)); let interval = Arc::new(AtomicUsize::new(10_000)); { let flag = flag.clone(); let counter = counter.clone(); rt.spawn(async move { iter(flag, counter, false) }); } // Now, hammer the injection queue until the interval reaches the expected range. let mut n = 0; loop { let curr = interval.load(Relaxed); if curr <= 1_000 && curr > 32 { n += 1; } else { n = 0; } if n == 3 { break; } if Arc::strong_count(&interval) <= 5_000 { let counter = counter.clone(); let interval = interval.clone(); rt.spawn(async move { let prev = counter.swap(0, Relaxed); interval.store(prev, Relaxed); }); } std::thread::yield_now(); } flag.store(false, Relaxed); } fn rt() -> runtime::Runtime { runtime::Runtime::new().unwrap() } #[cfg(tokio_unstable)] mod unstable { use super::*; #[test] fn test_disable_lifo_slot() { use std::sync::mpsc::{channel, RecvTimeoutError}; let rt = runtime::Builder::new_multi_thread() .disable_lifo_slot() .worker_threads(2) .build() .unwrap(); // Spawn a background thread to poke the runtime periodically. // // This is necessary because we may end up triggering the issue in: // <https://github.com/tokio-rs/tokio/issues/4730> // // Spawning a task will wake up the second worker, which will then steal // the task. However, the steal will fail if the task is in the LIFO // slot, because the LIFO slot cannot be stolen. // // Note that this only happens rarely. Most of the time, this thread is // not necessary. let (kill_bg_thread, recv) = channel::<()>(); let handle = rt.handle().clone(); let bg_thread = std::thread::spawn(move || { let one_sec = std::time::Duration::from_secs(1); while recv.recv_timeout(one_sec) == Err(RecvTimeoutError::Timeout) { handle.spawn(async {}); } }); rt.block_on(async { tokio::spawn(async { // Spawn another task and block the thread until completion. If the LIFO slot // is used then the test doesn't complete. futures::executor::block_on(tokio::spawn(async {})).unwrap(); }) .await .unwrap(); }); drop(kill_bg_thread); bg_thread.join().unwrap(); } #[test] fn runtime_id_is_same() { let rt = rt(); let handle1 = rt.handle(); let handle2 = rt.handle(); assert_eq!(handle1.id(), handle2.id()); } #[test] fn runtime_ids_different() { let rt1 = rt(); let rt2 = rt(); assert_ne!(rt1.handle().id(), rt2.handle().id()); } }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/tcp_peek.rs
tokio/tests/tcp_peek.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi"), not(miri)))] // Wasi doesn't support bind // No `socket` on miri. use tokio::io::AsyncReadExt; use tokio::net::TcpStream; use tokio_test::assert_ok; use std::thread; use std::{io::Write, net}; #[tokio::test] async fn peek() { let listener = net::TcpListener::bind("127.0.0.1:0").unwrap(); let addr = listener.local_addr().unwrap(); let t = thread::spawn(move || assert_ok!(listener.accept()).0); let left = net::TcpStream::connect(addr).unwrap(); left.set_nonblocking(true).unwrap(); let mut right = t.join().unwrap(); right.set_nonblocking(true).unwrap(); let _ = right.write(&[1, 2, 3, 4]).unwrap(); let mut left: TcpStream = left.try_into().unwrap(); let mut buf = [0u8; 16]; let n = assert_ok!(left.peek(&mut buf).await); assert_eq!([1, 2, 3, 4], buf[..n]); let n = assert_ok!(left.read(&mut buf).await); assert_eq!([1, 2, 3, 4], buf[..n]); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/tracing_time.rs
tokio/tests/tracing_time.rs
//! Tests for time resource instrumentation. //! //! These tests ensure that the instrumentation for tokio //! synchronization primitives is correct. #![warn(rust_2018_idioms)] #![cfg(all(tokio_unstable, feature = "tracing", target_has_atomic = "64"))] use std::time::Duration; use tracing_mock::{expect, subscriber}; #[tokio::test] async fn test_sleep_creates_span() { let sleep_span_id = expect::id(); let sleep_span = expect::span() .with_id(sleep_span_id.clone()) .named("runtime.resource") .with_target("tokio::time::sleep"); let state_update = expect::event() .with_target("runtime::resource::state_update") .with_fields( expect::field("duration") // FIXME(hds): This value isn't stable and doesn't seem to make sense. We're not // going to test on it until the resource instrumentation has been // refactored and we know that we're getting a stable value here. //.with_value(&(7_u64 + 1)) .and(expect::field("duration.op").with_value(&"override")), ); let async_op_span_id = expect::id(); let async_op_span = expect::span() .with_id(async_op_span_id.clone()) .named("runtime.resource.async_op") .with_target("tokio::time::sleep"); let async_op_poll_span = expect::span() .named("runtime.resource.async_op.poll") .with_target("tokio::time::sleep"); let (subscriber, handle) = subscriber::mock() .new_span(sleep_span.clone().with_ancestry(expect::is_explicit_root())) .enter(sleep_span.clone()) .event(state_update) .new_span( async_op_span .clone() .with_ancestry(expect::has_contextual_parent(&sleep_span_id)) .with_fields(expect::field("source").with_value(&"Sleep::new_timeout")), ) .exit(sleep_span.clone()) .enter(async_op_span.clone()) .new_span( async_op_poll_span .clone() .with_ancestry(expect::has_contextual_parent(&async_op_span_id)), ) .exit(async_op_span.clone()) .drop_span(async_op_span) .drop_span(async_op_poll_span) .drop_span(sleep_span) .run_with_handle(); { let _guard = tracing::subscriber::set_default(subscriber); _ = tokio::time::sleep(Duration::from_millis(7)); } handle.assert_finished(); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/fs_symlink_file_windows.rs
tokio/tests/fs_symlink_file_windows.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // WASI does not support all fs operations #![cfg(windows)] use tempfile::tempdir; use tokio::fs; #[tokio::test] async fn symlink_file_windows() { let dir = tempdir().unwrap(); let source_path = dir.path().join("foo.txt"); let dest_path = dir.path().join("bar.txt"); fs::write(&source_path, b"Hello File!").await.unwrap(); fs::symlink_file(&source_path, &dest_path).await.unwrap(); fs::write(&source_path, b"new data!").await.unwrap(); let from = fs::read(&source_path).await.unwrap(); let to = fs::read(&dest_path).await.unwrap(); assert_eq!(from, to); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/tracing_task.rs
tokio/tests/tracing_task.rs
//! Tests for task instrumentation. //! //! These tests ensure that the instrumentation for task spawning and task //! lifecycles is correct. #![warn(rust_2018_idioms)] #![cfg(all(tokio_unstable, feature = "tracing", target_has_atomic = "64"))] use std::{mem, time::Duration}; use tokio::task; use tracing_mock::{expect, span::NewSpan, subscriber}; #[tokio::test] async fn task_spawn_creates_span() { let task_span = expect::span() .named("runtime.spawn") .with_target("tokio::task"); let (subscriber, handle) = subscriber::mock() .new_span(&task_span) .enter(&task_span) .exit(&task_span) // The task span is entered once more when it gets dropped .enter(&task_span) .exit(&task_span) .drop_span(task_span) .run_with_handle(); { let _guard = tracing::subscriber::set_default(subscriber); tokio::spawn(futures::future::ready(())) .await .expect("failed to await join handle"); } handle.assert_finished(); } #[tokio::test] async fn task_spawn_loc_file_recorded() { let task_span = expect::span() .named("runtime.spawn") .with_target("tokio::task") .with_fields(expect::field("loc.file").with_value(&file!())); let (subscriber, handle) = subscriber::mock().new_span(task_span).run_with_handle(); { let _guard = tracing::subscriber::set_default(subscriber); tokio::spawn(futures::future::ready(())) .await .expect("failed to await join handle"); } handle.assert_finished(); } #[tokio::test] async fn task_builder_name_recorded() { let task_span = expect_task_named("test-task"); let (subscriber, handle) = subscriber::mock().new_span(task_span).run_with_handle(); { let _guard = tracing::subscriber::set_default(subscriber); task::Builder::new() .name("test-task") .spawn(futures::future::ready(())) .unwrap() .await .expect("failed to await join handle"); } handle.assert_finished(); } #[tokio::test] async fn task_builder_loc_file_recorded() { let task_span = expect::span() .named("runtime.spawn") .with_target("tokio::task") .with_fields(expect::field("loc.file").with_value(&file!())); let (subscriber, handle) = subscriber::mock().new_span(task_span).run_with_handle(); { let _guard = tracing::subscriber::set_default(subscriber); task::Builder::new() .spawn(futures::future::ready(())) .unwrap() .await .expect("failed to await join handle"); } handle.assert_finished(); } #[tokio::test] async fn task_spawn_sizes_recorded() { let future = futures::future::ready(()); let size = mem::size_of_val(&future) as u64; let task_span = expect::span() .named("runtime.spawn") .with_target("tokio::task") // TODO(hds): check that original_size.bytes is NOT recorded when this can be done in // tracing-mock without listing every other field. .with_fields(expect::field("size.bytes").with_value(&size)); let (subscriber, handle) = subscriber::mock().new_span(task_span).run_with_handle(); { let _guard = tracing::subscriber::set_default(subscriber); task::Builder::new() .spawn(future) .unwrap() .await .expect("failed to await join handle"); } handle.assert_finished(); } #[tokio::test] async fn task_big_spawn_sizes_recorded() { let future = { async fn big<const N: usize>() { let mut a = [0_u8; N]; for (idx, item) in a.iter_mut().enumerate() { *item = (idx % 256) as u8; } tokio::time::sleep(Duration::from_millis(10)).await; for (idx, item) in a.iter_mut().enumerate() { assert_eq!(*item, (idx % 256) as u8); } } // This is larger than the release auto-boxing threshold big::<20_000>() }; fn boxed_size<T>(_: &T) -> usize { mem::size_of::<Box<T>>() } let size = mem::size_of_val(&future) as u64; let boxed_size = boxed_size(&future); let task_span = expect::span() .named("runtime.spawn") .with_target("tokio::task") .with_fields( expect::field("size.bytes") .with_value(&boxed_size) .and(expect::field("original_size.bytes").with_value(&size)), ); let (subscriber, handle) = subscriber::mock().new_span(task_span).run_with_handle(); { let _guard = tracing::subscriber::set_default(subscriber); task::Builder::new() .spawn(future) .unwrap() .await .expect("failed to await join handle"); } handle.assert_finished(); } /// Expect a task with name /// /// This is a convenience function to create the expectation for a new task /// with the `task.name` field set to the provided name. fn expect_task_named(name: &str) -> NewSpan { expect::span() .named("runtime.spawn") .with_target("tokio::task") .with_fields( expect::field("task.name").with_value(&tracing::field::debug(format_args!("{name}"))), ) }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_panic.rs
tokio/tests/io_panic.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // Wasi does not support panic recovery #![cfg(panic = "unwind")] use std::task::{Context, Poll}; use std::{error::Error, pin::Pin}; use tokio::io::{self, split, AsyncRead, AsyncWrite, ReadBuf}; mod support { pub mod panic; } use support::panic::test_panic; struct RW; impl AsyncRead for RW { fn poll_read( self: Pin<&mut Self>, _cx: &mut Context<'_>, buf: &mut ReadBuf<'_>, ) -> Poll<io::Result<()>> { buf.put_slice(b"z"); Poll::Ready(Ok(())) } } impl AsyncWrite for RW { fn poll_write( self: Pin<&mut Self>, _cx: &mut Context<'_>, _buf: &[u8], ) -> Poll<Result<usize, io::Error>> { Poll::Ready(Ok(1)) } fn poll_flush(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<Result<(), io::Error>> { Poll::Ready(Ok(())) } fn poll_shutdown(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<Result<(), io::Error>> { Poll::Ready(Ok(())) } } #[cfg(unix)] mod unix { use std::os::unix::prelude::{AsRawFd, RawFd}; pub struct MockFd; impl AsRawFd for MockFd { fn as_raw_fd(&self) -> RawFd { 0 } } } #[test] fn read_buf_initialize_unfilled_to_panic_caller() -> Result<(), Box<dyn Error>> { let panic_location_file = test_panic(|| { let mut buffer = Vec::<u8>::new(); let mut read_buf = ReadBuf::new(&mut buffer); read_buf.initialize_unfilled_to(2); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] fn read_buf_advance_panic_caller() -> Result<(), Box<dyn Error>> { let panic_location_file = test_panic(|| { let mut buffer = Vec::<u8>::new(); let mut read_buf = ReadBuf::new(&mut buffer); read_buf.advance(2); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] fn read_buf_set_filled_panic_caller() -> Result<(), Box<dyn Error>> { let panic_location_file = test_panic(|| { let mut buffer = Vec::<u8>::new(); let mut read_buf = ReadBuf::new(&mut buffer); read_buf.set_filled(2); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] fn read_buf_put_slice_panic_caller() -> Result<(), Box<dyn Error>> { let panic_location_file = test_panic(|| { let mut buffer = Vec::<u8>::new(); let mut read_buf = ReadBuf::new(&mut buffer); let new_slice = [0x40_u8, 0x41_u8]; read_buf.put_slice(&new_slice); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] fn unsplit_panic_caller() -> Result<(), Box<dyn Error>> { let panic_location_file = test_panic(|| { let (r1, _w1) = split(RW); let (_r2, w2) = split(RW); r1.unsplit(w2); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] #[cfg(unix)] fn async_fd_new_panic_caller() -> Result<(), Box<dyn Error>> { use tokio::io::unix::AsyncFd; use tokio::runtime::Builder; let panic_location_file = test_panic(|| { // Runtime without `enable_io` so it has no IO driver set. let rt = Builder::new_current_thread().build().unwrap(); rt.block_on(async { let fd = unix::MockFd; let _ = AsyncFd::new(fd); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] #[cfg(unix)] fn async_fd_with_interest_panic_caller() -> Result<(), Box<dyn Error>> { use tokio::io::unix::AsyncFd; use tokio::io::Interest; use tokio::runtime::Builder; let panic_location_file = test_panic(|| { // Runtime without `enable_io` so it has no IO driver set. let rt = Builder::new_current_thread().build().unwrap(); rt.block_on(async { let fd = unix::MockFd; let _ = AsyncFd::with_interest(fd, Interest::READABLE); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] #[cfg(unix)] fn async_fd_try_new_panic_caller() -> Result<(), Box<dyn Error>> { use tokio::io::unix::AsyncFd; use tokio::runtime::Builder; let panic_location_file = test_panic(|| { // Runtime without `enable_io` so it has no IO driver set. let rt = Builder::new_current_thread().build().unwrap(); rt.block_on(async { let fd = unix::MockFd; let _ = AsyncFd::try_new(fd); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] #[cfg(unix)] fn async_fd_try_with_interest_panic_caller() -> Result<(), Box<dyn Error>> { use tokio::io::unix::AsyncFd; use tokio::io::Interest; use tokio::runtime::Builder; let panic_location_file = test_panic(|| { // Runtime without `enable_io` so it has no IO driver set. let rt = Builder::new_current_thread().build().unwrap(); rt.block_on(async { let fd = unix::MockFd; let _ = AsyncFd::try_with_interest(fd, Interest::READABLE); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/signal_ctrl_c.rs
tokio/tests/signal_ctrl_c.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] #![cfg(unix)] #![cfg(not(miri))] // No `sigaction` on Miri mod support { pub mod signal; } use support::signal::send_signal; use tokio::signal; use tokio_test::assert_ok; #[tokio::test] async fn ctrl_c() { let ctrl_c = signal::ctrl_c(); tokio::spawn(async { send_signal(libc::SIGINT); }); assert_ok!(ctrl_c.await); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/fs_rename.rs
tokio/tests/fs_rename.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // WASI does not support all fs operations use tempfile::tempdir; use tokio::fs; #[tokio::test] async fn rename_file() { let temp_dir = tempdir().unwrap(); let file_path = temp_dir.path().join("a.txt"); fs::write(&file_path, b"Hello File!").await.unwrap(); assert!(fs::try_exists(&file_path).await.unwrap()); let new_file_path = temp_dir.path().join("b.txt"); fs::rename(&file_path, &new_file_path).await.unwrap(); assert!(fs::try_exists(new_file_path).await.unwrap()); // original file should no longer exist match fs::try_exists(file_path).await { Ok(exists) => assert!(!exists), Err(_) => println!("ignored try_exists error after rename"), }; }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/tracing_sync.rs
tokio/tests/tracing_sync.rs
//! Tests for sync instrumentation. //! //! These tests ensure that the instrumentation for tokio //! synchronization primitives is correct. #![warn(rust_2018_idioms)] #![cfg(all(tokio_unstable, feature = "tracing", target_has_atomic = "64"))] use tokio::sync; use tracing_mock::{expect, subscriber}; #[tokio::test] async fn test_barrier_creates_span() { let barrier_span = expect::span() .named("runtime.resource") .with_target("tokio::sync::barrier"); let size_event = expect::event() .with_target("runtime::resource::state_update") .with_fields(expect::field("size").with_value(&1_u64)); let arrived_event = expect::event() .with_target("runtime::resource::state_update") .with_fields(expect::field("arrived").with_value(&0_i64)); let (subscriber, handle) = subscriber::mock() .new_span( barrier_span .clone() .with_ancestry(expect::is_explicit_root()), ) .enter(&barrier_span) .event(size_event) .event(arrived_event) .exit(&barrier_span) .drop_span(&barrier_span) .run_with_handle(); { let _guard = tracing::subscriber::set_default(subscriber); let _ = sync::Barrier::new(1); } handle.assert_finished(); } #[tokio::test] async fn test_mutex_creates_span() { let mutex_span = expect::span() .named("runtime.resource") .with_target("tokio::sync::mutex"); let locked_event = expect::event() .with_target("runtime::resource::state_update") .with_fields(expect::field("locked").with_value(&false)); let batch_semaphore_span = expect::span() .named("runtime.resource") .with_target("tokio::sync::batch_semaphore"); let batch_semaphore_permits_event = expect::event() .with_target("runtime::resource::state_update") .with_fields(expect::field("permits").with_value(&1u64)) .with_fields(expect::field("permits.op").with_value(&"override")); let (subscriber, handle) = subscriber::mock() .new_span(mutex_span.clone().with_ancestry(expect::is_explicit_root())) .enter(&mutex_span) .event(locked_event) .new_span( batch_semaphore_span .clone() .with_ancestry(expect::is_explicit_root()), ) .enter(&batch_semaphore_span) .event(batch_semaphore_permits_event) .exit(&batch_semaphore_span) .exit(&mutex_span) .drop_span(&mutex_span) .drop_span(&batch_semaphore_span) .run_with_handle(); { let _guard = tracing::subscriber::set_default(subscriber); let _ = sync::Mutex::new(true); } handle.assert_finished(); } #[tokio::test] async fn test_oneshot_creates_span() { let oneshot_span_id = expect::id(); let oneshot_span = expect::span() .with_id(oneshot_span_id.clone()) .named("runtime.resource") .with_target("tokio::sync::oneshot"); let initial_tx_dropped_event = expect::event() .with_target("runtime::resource::state_update") .with_fields(expect::field("tx_dropped").with_value(&false)) .with_fields(expect::field("tx_dropped.op").with_value(&"override")); let final_tx_dropped_event = expect::event() .with_target("runtime::resource::state_update") .with_fields(expect::field("tx_dropped").with_value(&true)) .with_fields(expect::field("tx_dropped.op").with_value(&"override")); let initial_rx_dropped_event = expect::event() .with_target("runtime::resource::state_update") .with_fields(expect::field("rx_dropped").with_value(&false)) .with_fields(expect::field("rx_dropped.op").with_value(&"override")); let final_rx_dropped_event = expect::event() .with_target("runtime::resource::state_update") .with_fields(expect::field("rx_dropped").with_value(&true)) .with_fields(expect::field("rx_dropped.op").with_value(&"override")); let value_sent_event = expect::event() .with_target("runtime::resource::state_update") .with_fields(expect::field("value_sent").with_value(&false)) .with_fields(expect::field("value_sent.op").with_value(&"override")); let value_received_event = expect::event() .with_target("runtime::resource::state_update") .with_fields(expect::field("value_received").with_value(&false)) .with_fields(expect::field("value_received.op").with_value(&"override")); let async_op_span_id = expect::id(); let async_op_span = expect::span() .with_id(async_op_span_id.clone()) .named("runtime.resource.async_op") .with_target("tokio::sync::oneshot"); let async_op_poll_span = expect::span() .named("runtime.resource.async_op.poll") .with_target("tokio::sync::oneshot"); let (subscriber, handle) = subscriber::mock() .new_span( oneshot_span .clone() .with_ancestry(expect::is_explicit_root()), ) .enter(&oneshot_span) .event(initial_tx_dropped_event) .exit(&oneshot_span) .enter(&oneshot_span) .event(initial_rx_dropped_event) .exit(&oneshot_span) .enter(&oneshot_span) .event(value_sent_event) .exit(&oneshot_span) .enter(&oneshot_span) .event(value_received_event) .exit(&oneshot_span) .enter(&oneshot_span) .new_span( async_op_span .clone() .with_ancestry(expect::has_contextual_parent(&oneshot_span_id)), ) .exit(&oneshot_span) .enter(&async_op_span) .new_span( async_op_poll_span .clone() .with_ancestry(expect::has_contextual_parent(&async_op_span_id)), ) .exit(&async_op_span) .enter(&oneshot_span) .event(final_tx_dropped_event) .exit(&oneshot_span) .enter(&oneshot_span) .event(final_rx_dropped_event) .exit(&oneshot_span) .drop_span(oneshot_span) .drop_span(async_op_span) .drop_span(&async_op_poll_span) .run_with_handle(); { let _guard = tracing::subscriber::set_default(subscriber); let _ = sync::oneshot::channel::<bool>(); } handle.assert_finished(); } #[tokio::test] async fn test_rwlock_creates_span() { let rwlock_span = expect::span() .named("runtime.resource") .with_target("tokio::sync::rwlock"); let max_readers_event = expect::event() .with_target("runtime::resource::state_update") .with_fields(expect::field("max_readers").with_value(&0x1FFFFFFF_u64)); let write_locked_event = expect::event() .with_target("runtime::resource::state_update") .with_fields(expect::field("write_locked").with_value(&false)); let current_readers_event = expect::event() .with_target("runtime::resource::state_update") .with_fields(expect::field("current_readers").with_value(&0_i64)); let batch_semaphore_span = expect::span() .named("runtime.resource") .with_target("tokio::sync::batch_semaphore"); let batch_semaphore_permits_event = expect::event() .with_target("runtime::resource::state_update") .with_fields(expect::field("permits").with_value(&1u64)) .with_fields(expect::field("permits.op").with_value(&"override")); let (subscriber, handle) = subscriber::mock() .new_span( rwlock_span .clone() .with_ancestry(expect::is_explicit_root()), ) .enter(rwlock_span.clone()) .event(max_readers_event) .event(write_locked_event) .event(current_readers_event) .exit(rwlock_span.clone()) .enter(rwlock_span.clone()) .new_span(batch_semaphore_span.clone()) .enter(batch_semaphore_span.clone()) .event(batch_semaphore_permits_event) .exit(batch_semaphore_span.clone()) .exit(rwlock_span.clone()) .drop_span(rwlock_span) .drop_span(batch_semaphore_span) .run_with_handle(); { let _guard = tracing::subscriber::set_default(subscriber); let _ = sync::RwLock::new(true); } handle.assert_finished(); } #[tokio::test] async fn test_semaphore_creates_span() { let semaphore_span = expect::span() .named("runtime.resource") .with_target("tokio::sync::semaphore"); let batch_semaphore_span = expect::span() .named("runtime.resource") .with_target("tokio::sync::batch_semaphore"); let batch_semaphore_permits_event = expect::event() .with_target("runtime::resource::state_update") .with_fields(expect::field("permits").with_value(&1u64)) .with_fields(expect::field("permits.op").with_value(&"override")); let (subscriber, handle) = subscriber::mock() .new_span( semaphore_span .clone() .with_ancestry(expect::is_explicit_root()), ) .enter(semaphore_span.clone()) .new_span(batch_semaphore_span.clone()) .enter(batch_semaphore_span.clone()) .event(batch_semaphore_permits_event) .exit(batch_semaphore_span.clone()) .exit(semaphore_span.clone()) .drop_span(semaphore_span) .drop_span(batch_semaphore_span) .run_with_handle(); { let _guard = tracing::subscriber::set_default(subscriber); let _ = sync::Semaphore::new(1); } handle.assert_finished(); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_buf_reader.rs
tokio/tests/io_buf_reader.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] // https://github.com/rust-lang/futures-rs/blob/1803948ff091b4eabf7f3bf39e16bbbdefca5cc8/futures/tests/io_buf_reader.rs use futures::task::{noop_waker_ref, Context, Poll}; use std::cmp; use std::io::{self, Cursor}; use std::pin::Pin; use tokio::io::{ AsyncBufRead, AsyncBufReadExt, AsyncRead, AsyncReadExt, AsyncSeek, AsyncSeekExt, AsyncWriteExt, BufReader, ReadBuf, SeekFrom, }; use tokio_test::task::spawn; use tokio_test::{assert_pending, assert_ready}; macro_rules! run_fill_buf { ($reader:expr) => {{ let mut cx = Context::from_waker(noop_waker_ref()); loop { if let Poll::Ready(x) = Pin::new(&mut $reader).poll_fill_buf(&mut cx) { break x; } } }}; } struct MaybePending<'a> { inner: &'a [u8], ready_read: bool, ready_fill_buf: bool, } impl<'a> MaybePending<'a> { fn new(inner: &'a [u8]) -> Self { Self { inner, ready_read: false, ready_fill_buf: false, } } } impl AsyncRead for MaybePending<'_> { fn poll_read( mut self: Pin<&mut Self>, cx: &mut Context<'_>, buf: &mut ReadBuf<'_>, ) -> Poll<io::Result<()>> { if self.ready_read { self.ready_read = false; Pin::new(&mut self.inner).poll_read(cx, buf) } else { self.ready_read = true; cx.waker().wake_by_ref(); Poll::Pending } } } impl AsyncBufRead for MaybePending<'_> { fn poll_fill_buf(mut self: Pin<&mut Self>, _: &mut Context<'_>) -> Poll<io::Result<&[u8]>> { if self.ready_fill_buf { self.ready_fill_buf = false; if self.inner.is_empty() { return Poll::Ready(Ok(&[])); } let len = cmp::min(2, self.inner.len()); Poll::Ready(Ok(&self.inner[0..len])) } else { self.ready_fill_buf = true; Poll::Pending } } fn consume(mut self: Pin<&mut Self>, amt: usize) { self.inner = &self.inner[amt..]; } } #[tokio::test] async fn test_buffered_reader() { let inner: &[u8] = &[5, 6, 7, 0, 1, 2, 3, 4]; let mut reader = BufReader::with_capacity(2, inner); let mut buf = [0, 0, 0]; let nread = reader.read(&mut buf).await.unwrap(); assert_eq!(nread, 3); assert_eq!(buf, [5, 6, 7]); assert_eq!(reader.buffer(), []); let mut buf = [0, 0]; let nread = reader.read(&mut buf).await.unwrap(); assert_eq!(nread, 2); assert_eq!(buf, [0, 1]); assert_eq!(reader.buffer(), []); let mut buf = [0]; let nread = reader.read(&mut buf).await.unwrap(); assert_eq!(nread, 1); assert_eq!(buf, [2]); assert_eq!(reader.buffer(), [3]); let mut buf = [0, 0, 0]; let nread = reader.read(&mut buf).await.unwrap(); assert_eq!(nread, 1); assert_eq!(buf, [3, 0, 0]); assert_eq!(reader.buffer(), []); let nread = reader.read(&mut buf).await.unwrap(); assert_eq!(nread, 1); assert_eq!(buf, [4, 0, 0]); assert_eq!(reader.buffer(), []); assert_eq!(reader.read(&mut buf).await.unwrap(), 0); } #[tokio::test] async fn test_buffered_reader_seek() { let inner: &[u8] = &[5, 6, 7, 0, 1, 2, 3, 4]; let mut reader = BufReader::with_capacity(2, Cursor::new(inner)); assert_eq!(reader.seek(SeekFrom::Start(3)).await.unwrap(), 3); assert_eq!(run_fill_buf!(reader).unwrap(), &[0, 1][..]); assert!(reader.seek(SeekFrom::Current(i64::MIN)).await.is_err()); assert_eq!(run_fill_buf!(reader).unwrap(), &[0, 1][..]); assert_eq!(reader.seek(SeekFrom::Current(1)).await.unwrap(), 4); assert_eq!(run_fill_buf!(reader).unwrap(), &[1, 2][..]); Pin::new(&mut reader).consume(1); assert_eq!(reader.seek(SeekFrom::Current(-2)).await.unwrap(), 3); } #[tokio::test] async fn test_buffered_reader_seek_underflow() { // gimmick reader that yields its position modulo 256 for each byte struct PositionReader { pos: u64, } impl AsyncRead for PositionReader { fn poll_read( mut self: Pin<&mut Self>, _: &mut Context<'_>, buf: &mut ReadBuf<'_>, ) -> Poll<io::Result<()>> { let b = buf.initialize_unfilled(); let len = b.len(); for x in b { *x = self.pos as u8; self.pos = self.pos.wrapping_add(1); } buf.advance(len); Poll::Ready(Ok(())) } } impl AsyncSeek for PositionReader { fn start_seek(mut self: Pin<&mut Self>, pos: SeekFrom) -> io::Result<()> { match pos { SeekFrom::Start(n) => { self.pos = n; } SeekFrom::Current(n) => { self.pos = self.pos.wrapping_add(n as u64); } SeekFrom::End(n) => { self.pos = u64::MAX.wrapping_add(n as u64); } } Ok(()) } fn poll_complete(self: Pin<&mut Self>, _: &mut Context<'_>) -> Poll<io::Result<u64>> { Poll::Ready(Ok(self.pos)) } } let mut reader = BufReader::with_capacity(5, PositionReader { pos: 0 }); assert_eq!(run_fill_buf!(reader).unwrap(), &[0, 1, 2, 3, 4][..]); assert_eq!(reader.seek(SeekFrom::End(-5)).await.unwrap(), u64::MAX - 5); assert_eq!(run_fill_buf!(reader).unwrap().len(), 5); // the following seek will require two underlying seeks let expected = 9_223_372_036_854_775_802; assert_eq!( reader.seek(SeekFrom::Current(i64::MIN)).await.unwrap(), expected ); assert_eq!(run_fill_buf!(reader).unwrap().len(), 5); // seeking to 0 should empty the buffer. assert_eq!(reader.seek(SeekFrom::Current(0)).await.unwrap(), expected); assert_eq!(reader.get_ref().pos, expected); } #[tokio::test] async fn test_short_reads() { /// A dummy reader intended at testing short-reads propagation. struct ShortReader { lengths: Vec<usize>, } impl AsyncRead for ShortReader { fn poll_read( mut self: Pin<&mut Self>, _: &mut Context<'_>, buf: &mut ReadBuf<'_>, ) -> Poll<io::Result<()>> { if !self.lengths.is_empty() { buf.advance(self.lengths.remove(0)); } Poll::Ready(Ok(())) } } let inner = ShortReader { lengths: vec![0, 1, 2, 0, 1, 0], }; let mut reader = BufReader::new(inner); let mut buf = [0, 0]; assert_eq!(reader.read(&mut buf).await.unwrap(), 0); assert_eq!(reader.read(&mut buf).await.unwrap(), 1); assert_eq!(reader.read(&mut buf).await.unwrap(), 2); assert_eq!(reader.read(&mut buf).await.unwrap(), 0); assert_eq!(reader.read(&mut buf).await.unwrap(), 1); assert_eq!(reader.read(&mut buf).await.unwrap(), 0); assert_eq!(reader.read(&mut buf).await.unwrap(), 0); } #[tokio::test] async fn maybe_pending() { let inner: &[u8] = &[5, 6, 7, 0, 1, 2, 3, 4]; let mut reader = BufReader::with_capacity(2, MaybePending::new(inner)); let mut buf = [0, 0, 0]; let nread = reader.read(&mut buf).await.unwrap(); assert_eq!(nread, 3); assert_eq!(buf, [5, 6, 7]); assert_eq!(reader.buffer(), []); let mut buf = [0, 0]; let nread = reader.read(&mut buf).await.unwrap(); assert_eq!(nread, 2); assert_eq!(buf, [0, 1]); assert_eq!(reader.buffer(), []); let mut buf = [0]; let nread = reader.read(&mut buf).await.unwrap(); assert_eq!(nread, 1); assert_eq!(buf, [2]); assert_eq!(reader.buffer(), [3]); let mut buf = [0, 0, 0]; let nread = reader.read(&mut buf).await.unwrap(); assert_eq!(nread, 1); assert_eq!(buf, [3, 0, 0]); assert_eq!(reader.buffer(), []); let nread = reader.read(&mut buf).await.unwrap(); assert_eq!(nread, 1); assert_eq!(buf, [4, 0, 0]); assert_eq!(reader.buffer(), []); assert_eq!(reader.read(&mut buf).await.unwrap(), 0); } #[tokio::test] async fn maybe_pending_buf_read() { let inner = MaybePending::new(&[0, 1, 2, 3, 1, 0]); let mut reader = BufReader::with_capacity(2, inner); let mut v = Vec::new(); reader.read_until(3, &mut v).await.unwrap(); assert_eq!(v, [0, 1, 2, 3]); v.clear(); reader.read_until(1, &mut v).await.unwrap(); assert_eq!(v, [1]); v.clear(); reader.read_until(8, &mut v).await.unwrap(); assert_eq!(v, [0]); v.clear(); reader.read_until(9, &mut v).await.unwrap(); assert_eq!(v, []); } // https://github.com/rust-lang/futures-rs/pull/1573#discussion_r281162309 #[tokio::test] async fn maybe_pending_seek() { struct MaybePendingSeek<'a> { inner: Cursor<&'a [u8]>, ready: bool, seek_res: Option<io::Result<()>>, } impl<'a> MaybePendingSeek<'a> { fn new(inner: &'a [u8]) -> Self { Self { inner: Cursor::new(inner), ready: true, seek_res: None, } } } impl AsyncRead for MaybePendingSeek<'_> { fn poll_read( mut self: Pin<&mut Self>, cx: &mut Context<'_>, buf: &mut ReadBuf<'_>, ) -> Poll<io::Result<()>> { Pin::new(&mut self.inner).poll_read(cx, buf) } } impl AsyncBufRead for MaybePendingSeek<'_> { fn poll_fill_buf( mut self: Pin<&mut Self>, cx: &mut Context<'_>, ) -> Poll<io::Result<&[u8]>> { let this: *mut Self = &mut *self as *mut _; Pin::new(&mut unsafe { &mut *this }.inner).poll_fill_buf(cx) } fn consume(mut self: Pin<&mut Self>, amt: usize) { Pin::new(&mut self.inner).consume(amt) } } impl AsyncSeek for MaybePendingSeek<'_> { fn start_seek(mut self: Pin<&mut Self>, pos: SeekFrom) -> io::Result<()> { self.seek_res = Some(Pin::new(&mut self.inner).start_seek(pos)); Ok(()) } fn poll_complete(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<io::Result<u64>> { if self.ready { self.ready = false; self.seek_res.take().unwrap_or(Ok(()))?; Pin::new(&mut self.inner).poll_complete(cx) } else { self.ready = true; cx.waker().wake_by_ref(); Poll::Pending } } } let inner: &[u8] = &[5, 6, 7, 0, 1, 2, 3, 4]; let mut reader = BufReader::with_capacity(2, MaybePendingSeek::new(inner)); assert_eq!(reader.seek(SeekFrom::Current(3)).await.unwrap(), 3); assert_eq!(run_fill_buf!(reader).unwrap(), &[0, 1][..]); assert!(reader.seek(SeekFrom::Current(i64::MIN)).await.is_err()); assert_eq!(run_fill_buf!(reader).unwrap(), &[0, 1][..]); assert_eq!(reader.seek(SeekFrom::Current(1)).await.unwrap(), 4); assert_eq!(run_fill_buf!(reader).unwrap(), &[1, 2][..]); Pin::new(&mut reader).consume(1); assert_eq!(reader.seek(SeekFrom::Current(-2)).await.unwrap(), 3); } // This tests the AsyncBufReadExt::fill_buf wrapper. #[tokio::test] async fn test_fill_buf_wrapper() { let (mut write, read) = tokio::io::duplex(16); let mut read = BufReader::new(read); write.write_all(b"hello world").await.unwrap(); assert_eq!(read.fill_buf().await.unwrap(), b"hello world"); read.consume(b"hello ".len()); assert_eq!(read.fill_buf().await.unwrap(), b"world"); assert_eq!(read.fill_buf().await.unwrap(), b"world"); read.consume(b"world".len()); let mut fill = spawn(read.fill_buf()); assert_pending!(fill.poll()); write.write_all(b"foo bar").await.unwrap(); assert_eq!(assert_ready!(fill.poll()).unwrap(), b"foo bar"); drop(fill); drop(write); assert_eq!(read.fill_buf().await.unwrap(), b"foo bar"); read.consume(b"foo bar".len()); assert_eq!(read.fill_buf().await.unwrap(), b""); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/join_handle_panic.rs
tokio/tests/join_handle_panic.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // Wasi doesn't support panic recovery #![cfg(panic = "unwind")] struct PanicsOnDrop; impl Drop for PanicsOnDrop { fn drop(&mut self) { panic!("I told you so"); } } #[tokio::test] async fn test_panics_do_not_propagate_when_dropping_join_handle() { let join_handle = tokio::spawn(async move { PanicsOnDrop }); // only drop the JoinHandle when the task has completed // (which is difficult to synchronize precisely) tokio::time::sleep(std::time::Duration::from_millis(3)).await; drop(join_handle); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/buffered.rs
tokio/tests/buffered.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // Wasi does not support bind() use tokio::net::TcpListener; use tokio_test::assert_ok; use std::io::prelude::*; use std::net::TcpStream; use std::thread; #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` on miri. async fn echo_server() { const N: usize = 1024; let srv = assert_ok!(TcpListener::bind("127.0.0.1:0").await); let addr = assert_ok!(srv.local_addr()); let msg = "foo bar baz"; let t = thread::spawn(move || { let mut s = assert_ok!(TcpStream::connect(addr)); let t2 = thread::spawn(move || { let mut s = assert_ok!(TcpStream::connect(addr)); let mut b = vec![0; msg.len() * N]; assert_ok!(s.read_exact(&mut b)); b }); let mut expected = Vec::<u8>::new(); for _i in 0..N { expected.extend(msg.as_bytes()); let res = assert_ok!(s.write(msg.as_bytes())); assert_eq!(res, msg.len()); } (expected, t2) }); let (mut a, _) = assert_ok!(srv.accept().await); let (mut b, _) = assert_ok!(srv.accept().await); let n = assert_ok!(tokio::io::copy(&mut a, &mut b).await); let (expected, t2) = t.join().unwrap(); let actual = t2.join().unwrap(); assert!(expected == actual); assert_eq!(n, msg.len() as u64 * 1024); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_async_read.rs
tokio/tests/io_async_read.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use tokio::io::AsyncRead; #[test] fn assert_obj_safe() { fn _assert<T>() {} _assert::<Box<dyn AsyncRead>>(); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_write_int.rs
tokio/tests/io_write_int.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use tokio::io::{AsyncWrite, AsyncWriteExt}; use std::io; use std::pin::Pin; use std::task::{Context, Poll}; #[tokio::test] async fn write_int_should_err_if_write_count_0() { struct Wr {} impl AsyncWrite for Wr { fn poll_write( self: Pin<&mut Self>, _cx: &mut Context<'_>, _buf: &[u8], ) -> Poll<io::Result<usize>> { Ok(0).into() } fn poll_flush(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<io::Result<()>> { Ok(()).into() } fn poll_shutdown(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<io::Result<()>> { Ok(()).into() } } let mut wr = Wr {}; // should be ok just to test these 2, other cases actually expanded by same macro. assert!(wr.write_i8(0).await.is_err()); assert!(wr.write_i32(12).await.is_err()); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/dump.rs
tokio/tests/dump.rs
#![cfg(all( tokio_unstable, feature = "taskdump", target_os = "linux", any(target_arch = "aarch64", target_arch = "x86", target_arch = "x86_64") ))] use std::hint::black_box; use tokio::runtime::{self, Handle}; #[inline(never)] async fn a() { black_box(b()).await } #[inline(never)] async fn b() { black_box(c()).await } #[inline(never)] async fn c() { loop { black_box(tokio::task::yield_now()).await } } #[test] fn current_thread() { let rt = runtime::Builder::new_current_thread() .enable_all() .build() .unwrap(); async fn dump() { let handle = Handle::current(); let dump = handle.dump().await; let tasks: Vec<_> = dump.tasks().iter().collect(); assert_eq!(tasks.len(), 3); for task in tasks { let id = task.id(); let trace = task.trace().to_string(); eprintln!("\n\n{id}:\n{trace}\n\n"); assert!(trace.contains("dump::a")); assert!(trace.contains("dump::b")); assert!(trace.contains("dump::c")); assert!(trace.contains("tokio::task::yield_now")); } } rt.block_on(async { tokio::select!( biased; _ = tokio::spawn(a()) => {}, _ = tokio::spawn(a()) => {}, _ = tokio::spawn(a()) => {}, _ = dump() => {}, ); }); } #[test] fn multi_thread() { let rt = runtime::Builder::new_multi_thread() .enable_all() .worker_threads(3) .build() .unwrap(); async fn dump() { let handle = Handle::current(); let dump = handle.dump().await; let tasks: Vec<_> = dump.tasks().iter().collect(); assert_eq!(tasks.len(), 3); for task in tasks { let id = task.id(); let trace = task.trace().to_string(); eprintln!("\n\n{id}:\n{trace}\n\n"); assert!(trace.contains("dump::a")); assert!(trace.contains("dump::b")); assert!(trace.contains("dump::c")); assert!(trace.contains("tokio::task::yield_now")); } } rt.block_on(async { tokio::select!( biased; _ = tokio::spawn(a()) => {}, _ = tokio::spawn(a()) => {}, _ = tokio::spawn(a()) => {}, _ = dump() => {}, ); }); } /// Regression tests for #6035. /// /// These tests ensure that dumping will not deadlock if a future completes /// during a trace. mod future_completes_during_trace { use super::*; use core::future::{poll_fn, Future}; /// A future that completes only during a trace. fn complete_during_trace() -> impl Future<Output = ()> + Send { use std::task::Poll; poll_fn(|cx| { if Handle::is_tracing() { Poll::Ready(()) } else { cx.waker().wake_by_ref(); Poll::Pending } }) } #[test] fn current_thread() { let rt = runtime::Builder::new_current_thread() .enable_all() .build() .unwrap(); async fn dump() { let handle = Handle::current(); let _dump = handle.dump().await; } rt.block_on(async { let _ = tokio::join!(tokio::spawn(complete_during_trace()), dump()); }); } #[test] fn multi_thread() { let rt = runtime::Builder::new_multi_thread() .enable_all() .build() .unwrap(); async fn dump() { let handle = Handle::current(); let _dump = handle.dump().await; tokio::task::yield_now().await; } rt.block_on(async { let _ = tokio::join!(tokio::spawn(complete_during_trace()), dump()); }); } } /// Regression test for #6051. /// /// This test ensures that tasks notified outside of a worker will not be /// traced, since doing so will un-set their notified bit prior to them being /// run and panic. #[test] fn notified_during_tracing() { let rt = runtime::Builder::new_multi_thread() .enable_all() .worker_threads(3) .build() .unwrap(); let timeout = async { tokio::time::sleep(tokio::time::Duration::from_secs(1)).await; }; let timer = rt.spawn(async { loop { tokio::time::sleep(tokio::time::Duration::from_nanos(1)).await; } }); let dump = async { loop { let handle = Handle::current(); let _dump = handle.dump().await; } }; rt.block_on(async { tokio::select!( biased; _ = timeout => {}, _ = timer => {}, _ = dump => {}, ); }); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/task_join_set.rs
tokio/tests/task_join_set.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use futures::future::{pending, FutureExt}; use std::panic; use tokio::sync::oneshot; use tokio::task::{JoinSet, LocalSet}; use tokio::time::Duration; fn rt() -> tokio::runtime::Runtime { tokio::runtime::Builder::new_current_thread() .build() .unwrap() } // Spawn `N` tasks that return their index (`i`). fn spawn_index_tasks(set: &mut JoinSet<usize>, n: usize, on: Option<&LocalSet>) { for i in 0..n { let rc = std::rc::Rc::new(i); match on { None => set.spawn_local(async move { *rc }), Some(local) => set.spawn_local_on(async move { *rc }, local), }; } } // Spawn `N` “pending” tasks that own a `oneshot::Sender`. // When the task is aborted the sender is dropped, which is observed // via the returned `Receiver`s. fn spawn_pending_tasks( set: &mut JoinSet<()>, receivers: &mut Vec<oneshot::Receiver<()>>, n: usize, on: Option<&LocalSet>, ) { for _ in 0..n { let (tx, rx) = oneshot::channel::<()>(); receivers.push(rx); let fut = async move { pending::<()>().await; drop(tx); }; match on { None => set.spawn_local(fut), Some(local) => set.spawn_local_on(fut, local), }; } } // Await every task in a JoinSet and assert every task returns its own index. async fn drain_joinset_and_assert(mut set: JoinSet<usize>, n: usize) { let mut seen = vec![false; n]; while let Some(res) = set.join_next().await { let idx = res.expect("task panicked"); seen[idx] = true; } assert!(seen.into_iter().all(|b| b)); assert!(set.is_empty()); } // Await every receiver and assert they all return `Err` because the // corresponding sender (inside an aborted task) was dropped. async fn await_receivers_and_assert(receivers: Vec<oneshot::Receiver<()>>) { for rx in receivers { assert!( rx.await.is_err(), "the task should have been aborted and the sender dropped" ); } } #[tokio::test(start_paused = true)] async fn test_with_sleep() { let mut set = JoinSet::new(); for i in 0..10 { set.spawn(async move { i }); assert_eq!(set.len(), 1 + i); } set.detach_all(); assert_eq!(set.len(), 0); assert!(set.join_next().await.is_none()); for i in 0..10 { set.spawn(async move { tokio::time::sleep(Duration::from_secs(i as u64)).await; i }); assert_eq!(set.len(), 1 + i); } let mut seen = [false; 10]; while let Some(res) = set.join_next().await.transpose().unwrap() { seen[res] = true; } for was_seen in &seen { assert!(was_seen); } assert!(set.join_next().await.is_none()); // Do it again. for i in 0..10 { set.spawn(async move { tokio::time::sleep(Duration::from_secs(i as u64)).await; i }); } let mut seen = [false; 10]; while let Some(res) = set.join_next().await.transpose().unwrap() { seen[res] = true; } for was_seen in &seen { assert!(was_seen); } assert!(set.join_next().await.is_none()); } #[tokio::test] async fn test_abort_on_drop() { let mut set = JoinSet::new(); let mut recvs = Vec::new(); for _ in 0..16 { let (send, recv) = oneshot::channel::<()>(); recvs.push(recv); set.spawn(async { // This task will never complete on its own. futures::future::pending::<()>().await; drop(send); }); } drop(set); for recv in recvs { // The task is aborted soon and we will receive an error. assert!(recv.await.is_err()); } } #[tokio::test] async fn alternating() { let mut set = JoinSet::new(); assert_eq!(set.len(), 0); set.spawn(async {}); assert_eq!(set.len(), 1); set.spawn(async {}); assert_eq!(set.len(), 2); for _ in 0..16 { let () = set.join_next().await.unwrap().unwrap(); assert_eq!(set.len(), 1); set.spawn(async {}); assert_eq!(set.len(), 2); } } #[tokio::test(start_paused = true)] async fn abort_tasks() { let mut set = JoinSet::new(); let mut num_canceled = 0; let mut num_completed = 0; for i in 0..16 { let abort = set.spawn(async move { tokio::time::sleep(Duration::from_secs(i as u64)).await; i }); if i % 2 != 0 { // abort odd-numbered tasks. abort.abort(); } } loop { match set.join_next().await { Some(Ok(res)) => { num_completed += 1; assert_eq!(res % 2, 0); } Some(Err(e)) => { assert!(e.is_cancelled()); num_canceled += 1; } None => break, } } assert_eq!(num_canceled, 8); assert_eq!(num_completed, 8); } #[test] fn runtime_gone() { let mut set = JoinSet::new(); { let rt = rt(); set.spawn_on(async { 1 }, rt.handle()); drop(rt); } assert!(rt() .block_on(set.join_next()) .unwrap() .unwrap_err() .is_cancelled()); } #[tokio::test] async fn join_all() { let mut set: JoinSet<i32> = JoinSet::new(); for _ in 0..5 { set.spawn(async { 1 }); } let res: Vec<i32> = set.join_all().await; assert_eq!(res.len(), 5); for itm in res.into_iter() { assert_eq!(itm, 1) } } #[cfg(panic = "unwind")] #[tokio::test(start_paused = true)] async fn task_panics() { let mut set: JoinSet<()> = JoinSet::new(); let (tx, mut rx) = oneshot::channel(); assert_eq!(set.len(), 0); set.spawn(async move { tokio::time::sleep(Duration::from_secs(2)).await; tx.send(()).unwrap(); }); assert_eq!(set.len(), 1); set.spawn(async { tokio::time::sleep(Duration::from_secs(1)).await; panic!(); }); assert_eq!(set.len(), 2); let panic = tokio::spawn(set.join_all()).await.unwrap_err(); assert!(rx.try_recv().is_err()); assert!(panic.is_panic()); } #[tokio::test(start_paused = true)] async fn abort_all() { let mut set: JoinSet<()> = JoinSet::new(); for _ in 0..5 { set.spawn(futures::future::pending()); } for _ in 0..5 { set.spawn(async { tokio::time::sleep(Duration::from_secs(1)).await; }); } // The join set will now have 5 pending tasks and 5 ready tasks. tokio::time::sleep(Duration::from_secs(2)).await; set.abort_all(); assert_eq!(set.len(), 10); let mut count = 0; while let Some(res) = set.join_next().await { if let Err(err) = res { assert!(err.is_cancelled()); } count += 1; } assert_eq!(count, 10); assert_eq!(set.len(), 0); } // This ensures that `join_next` works correctly when the coop budget is // exhausted. #[tokio::test(flavor = "current_thread")] async fn join_set_coop() { // Large enough to trigger coop. const TASK_NUM: u32 = 1000; static SEM: tokio::sync::Semaphore = tokio::sync::Semaphore::const_new(0); let mut set = JoinSet::new(); for _ in 0..TASK_NUM { set.spawn(async { SEM.add_permits(1); }); } // Wait for all tasks to complete. // // Since this is a `current_thread` runtime, there's no race condition // between the last permit being added and the task completing. let _ = SEM.acquire_many(TASK_NUM).await.unwrap(); let mut count = 0; let mut coop_count = 0; loop { match set.join_next().now_or_never() { Some(Some(Ok(()))) => {} Some(Some(Err(err))) => panic!("failed: {err}"), None => { coop_count += 1; tokio::task::yield_now().await; continue; } Some(None) => break, } count += 1; } assert!(coop_count >= 1); assert_eq!(count, TASK_NUM); } #[tokio::test(flavor = "current_thread")] async fn try_join_next() { const TASK_NUM: u32 = 1000; let (send, recv) = tokio::sync::watch::channel(()); let mut set = JoinSet::new(); for _ in 0..TASK_NUM { let mut recv = recv.clone(); set.spawn(async move { recv.changed().await.unwrap() }); } drop(recv); assert!(set.try_join_next().is_none()); send.send_replace(()); send.closed().await; let mut count = 0; loop { match set.try_join_next() { Some(Ok(())) => { count += 1; } Some(Err(err)) => panic!("failed: {err}"), None => { break; } } } assert_eq!(count, TASK_NUM); } #[cfg(tokio_unstable)] #[tokio::test(flavor = "current_thread")] async fn try_join_next_with_id() { const TASK_NUM: u32 = 1000; let (send, recv) = tokio::sync::watch::channel(()); let mut set = JoinSet::new(); let mut spawned = std::collections::HashSet::with_capacity(TASK_NUM as usize); for _ in 0..TASK_NUM { let mut recv = recv.clone(); let handle = set.spawn(async move { recv.changed().await.unwrap() }); spawned.insert(handle.id()); } drop(recv); assert!(set.try_join_next_with_id().is_none()); send.send_replace(()); send.closed().await; let mut count = 0; let mut joined = std::collections::HashSet::with_capacity(TASK_NUM as usize); loop { match set.try_join_next_with_id() { Some(Ok((id, ()))) => { count += 1; joined.insert(id); } Some(Err(err)) => panic!("failed: {err}"), None => { break; } } } assert_eq!(count, TASK_NUM); assert_eq!(joined, spawned); } #[tokio::test] async fn extend() { let mut set: JoinSet<_> = (0..5).map(|i| async move { i }).collect(); set.extend((5..10).map(|i| async move { i })); let mut seen = [false; 10]; while let Some(res) = set.join_next().await { let idx = res.unwrap(); seen[idx] = true; } for s in &seen { assert!(s); } } mod spawn_local { use super::*; #[test] #[should_panic( expected = "`spawn_local` called from outside of a `task::LocalSet` or `runtime::LocalRuntime`" )] fn panic_outside_any_runtime() { let mut set = JoinSet::new(); set.spawn_local(async {}); } #[tokio::test(flavor = "multi_thread")] #[should_panic( expected = "`spawn_local` called from outside of a `task::LocalSet` or `runtime::LocalRuntime`" )] async fn panic_in_multi_thread_runtime() { let mut set = JoinSet::new(); set.spawn_local(async {}); } #[cfg(tokio_unstable)] mod local_runtime { use super::*; /// Spawn several tasks, and then join all tasks. #[tokio::test(flavor = "local")] async fn spawn_then_join_next() { const N: usize = 8; let mut set = JoinSet::new(); spawn_index_tasks(&mut set, N, None); assert!(set.try_join_next().is_none()); drain_joinset_and_assert(set, N).await; } /// Spawn several pending-forever tasks, and then shutdown the [`JoinSet`]. #[tokio::test(flavor = "local")] async fn spawn_then_shutdown() { const N: usize = 8; let mut set = JoinSet::new(); let mut receivers = Vec::new(); spawn_pending_tasks(&mut set, &mut receivers, N, None); assert!(set.try_join_next().is_none()); set.shutdown().await; assert!(set.is_empty()); await_receivers_and_assert(receivers).await; } /// Spawn several pending-forever tasks, and then drop the [`JoinSet`]. #[tokio::test(flavor = "local")] async fn spawn_then_drop() { const N: usize = 8; let mut set = JoinSet::new(); let mut receivers = Vec::new(); spawn_pending_tasks(&mut set, &mut receivers, N, None); assert!(set.try_join_next().is_none()); drop(set); await_receivers_and_assert(receivers).await; } } mod local_set { use super::*; /// Spawn several tasks, and then join all tasks. #[tokio::test(flavor = "current_thread")] async fn spawn_then_join_next() { const N: usize = 8; let local = LocalSet::new(); local .run_until(async move { let mut set = JoinSet::new(); spawn_index_tasks(&mut set, N, None); drain_joinset_and_assert(set, N).await; }) .await; } /// Spawn several pending-forever tasks, and then shutdown the [`JoinSet`]. #[tokio::test(flavor = "current_thread")] async fn spawn_then_shutdown() { const N: usize = 8; let local = LocalSet::new(); local .run_until(async { let mut set = JoinSet::new(); let mut receivers = Vec::new(); spawn_pending_tasks(&mut set, &mut receivers, N, None); assert!(set.try_join_next().is_none()); set.shutdown().await; assert!(set.is_empty()); await_receivers_and_assert(receivers).await; }) .await; } /// Spawn several pending-forever tasks, and then drop the [`JoinSet`]. #[tokio::test(flavor = "current_thread")] async fn spawn_then_drop() { const N: usize = 8; let local = LocalSet::new(); local .run_until(async { let mut set = JoinSet::new(); let mut receivers = Vec::new(); spawn_pending_tasks(&mut set, &mut receivers, N, None); assert!(set.try_join_next().is_none()); drop(set); await_receivers_and_assert(receivers).await; }) .await; } } } mod spawn_local_on { use super::*; #[cfg(tokio_unstable)] mod local_runtime { use super::*; /// Spawn several tasks, and then join all tasks. #[tokio::test(flavor = "local")] async fn spawn_then_join_next() { const N: usize = 8; let local = LocalSet::new(); let mut set = JoinSet::new(); spawn_index_tasks(&mut set, N, Some(&local)); assert!(set.try_join_next().is_none()); local .run_until(async move { drain_joinset_and_assert(set, N).await; }) .await; } } mod local_set { use super::*; /// Spawn several tasks, and then join all tasks. #[tokio::test(flavor = "current_thread")] async fn spawn_then_join_next() { const N: usize = 8; let local = LocalSet::new(); let mut pending_set = JoinSet::new(); spawn_index_tasks(&mut pending_set, N, Some(&local)); assert!(pending_set.try_join_next().is_none()); local .run_until(async move { drain_joinset_and_assert(pending_set, N).await; }) .await; } /// Spawn several pending-forever tasks, and then shutdown the [`JoinSet`]. #[tokio::test(flavor = "current_thread")] async fn spawn_then_shutdown() { const N: usize = 8; let local = LocalSet::new(); let mut set = JoinSet::new(); let mut receivers = Vec::new(); spawn_pending_tasks(&mut set, &mut receivers, N, Some(&local)); assert!(set.try_join_next().is_none()); local .run_until(async move { set.shutdown().await; assert!(set.is_empty()); await_receivers_and_assert(receivers).await; }) .await; } /// Spawn several pending-forever tasks and then drop the [`JoinSet`] /// before the `LocalSet` is driven and while the `LocalSet` is already driven. #[tokio::test(flavor = "current_thread")] async fn spawn_then_drop() { const N: usize = 8; { let local = LocalSet::new(); let mut set = JoinSet::new(); let mut receivers = Vec::new(); spawn_pending_tasks(&mut set, &mut receivers, N, Some(&local)); assert!(set.try_join_next().is_none()); drop(set); local .run_until(async move { await_receivers_and_assert(receivers).await; }) .await; } { let local = LocalSet::new(); let mut set = JoinSet::new(); let mut receivers = Vec::new(); spawn_pending_tasks(&mut set, &mut receivers, N, Some(&local)); assert!(set.try_join_next().is_none()); local .run_until(async move { drop(set); await_receivers_and_assert(receivers).await; }) .await; } } } }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_buf_writer.rs
tokio/tests/io_buf_writer.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] // https://github.com/rust-lang/futures-rs/blob/1803948ff091b4eabf7f3bf39e16bbbdefca5cc8/futures/tests/io_buf_writer.rs use futures::task::{Context, Poll}; use std::io::{self, Cursor}; use std::pin::Pin; use tokio::io::{AsyncSeek, AsyncSeekExt, AsyncWrite, AsyncWriteExt, BufWriter, SeekFrom}; use futures::future; use tokio_test::assert_ok; use std::cmp; use std::io::IoSlice; mod support { pub(crate) mod io_vec; } use support::io_vec::IoBufs; struct MaybePending { inner: Vec<u8>, ready: bool, } impl MaybePending { fn new(inner: Vec<u8>) -> Self { Self { inner, ready: false, } } } impl AsyncWrite for MaybePending { fn poll_write( mut self: Pin<&mut Self>, cx: &mut Context<'_>, buf: &[u8], ) -> Poll<io::Result<usize>> { if self.ready { self.ready = false; Pin::new(&mut self.inner).poll_write(cx, buf) } else { self.ready = true; cx.waker().wake_by_ref(); Poll::Pending } } fn poll_flush(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<io::Result<()>> { Pin::new(&mut self.inner).poll_flush(cx) } fn poll_shutdown(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<io::Result<()>> { Pin::new(&mut self.inner).poll_shutdown(cx) } } async fn write_vectored<W>(writer: &mut W, bufs: &[IoSlice<'_>]) -> io::Result<usize> where W: AsyncWrite + Unpin, { let mut writer = Pin::new(writer); future::poll_fn(|cx| writer.as_mut().poll_write_vectored(cx, bufs)).await } #[tokio::test] async fn buf_writer() { let mut writer = BufWriter::with_capacity(2, Vec::new()); assert_eq!(writer.write(&[0, 1]).await.unwrap(), 2); assert_eq!(writer.buffer(), []); assert_eq!(*writer.get_ref(), [0, 1]); assert_eq!(writer.write(&[2]).await.unwrap(), 1); assert_eq!(writer.buffer(), [2]); assert_eq!(*writer.get_ref(), [0, 1]); assert_eq!(writer.write(&[3]).await.unwrap(), 1); assert_eq!(writer.buffer(), [2, 3]); assert_eq!(*writer.get_ref(), [0, 1]); writer.flush().await.unwrap(); assert_eq!(writer.buffer(), []); assert_eq!(*writer.get_ref(), [0, 1, 2, 3]); assert_eq!(writer.write(&[4]).await.unwrap(), 1); assert_eq!(writer.write(&[5]).await.unwrap(), 1); assert_eq!(writer.buffer(), [4, 5]); assert_eq!(*writer.get_ref(), [0, 1, 2, 3]); assert_eq!(writer.write(&[6]).await.unwrap(), 1); assert_eq!(writer.buffer(), [6]); assert_eq!(*writer.get_ref(), [0, 1, 2, 3, 4, 5]); assert_eq!(writer.write(&[7, 8]).await.unwrap(), 2); assert_eq!(writer.buffer(), []); assert_eq!(*writer.get_ref(), [0, 1, 2, 3, 4, 5, 6, 7, 8]); assert_eq!(writer.write(&[9, 10, 11]).await.unwrap(), 3); assert_eq!(writer.buffer(), []); assert_eq!(*writer.get_ref(), [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11]); writer.flush().await.unwrap(); assert_eq!(writer.buffer(), []); assert_eq!(*writer.get_ref(), [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11]); } #[tokio::test] async fn buf_writer_inner_flushes() { let mut w = BufWriter::with_capacity(3, Vec::new()); assert_eq!(w.write(&[0, 1]).await.unwrap(), 2); assert_eq!(*w.get_ref(), []); w.flush().await.unwrap(); let w = w.into_inner(); assert_eq!(w, [0, 1]); } #[tokio::test] async fn buf_writer_seek() { let mut w = BufWriter::with_capacity(3, Cursor::new(Vec::new())); w.write_all(&[0, 1, 2, 3, 4, 5]).await.unwrap(); w.write_all(&[6, 7]).await.unwrap(); assert_eq!(w.seek(SeekFrom::Current(0)).await.unwrap(), 8); assert_eq!(&w.get_ref().get_ref()[..], &[0, 1, 2, 3, 4, 5, 6, 7][..]); assert_eq!(w.seek(SeekFrom::Start(2)).await.unwrap(), 2); w.write_all(&[8, 9]).await.unwrap(); w.flush().await.unwrap(); assert_eq!(&w.into_inner().into_inner()[..], &[0, 1, 8, 9, 4, 5, 6, 7]); } #[tokio::test] async fn maybe_pending_buf_writer() { let mut writer = BufWriter::with_capacity(2, MaybePending::new(Vec::new())); assert_eq!(writer.write(&[0, 1]).await.unwrap(), 2); assert_eq!(writer.buffer(), []); assert_eq!(&writer.get_ref().inner, &[0, 1]); assert_eq!(writer.write(&[2]).await.unwrap(), 1); assert_eq!(writer.buffer(), [2]); assert_eq!(&writer.get_ref().inner, &[0, 1]); assert_eq!(writer.write(&[3]).await.unwrap(), 1); assert_eq!(writer.buffer(), [2, 3]); assert_eq!(&writer.get_ref().inner, &[0, 1]); writer.flush().await.unwrap(); assert_eq!(writer.buffer(), []); assert_eq!(&writer.get_ref().inner, &[0, 1, 2, 3]); assert_eq!(writer.write(&[4]).await.unwrap(), 1); assert_eq!(writer.write(&[5]).await.unwrap(), 1); assert_eq!(writer.buffer(), [4, 5]); assert_eq!(&writer.get_ref().inner, &[0, 1, 2, 3]); assert_eq!(writer.write(&[6]).await.unwrap(), 1); assert_eq!(writer.buffer(), [6]); assert_eq!(writer.get_ref().inner, &[0, 1, 2, 3, 4, 5]); assert_eq!(writer.write(&[7, 8]).await.unwrap(), 2); assert_eq!(writer.buffer(), []); assert_eq!(writer.get_ref().inner, &[0, 1, 2, 3, 4, 5, 6, 7, 8]); assert_eq!(writer.write(&[9, 10, 11]).await.unwrap(), 3); assert_eq!(writer.buffer(), []); assert_eq!( writer.get_ref().inner, &[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11] ); writer.flush().await.unwrap(); assert_eq!(writer.buffer(), []); assert_eq!( &writer.get_ref().inner, &[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11] ); } #[tokio::test] async fn maybe_pending_buf_writer_inner_flushes() { let mut w = BufWriter::with_capacity(3, MaybePending::new(Vec::new())); assert_eq!(w.write(&[0, 1]).await.unwrap(), 2); assert_eq!(&w.get_ref().inner, &[]); w.flush().await.unwrap(); let w = w.into_inner().inner; assert_eq!(w, [0, 1]); } #[tokio::test] async fn maybe_pending_buf_writer_seek() { struct MaybePendingSeek { inner: Cursor<Vec<u8>>, ready_write: bool, ready_seek: bool, seek_res: Option<io::Result<()>>, } impl MaybePendingSeek { fn new(inner: Vec<u8>) -> Self { Self { inner: Cursor::new(inner), ready_write: false, ready_seek: false, seek_res: None, } } } impl AsyncWrite for MaybePendingSeek { fn poll_write( mut self: Pin<&mut Self>, cx: &mut Context<'_>, buf: &[u8], ) -> Poll<io::Result<usize>> { if self.ready_write { self.ready_write = false; Pin::new(&mut self.inner).poll_write(cx, buf) } else { self.ready_write = true; cx.waker().wake_by_ref(); Poll::Pending } } fn poll_flush(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<io::Result<()>> { Pin::new(&mut self.inner).poll_flush(cx) } fn poll_shutdown(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<io::Result<()>> { Pin::new(&mut self.inner).poll_shutdown(cx) } } impl AsyncSeek for MaybePendingSeek { fn start_seek(mut self: Pin<&mut Self>, pos: SeekFrom) -> io::Result<()> { self.seek_res = Some(Pin::new(&mut self.inner).start_seek(pos)); Ok(()) } fn poll_complete(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<io::Result<u64>> { if self.ready_seek { self.ready_seek = false; self.seek_res.take().unwrap_or(Ok(()))?; Pin::new(&mut self.inner).poll_complete(cx) } else { self.ready_seek = true; cx.waker().wake_by_ref(); Poll::Pending } } } let mut w = BufWriter::with_capacity(3, MaybePendingSeek::new(Vec::new())); w.write_all(&[0, 1, 2, 3, 4, 5]).await.unwrap(); w.write_all(&[6, 7]).await.unwrap(); assert_eq!(w.seek(SeekFrom::Current(0)).await.unwrap(), 8); assert_eq!( &w.get_ref().inner.get_ref()[..], &[0, 1, 2, 3, 4, 5, 6, 7][..] ); assert_eq!(w.seek(SeekFrom::Start(2)).await.unwrap(), 2); w.write_all(&[8, 9]).await.unwrap(); w.flush().await.unwrap(); assert_eq!( &w.into_inner().inner.into_inner()[..], &[0, 1, 8, 9, 4, 5, 6, 7] ); } struct MockWriter { data: Vec<u8>, write_len: usize, vectored: bool, } impl MockWriter { fn new(write_len: usize) -> Self { MockWriter { data: Vec::new(), write_len, vectored: false, } } fn vectored(write_len: usize) -> Self { MockWriter { data: Vec::new(), write_len, vectored: true, } } fn write_up_to(&mut self, buf: &[u8], limit: usize) -> usize { let len = cmp::min(buf.len(), limit); self.data.extend_from_slice(&buf[..len]); len } } impl AsyncWrite for MockWriter { fn poll_write( self: Pin<&mut Self>, _: &mut Context<'_>, buf: &[u8], ) -> Poll<Result<usize, io::Error>> { let this = self.get_mut(); let n = this.write_up_to(buf, this.write_len); Ok(n).into() } fn poll_write_vectored( self: Pin<&mut Self>, _: &mut Context<'_>, bufs: &[IoSlice<'_>], ) -> Poll<Result<usize, io::Error>> { let this = self.get_mut(); let mut total_written = 0; for buf in bufs { let n = this.write_up_to(buf, this.write_len - total_written); total_written += n; if total_written == this.write_len { break; } } Ok(total_written).into() } fn is_write_vectored(&self) -> bool { self.vectored } fn poll_flush(self: Pin<&mut Self>, _: &mut Context<'_>) -> Poll<Result<(), io::Error>> { Ok(()).into() } fn poll_shutdown(self: Pin<&mut Self>, _: &mut Context<'_>) -> Poll<Result<(), io::Error>> { Ok(()).into() } } #[tokio::test] async fn write_vectored_empty_on_non_vectored() { let mut w = BufWriter::new(MockWriter::new(4)); let n = assert_ok!(write_vectored(&mut w, &[]).await); assert_eq!(n, 0); let io_vec = [IoSlice::new(&[]); 3]; let n = assert_ok!(write_vectored(&mut w, &io_vec).await); assert_eq!(n, 0); assert_ok!(w.flush().await); assert!(w.get_ref().data.is_empty()); } #[tokio::test] async fn write_vectored_empty_on_vectored() { let mut w = BufWriter::new(MockWriter::vectored(4)); let n = assert_ok!(write_vectored(&mut w, &[]).await); assert_eq!(n, 0); let io_vec = [IoSlice::new(&[]); 3]; let n = assert_ok!(write_vectored(&mut w, &io_vec).await); assert_eq!(n, 0); assert_ok!(w.flush().await); assert!(w.get_ref().data.is_empty()); } #[tokio::test] async fn write_vectored_basic_on_non_vectored() { let msg = b"foo bar baz"; let bufs = [ IoSlice::new(&msg[0..4]), IoSlice::new(&msg[4..8]), IoSlice::new(&msg[8..]), ]; let mut w = BufWriter::new(MockWriter::new(4)); let n = assert_ok!(write_vectored(&mut w, &bufs).await); assert_eq!(n, msg.len()); assert!(w.buffer() == &msg[..]); assert_ok!(w.flush().await); assert_eq!(w.get_ref().data, msg); } #[tokio::test] async fn write_vectored_basic_on_vectored() { let msg = b"foo bar baz"; let bufs = [ IoSlice::new(&msg[0..4]), IoSlice::new(&msg[4..8]), IoSlice::new(&msg[8..]), ]; let mut w = BufWriter::new(MockWriter::vectored(4)); let n = assert_ok!(write_vectored(&mut w, &bufs).await); assert_eq!(n, msg.len()); assert!(w.buffer() == &msg[..]); assert_ok!(w.flush().await); assert_eq!(w.get_ref().data, msg); } #[tokio::test] async fn write_vectored_large_total_on_non_vectored() { let msg = b"foo bar baz"; let mut bufs = [ IoSlice::new(&msg[0..4]), IoSlice::new(&msg[4..8]), IoSlice::new(&msg[8..]), ]; let io_vec = IoBufs::new(&mut bufs); let mut w = BufWriter::with_capacity(8, MockWriter::new(4)); let n = assert_ok!(write_vectored(&mut w, &io_vec).await); assert_eq!(n, 8); assert!(w.buffer() == &msg[..8]); let io_vec = io_vec.advance(n); let n = assert_ok!(write_vectored(&mut w, &io_vec).await); assert_eq!(n, 3); assert!(w.get_ref().data.as_slice() == &msg[..8]); assert!(w.buffer() == &msg[8..]); } #[tokio::test] async fn write_vectored_large_total_on_vectored() { let msg = b"foo bar baz"; let mut bufs = [ IoSlice::new(&msg[0..4]), IoSlice::new(&msg[4..8]), IoSlice::new(&msg[8..]), ]; let io_vec = IoBufs::new(&mut bufs); let mut w = BufWriter::with_capacity(8, MockWriter::vectored(10)); let n = assert_ok!(write_vectored(&mut w, &io_vec).await); assert_eq!(n, 10); assert!(w.buffer().is_empty()); let io_vec = io_vec.advance(n); let n = assert_ok!(write_vectored(&mut w, &io_vec).await); assert_eq!(n, 1); assert!(w.get_ref().data.as_slice() == &msg[..10]); assert!(w.buffer() == &msg[10..]); } struct VectoredWriteHarness { writer: BufWriter<MockWriter>, buf_capacity: usize, } impl VectoredWriteHarness { fn new(buf_capacity: usize) -> Self { VectoredWriteHarness { writer: BufWriter::with_capacity(buf_capacity, MockWriter::new(4)), buf_capacity, } } fn with_vectored_backend(buf_capacity: usize) -> Self { VectoredWriteHarness { writer: BufWriter::with_capacity(buf_capacity, MockWriter::vectored(4)), buf_capacity, } } async fn write_all<'a, 'b>(&mut self, mut io_vec: IoBufs<'a, 'b>) -> usize { let mut total_written = 0; while !io_vec.is_empty() { let n = assert_ok!(write_vectored(&mut self.writer, &io_vec).await); assert!(n != 0); assert!(self.writer.buffer().len() <= self.buf_capacity); total_written += n; io_vec = io_vec.advance(n); } total_written } async fn flush(&mut self) -> &[u8] { assert_ok!(self.writer.flush().await); &self.writer.get_ref().data } } #[tokio::test] async fn write_vectored_odd_on_non_vectored() { let msg = b"foo bar baz"; let mut bufs = [ IoSlice::new(&msg[0..4]), IoSlice::new(&[]), IoSlice::new(&msg[4..9]), IoSlice::new(&msg[9..]), ]; let mut h = VectoredWriteHarness::new(8); let bytes_written = h.write_all(IoBufs::new(&mut bufs)).await; assert_eq!(bytes_written, msg.len()); assert_eq!(h.flush().await, msg); } #[tokio::test] async fn write_vectored_odd_on_vectored() { let msg = b"foo bar baz"; let mut bufs = [ IoSlice::new(&msg[0..4]), IoSlice::new(&[]), IoSlice::new(&msg[4..9]), IoSlice::new(&msg[9..]), ]; let mut h = VectoredWriteHarness::with_vectored_backend(8); let bytes_written = h.write_all(IoBufs::new(&mut bufs)).await; assert_eq!(bytes_written, msg.len()); assert_eq!(h.flush().await, msg); } #[tokio::test] async fn write_vectored_large_slice_on_non_vectored() { let msg = b"foo bar baz"; let mut bufs = [ IoSlice::new(&[]), IoSlice::new(&msg[..9]), IoSlice::new(&msg[9..]), ]; let mut h = VectoredWriteHarness::new(8); let bytes_written = h.write_all(IoBufs::new(&mut bufs)).await; assert_eq!(bytes_written, msg.len()); assert_eq!(h.flush().await, msg); } #[tokio::test] async fn write_vectored_large_slice_on_vectored() { let msg = b"foo bar baz"; let mut bufs = [ IoSlice::new(&[]), IoSlice::new(&msg[..9]), IoSlice::new(&msg[9..]), ]; let mut h = VectoredWriteHarness::with_vectored_backend(8); let bytes_written = h.write_all(IoBufs::new(&mut bufs)).await; assert_eq!(bytes_written, msg.len()); assert_eq!(h.flush().await, msg); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/udp.rs
tokio/tests/udp.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi"), not(miri)))] // Wasi does not support bind or UDP // No `socket` on miri. use std::future::poll_fn; use std::io; use std::sync::Arc; use tokio::{io::ReadBuf, net::UdpSocket}; use tokio_test::assert_ok; const MSG: &[u8] = b"hello"; const MSG_LEN: usize = MSG.len(); #[tokio::test] async fn send_recv() -> std::io::Result<()> { let sender = UdpSocket::bind("127.0.0.1:0").await?; let receiver = UdpSocket::bind("127.0.0.1:0").await?; sender.connect(receiver.local_addr()?).await?; receiver.connect(sender.local_addr()?).await?; sender.send(MSG).await?; let mut recv_buf = [0u8; 32]; let len = receiver.recv(&mut recv_buf[..]).await?; assert_eq!(&recv_buf[..len], MSG); Ok(()) } #[tokio::test] async fn send_recv_poll() -> std::io::Result<()> { let sender = UdpSocket::bind("127.0.0.1:0").await?; let receiver = UdpSocket::bind("127.0.0.1:0").await?; sender.connect(receiver.local_addr()?).await?; receiver.connect(sender.local_addr()?).await?; poll_fn(|cx| sender.poll_send(cx, MSG)).await?; let mut recv_buf = [0u8; 32]; let mut read = ReadBuf::new(&mut recv_buf); poll_fn(|cx| receiver.poll_recv(cx, &mut read)).await?; assert_eq!(read.filled(), MSG); Ok(()) } #[tokio::test] async fn send_to_recv_from() -> std::io::Result<()> { let sender = UdpSocket::bind("127.0.0.1:0").await?; let receiver = UdpSocket::bind("127.0.0.1:0").await?; let receiver_addr = receiver.local_addr()?; sender.send_to(MSG, &receiver_addr).await?; let mut recv_buf = [0u8; 32]; let (len, addr) = receiver.recv_from(&mut recv_buf[..]).await?; assert_eq!(&recv_buf[..len], MSG); assert_eq!(addr, sender.local_addr()?); Ok(()) } #[tokio::test] async fn send_to_recv_from_poll() -> std::io::Result<()> { let sender = UdpSocket::bind("127.0.0.1:0").await?; let receiver = UdpSocket::bind("127.0.0.1:0").await?; let receiver_addr = receiver.local_addr()?; poll_fn(|cx| sender.poll_send_to(cx, MSG, receiver_addr)).await?; let mut recv_buf = [0u8; 32]; let mut read = ReadBuf::new(&mut recv_buf); let addr = poll_fn(|cx| receiver.poll_recv_from(cx, &mut read)).await?; assert_eq!(read.filled(), MSG); assert_eq!(addr, sender.local_addr()?); Ok(()) } #[tokio::test] async fn send_to_peek_from() -> std::io::Result<()> { let sender = UdpSocket::bind("127.0.0.1:0").await?; let receiver = UdpSocket::bind("127.0.0.1:0").await?; let receiver_addr = receiver.local_addr()?; poll_fn(|cx| sender.poll_send_to(cx, MSG, receiver_addr)).await?; // peek let mut recv_buf = [0u8; 32]; let (n, addr) = receiver.peek_from(&mut recv_buf).await?; assert_eq!(&recv_buf[..n], MSG); assert_eq!(addr, sender.local_addr()?); // peek let mut recv_buf = [0u8; 32]; let (n, addr) = receiver.peek_from(&mut recv_buf).await?; assert_eq!(&recv_buf[..n], MSG); assert_eq!(addr, sender.local_addr()?); let mut recv_buf = [0u8; 32]; let (n, addr) = receiver.recv_from(&mut recv_buf).await?; assert_eq!(&recv_buf[..n], MSG); assert_eq!(addr, sender.local_addr()?); Ok(()) } #[tokio::test] async fn send_to_try_peek_from() -> std::io::Result<()> { let sender = UdpSocket::bind("127.0.0.1:0").await?; let receiver = UdpSocket::bind("127.0.0.1:0").await?; let receiver_addr = receiver.local_addr()?; poll_fn(|cx| sender.poll_send_to(cx, MSG, receiver_addr)).await?; // peek let mut recv_buf = [0u8; 32]; loop { match receiver.try_peek_from(&mut recv_buf) { Ok((n, addr)) => { assert_eq!(&recv_buf[..n], MSG); assert_eq!(addr, sender.local_addr()?); break; } Err(e) if e.kind() == io::ErrorKind::WouldBlock => { receiver.readable().await?; } Err(e) => return Err(e), } } // peek let mut recv_buf = [0u8; 32]; let (n, addr) = receiver.peek_from(&mut recv_buf).await?; assert_eq!(&recv_buf[..n], MSG); assert_eq!(addr, sender.local_addr()?); let mut recv_buf = [0u8; 32]; let (n, addr) = receiver.recv_from(&mut recv_buf).await?; assert_eq!(&recv_buf[..n], MSG); assert_eq!(addr, sender.local_addr()?); Ok(()) } #[tokio::test] async fn send_to_peek_from_poll() -> std::io::Result<()> { let sender = UdpSocket::bind("127.0.0.1:0").await?; let receiver = UdpSocket::bind("127.0.0.1:0").await?; let receiver_addr = receiver.local_addr()?; poll_fn(|cx| sender.poll_send_to(cx, MSG, receiver_addr)).await?; let mut recv_buf = [0u8; 32]; let mut read = ReadBuf::new(&mut recv_buf); let addr = poll_fn(|cx| receiver.poll_peek_from(cx, &mut read)).await?; assert_eq!(read.filled(), MSG); assert_eq!(addr, sender.local_addr()?); let mut recv_buf = [0u8; 32]; let mut read = ReadBuf::new(&mut recv_buf); poll_fn(|cx| receiver.poll_peek_from(cx, &mut read)).await?; assert_eq!(read.filled(), MSG); let mut recv_buf = [0u8; 32]; let mut read = ReadBuf::new(&mut recv_buf); poll_fn(|cx| receiver.poll_recv_from(cx, &mut read)).await?; assert_eq!(read.filled(), MSG); Ok(()) } #[tokio::test] async fn peek_sender() -> std::io::Result<()> { let sender = UdpSocket::bind("127.0.0.1:0").await?; let receiver = UdpSocket::bind("127.0.0.1:0").await?; let sender_addr = sender.local_addr()?; let receiver_addr = receiver.local_addr()?; let msg = b"Hello, world!"; sender.send_to(msg, receiver_addr).await?; let peeked_sender = receiver.peek_sender().await?; assert_eq!(peeked_sender, sender_addr); // Assert that `peek_sender()` returns the right sender but // doesn't remove from the receive queue. let mut recv_buf = [0u8; 32]; let (read, received_sender) = receiver.recv_from(&mut recv_buf).await?; assert_eq!(&recv_buf[..read], msg); assert_eq!(received_sender, peeked_sender); Ok(()) } #[tokio::test] async fn poll_peek_sender() -> std::io::Result<()> { let sender = UdpSocket::bind("127.0.0.1:0").await?; let receiver = UdpSocket::bind("127.0.0.1:0").await?; let sender_addr = sender.local_addr()?; let receiver_addr = receiver.local_addr()?; let msg = b"Hello, world!"; poll_fn(|cx| sender.poll_send_to(cx, msg, receiver_addr)).await?; let peeked_sender = poll_fn(|cx| receiver.poll_peek_sender(cx)).await?; assert_eq!(peeked_sender, sender_addr); // Assert that `poll_peek_sender()` returns the right sender but // doesn't remove from the receive queue. let mut recv_buf = [0u8; 32]; let mut read = ReadBuf::new(&mut recv_buf); let received_sender = poll_fn(|cx| receiver.poll_recv_from(cx, &mut read)).await?; assert_eq!(read.filled(), msg); assert_eq!(received_sender, peeked_sender); Ok(()) } #[tokio::test] async fn try_peek_sender() -> std::io::Result<()> { let sender = UdpSocket::bind("127.0.0.1:0").await?; let receiver = UdpSocket::bind("127.0.0.1:0").await?; let sender_addr = sender.local_addr()?; let receiver_addr = receiver.local_addr()?; let msg = b"Hello, world!"; sender.send_to(msg, receiver_addr).await?; let peeked_sender = loop { match receiver.try_peek_sender() { Ok(peeked_sender) => break peeked_sender, Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => { receiver.readable().await?; } Err(e) => return Err(e), } }; assert_eq!(peeked_sender, sender_addr); // Assert that `try_peek_sender()` returns the right sender but // didn't remove from the receive queue. let mut recv_buf = [0u8; 32]; // We already peeked the sender so there must be data in the receive queue. let (read, received_sender) = receiver.try_recv_from(&mut recv_buf).unwrap(); assert_eq!(&recv_buf[..read], msg); assert_eq!(received_sender, peeked_sender); Ok(()) } #[tokio::test] async fn split() -> std::io::Result<()> { let socket = UdpSocket::bind("127.0.0.1:0").await?; let s = Arc::new(socket); let r = s.clone(); let addr = s.local_addr()?; tokio::spawn(async move { s.send_to(MSG, &addr).await.unwrap(); }); let mut recv_buf = [0u8; 32]; let (len, _) = r.recv_from(&mut recv_buf[..]).await?; assert_eq!(&recv_buf[..len], MSG); Ok(()) } #[tokio::test] async fn split_chan() -> std::io::Result<()> { // setup UdpSocket that will echo all sent items let socket = UdpSocket::bind("127.0.0.1:0").await?; let addr = socket.local_addr().unwrap(); let s = Arc::new(socket); let r = s.clone(); let (tx, mut rx) = tokio::sync::mpsc::channel::<(Vec<u8>, std::net::SocketAddr)>(1_000); tokio::spawn(async move { while let Some((bytes, addr)) = rx.recv().await { s.send_to(&bytes, &addr).await.unwrap(); } }); tokio::spawn(async move { let mut buf = [0u8; 32]; loop { let (len, addr) = r.recv_from(&mut buf).await.unwrap(); tx.send((buf[..len].to_vec(), addr)).await.unwrap(); } }); // test that we can send a value and get back some response let sender = UdpSocket::bind("127.0.0.1:0").await?; sender.send_to(MSG, addr).await?; let mut recv_buf = [0u8; 32]; let (len, _) = sender.recv_from(&mut recv_buf).await?; assert_eq!(&recv_buf[..len], MSG); Ok(()) } #[tokio::test] async fn split_chan_poll() -> std::io::Result<()> { // setup UdpSocket that will echo all sent items let socket = UdpSocket::bind("127.0.0.1:0").await?; let addr = socket.local_addr().unwrap(); let s = Arc::new(socket); let r = s.clone(); let (tx, mut rx) = tokio::sync::mpsc::channel::<(Vec<u8>, std::net::SocketAddr)>(1_000); tokio::spawn(async move { while let Some((bytes, addr)) = rx.recv().await { poll_fn(|cx| s.poll_send_to(cx, &bytes, addr)) .await .unwrap(); } }); tokio::spawn(async move { let mut recv_buf = [0u8; 32]; let mut read = ReadBuf::new(&mut recv_buf); loop { let addr = poll_fn(|cx| r.poll_recv_from(cx, &mut read)).await.unwrap(); tx.send((read.filled().to_vec(), addr)).await.unwrap(); } }); // test that we can send a value and get back some response let sender = UdpSocket::bind("127.0.0.1:0").await?; poll_fn(|cx| sender.poll_send_to(cx, MSG, addr)).await?; let mut recv_buf = [0u8; 32]; let mut read = ReadBuf::new(&mut recv_buf); let _ = poll_fn(|cx| sender.poll_recv_from(cx, &mut read)).await?; assert_eq!(read.filled(), MSG); Ok(()) } // # Note // // This test is purposely written such that each time `sender` sends data on // the socket, `receiver` awaits the data. On Unix, it would be okay waiting // until the end of the test to receive all the data. On Windows, this would // **not** be okay because it's resources are completion based (via IOCP). // If data is sent and not yet received, attempting to send more data will // result in `ErrorKind::WouldBlock` until the first operation completes. #[tokio::test] async fn try_send_spawn() { const MSG2: &[u8] = b"world!"; const MSG2_LEN: usize = MSG2.len(); let sender = UdpSocket::bind("127.0.0.1:0").await.unwrap(); let receiver = UdpSocket::bind("127.0.0.1:0").await.unwrap(); receiver .connect(sender.local_addr().unwrap()) .await .unwrap(); sender.writable().await.unwrap(); let sent = &sender .try_send_to(MSG, receiver.local_addr().unwrap()) .unwrap(); assert_eq!(sent, &MSG_LEN); let mut buf = [0u8; 32]; let mut received = receiver.recv(&mut buf[..]).await.unwrap(); sender .connect(receiver.local_addr().unwrap()) .await .unwrap(); let sent = &sender.try_send(MSG2).unwrap(); assert_eq!(sent, &MSG2_LEN); received += receiver.recv(&mut buf[..]).await.unwrap(); std::thread::spawn(move || { let sent = &sender.try_send(MSG).unwrap(); assert_eq!(sent, &MSG_LEN); }) .join() .unwrap(); received += receiver.recv(&mut buf[..]).await.unwrap(); assert_eq!(received, MSG_LEN * 2 + MSG2_LEN); } #[tokio::test] async fn try_send_recv() { // Create listener let server = UdpSocket::bind("127.0.0.1:0").await.unwrap(); // Create socket pair let client = UdpSocket::bind("127.0.0.1:0").await.unwrap(); // Connect the two client.connect(server.local_addr().unwrap()).await.unwrap(); server.connect(client.local_addr().unwrap()).await.unwrap(); for _ in 0..5 { loop { client.writable().await.unwrap(); match client.try_send(b"hello world") { Ok(n) => { assert_eq!(n, 11); break; } Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("{e:?}"), } } loop { server.readable().await.unwrap(); let mut buf = [0; 512]; match server.try_recv(&mut buf) { Ok(n) => { assert_eq!(n, 11); assert_eq!(&buf[0..11], &b"hello world"[..]); break; } Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("{e:?}"), } } } } #[tokio::test] async fn try_send_to_recv_from() { // Create listener let server = UdpSocket::bind("127.0.0.1:0").await.unwrap(); let saddr = server.local_addr().unwrap(); // Create socket pair let client = UdpSocket::bind("127.0.0.1:0").await.unwrap(); let caddr = client.local_addr().unwrap(); for _ in 0..5 { loop { client.writable().await.unwrap(); match client.try_send_to(b"hello world", saddr) { Ok(n) => { assert_eq!(n, 11); break; } Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("{e:?}"), } } loop { server.readable().await.unwrap(); let mut buf = [0; 512]; match server.try_recv_from(&mut buf) { Ok((n, addr)) => { assert_eq!(n, 11); assert_eq!(addr, caddr); assert_eq!(&buf[0..11], &b"hello world"[..]); break; } Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("{e:?}"), } } } } #[tokio::test] async fn try_recv_buf() { // Create listener let server = UdpSocket::bind("127.0.0.1:0").await.unwrap(); // Create socket pair let client = UdpSocket::bind("127.0.0.1:0").await.unwrap(); // Connect the two client.connect(server.local_addr().unwrap()).await.unwrap(); server.connect(client.local_addr().unwrap()).await.unwrap(); for _ in 0..5 { loop { client.writable().await.unwrap(); match client.try_send(b"hello world") { Ok(n) => { assert_eq!(n, 11); break; } Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("{e:?}"), } } loop { server.readable().await.unwrap(); let mut buf = Vec::with_capacity(512); match server.try_recv_buf(&mut buf) { Ok(n) => { assert_eq!(n, 11); assert_eq!(&buf[0..11], &b"hello world"[..]); break; } Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("{e:?}"), } } } } #[tokio::test] async fn recv_buf() -> std::io::Result<()> { let sender = UdpSocket::bind("127.0.0.1:0").await?; let receiver = UdpSocket::bind("127.0.0.1:0").await?; sender.connect(receiver.local_addr()?).await?; receiver.connect(sender.local_addr()?).await?; sender.send(MSG).await?; let mut recv_buf = Vec::with_capacity(32); let len = receiver.recv_buf(&mut recv_buf).await?; assert_eq!(len, MSG_LEN); assert_eq!(&recv_buf[..len], MSG); Ok(()) } #[tokio::test] async fn try_recv_buf_from() { // Create listener let server = UdpSocket::bind("127.0.0.1:0").await.unwrap(); let saddr = server.local_addr().unwrap(); // Create socket pair let client = UdpSocket::bind("127.0.0.1:0").await.unwrap(); let caddr = client.local_addr().unwrap(); for _ in 0..5 { loop { client.writable().await.unwrap(); match client.try_send_to(b"hello world", saddr) { Ok(n) => { assert_eq!(n, 11); break; } Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("{e:?}"), } } loop { server.readable().await.unwrap(); let mut buf = Vec::with_capacity(512); match server.try_recv_buf_from(&mut buf) { Ok((n, addr)) => { assert_eq!(n, 11); assert_eq!(addr, caddr); assert_eq!(&buf[0..11], &b"hello world"[..]); break; } Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("{e:?}"), } } } } #[tokio::test] async fn recv_buf_from() -> std::io::Result<()> { let sender = UdpSocket::bind("127.0.0.1:0").await?; let receiver = UdpSocket::bind("127.0.0.1:0").await?; sender.connect(receiver.local_addr()?).await?; sender.send(MSG).await?; let mut recv_buf = Vec::with_capacity(32); let (len, caddr) = receiver.recv_buf_from(&mut recv_buf).await?; assert_eq!(len, MSG_LEN); assert_eq!(&recv_buf[..len], MSG); assert_eq!(caddr, sender.local_addr()?); Ok(()) } #[tokio::test] async fn poll_ready() { // Create listener let server = UdpSocket::bind("127.0.0.1:0").await.unwrap(); let saddr = server.local_addr().unwrap(); // Create socket pair let client = UdpSocket::bind("127.0.0.1:0").await.unwrap(); let caddr = client.local_addr().unwrap(); for _ in 0..5 { loop { assert_ok!(poll_fn(|cx| client.poll_send_ready(cx)).await); match client.try_send_to(b"hello world", saddr) { Ok(n) => { assert_eq!(n, 11); break; } Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("{e:?}"), } } loop { assert_ok!(poll_fn(|cx| server.poll_recv_ready(cx)).await); let mut buf = Vec::with_capacity(512); match server.try_recv_buf_from(&mut buf) { Ok((n, addr)) => { assert_eq!(n, 11); assert_eq!(addr, caddr); assert_eq!(&buf[0..11], &b"hello world"[..]); break; } Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("{e:?}"), } } } } /// Macro to create a simple test to set and get a socket option. macro_rules! test { // Test using the `arg`ument as expected return value. ($( #[ $attr: meta ] )* $get_fn: ident, $set_fn: ident ( $arg: expr ) ) => { test!($( #[$attr] )* $get_fn, $set_fn($arg), $arg); }; ($( #[ $attr: meta ] )* $get_fn: ident, $set_fn: ident ( $arg: expr ), $expected: expr ) => { #[tokio::test] $( #[$attr] )* async fn $get_fn() { test!(__ "127.0.0.1:0", $get_fn, $set_fn($arg), $expected); #[cfg(not(target_os = "vita"))] test!(__ "[::1]:0", $get_fn, $set_fn($arg), $expected); } }; // Only test using a IPv4 socket. (IPv4 $get_fn: ident, $set_fn: ident ( $arg: expr ) ) => { #[tokio::test] async fn $get_fn() { test!(__ "127.0.0.1:0", $get_fn, $set_fn($arg), $arg); } }; // Only test using a IPv6 socket. (IPv6 $get_fn: ident, $set_fn: ident ( $arg: expr ) ) => { #[tokio::test] async fn $get_fn() { test!(__ "[::1]:0", $get_fn, $set_fn($arg), $arg); } }; // Internal to this macro. (__ $addr: literal, $get_fn: ident, $set_fn: ident ( $arg: expr ), $expected: expr ) => { let socket = UdpSocket::bind($addr).await.expect("failed to create `UdpSocket`"); let initial = socket.$get_fn().expect("failed to get initial value"); let arg = $arg; assert_ne!(initial, arg, "initial value and argument are the same"); socket.$set_fn(arg).expect("failed to set option"); let got = socket.$get_fn().expect("failed to get value"); let expected = $expected; assert_eq!(got, expected, "set and get values differ"); }; } test!(broadcast, set_broadcast(true)); test!(IPv4 multicast_loop_v4, set_multicast_loop_v4(false)); #[cfg(target_os = "linux")] // broken on non-Linux platforms https://github.com/rust-lang/socket2/pull/630 test!(multicast_ttl_v4, set_multicast_ttl_v4(40)); test!(IPv6 multicast_loop_v6, set_multicast_loop_v6(false)); #[cfg(any( target_os = "android", target_os = "dragonfly", target_os = "freebsd", target_os = "fuchsia", target_os = "linux", target_os = "macos", target_os = "netbsd", target_os = "openbsd", target_os = "cygwin", ))] test!(IPv6 tclass_v6, set_tclass_v6(96)); test!(IPv4 ttl, set_ttl(40)); #[cfg(not(any( target_os = "fuchsia", target_os = "redox", target_os = "solaris", target_os = "illumos", target_os = "haiku" )))] test!(IPv4 tos_v4, set_tos_v4(96));
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/signal_twice.rs
tokio/tests/signal_twice.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] #![cfg(unix)] #![cfg(not(miri))] // No `sigaction` on Miri. mod support { pub mod signal; } use support::signal::send_signal; use tokio::signal::unix::{signal, SignalKind}; #[tokio::test] async fn twice() { let kind = SignalKind::user_defined1(); let mut sig = signal(kind).expect("failed to get signal"); for _ in 0..2 { send_signal(libc::SIGUSR1); assert!(sig.recv().await.is_some()); } }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/coop_budget.rs
tokio/tests/coop_budget.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", target_os = "linux"))] use std::sync::atomic::{AtomicUsize, Ordering}; use std::sync::Arc; use tokio::net::UdpSocket; use tokio::task::coop::{consume_budget, has_budget_remaining}; const BUDGET: usize = 128; /// Ensure that UDP sockets have functional budgeting /// /// # Design /// Two sockets communicate by spamming packets from one to the other. /// /// In Linux, this packet will be slammed through the entire network stack and into the receiver's buffer during the /// send system call because we are using the loopback interface. /// This happens because the softirq chain invoked on send when using the loopback interface covers virtually the /// entirety of the lifecycle of a packet within the kernel network stack. /// /// As a result, neither socket will ever encounter an EWOULDBLOCK, and the only way for these to yield during the loop /// is through budgeting. /// /// A second task runs in the background and increments a counter before yielding, allowing us to know how many times sockets yielded. /// Since we are both sending and receiving, that should happen once per 64 packets, because budgets are of size 128 /// and there are two budget events per packet, a send and a recv. #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` on miri. async fn coop_budget_udp_send_recv() { const N_ITERATIONS: usize = 1024; const PACKET: &[u8] = b"Hello, world"; const PACKET_LEN: usize = 12; assert_eq!( PACKET_LEN, PACKET.len(), "Defect in test, programmer can't do math" ); // bind each socket to a dynamic port, forcing IPv4 addressing on the localhost interface let tx = UdpSocket::bind("127.0.0.1:0").await.unwrap(); let rx = UdpSocket::bind("127.0.0.1:0").await.unwrap(); tx.connect(rx.local_addr().unwrap()).await.unwrap(); rx.connect(tx.local_addr().unwrap()).await.unwrap(); let tracker = Arc::new(AtomicUsize::default()); let tracker_clone = Arc::clone(&tracker); tokio::task::yield_now().await; tokio::spawn(async move { loop { tracker_clone.fetch_add(1, Ordering::SeqCst); tokio::task::yield_now().await; } }); for _ in 0..N_ITERATIONS { tx.send(PACKET).await.unwrap(); let mut tmp = [0; PACKET_LEN]; // ensure that we aren't somehow accumulating other assert_eq!( PACKET_LEN, rx.recv(&mut tmp).await.unwrap(), "Defect in test case, received unexpected result from socket" ); assert_eq!( PACKET, &tmp, "Defect in test case, received unexpected result from socket" ); } assert_eq!(N_ITERATIONS / (BUDGET / 2), tracker.load(Ordering::SeqCst)); } #[tokio::test] async fn test_has_budget_remaining() { // At the beginning, budget should be available. assert!(has_budget_remaining()); // Deplete the budget for _ in 0..BUDGET { consume_budget().await; } assert!(!has_budget_remaining()); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_poll_aio.rs
tokio/tests/io_poll_aio.rs
#![warn(rust_2018_idioms)] #![cfg(all(target_os = "freebsd", feature = "net"))] use mio_aio::{AioFsyncMode, SourceApi}; use std::{ future::Future, io, mem, os::fd::AsFd, os::unix::io::{AsRawFd, RawFd}, pin::{pin, Pin}, task::{Context, Poll}, }; use tempfile::tempfile; use tokio::io::bsd::{Aio, AioSource}; use tokio_test::assert_pending; mod aio { use super::*; #[derive(Debug)] struct TokioSource<'fd>(mio_aio::Source<nix::sys::aio::AioFsync<'fd>>); impl<'fd> AioSource for TokioSource<'fd> { fn register(&mut self, kq: RawFd, token: usize) { self.0.register_raw(kq, token) } fn deregister(&mut self) { self.0.deregister_raw() } } /// A very crude implementation of an AIO-based future struct FsyncFut<'fd>(Aio<TokioSource<'fd>>); impl<'fd> FsyncFut<'fd> { pub fn submit(self: Pin<&mut Self>) -> io::Result<()> { let p = unsafe { self.map_unchecked_mut(|s| &mut s.0 .0) }; match p.submit() { Ok(()) => Ok(()), Err(e) => Err(io::Error::from_raw_os_error(e as i32)), } } } impl<'fd> Future for FsyncFut<'fd> { type Output = io::Result<()>; fn poll(self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Self::Output> { let poll_result = self.0.poll_ready(cx); match poll_result { Poll::Pending => Poll::Pending, Poll::Ready(Err(e)) => Poll::Ready(Err(e)), Poll::Ready(Ok(_ev)) => { // At this point, we could clear readiness. But there's no // point, since we're about to drop the Aio. let p = unsafe { self.map_unchecked_mut(|s| &mut s.0 .0) }; let result = p.aio_return(); match result { Ok(r) => Poll::Ready(Ok(r)), Err(e) => Poll::Ready(Err(io::Error::from_raw_os_error(e as i32))), } } } } } /// Low-level AIO Source /// /// An example bypassing mio_aio and Nix to demonstrate how the kevent /// registration actually works, under the hood. struct LlSource(Pin<Box<libc::aiocb>>); impl LlSource { fn fsync(mut self: Pin<&mut Self>) { let r = unsafe { let p = self.0.as_mut().get_unchecked_mut(); libc::aio_fsync(libc::O_SYNC, p) }; assert_eq!(0, r); } } impl AioSource for LlSource { fn register(&mut self, kq: RawFd, token: usize) { let mut sev: libc::sigevent = unsafe { mem::MaybeUninit::zeroed().assume_init() }; sev.sigev_notify = libc::SIGEV_KEVENT; sev.sigev_signo = kq; sev.sigev_value = libc::sigval { sival_ptr: token as *mut libc::c_void, }; self.0.aio_sigevent = sev; } fn deregister(&mut self) { unsafe { self.0.aio_sigevent = mem::zeroed(); } } } struct LlFut(Aio<LlSource>); impl LlFut { pub fn fsync(self: Pin<&mut Self>) { let p = unsafe { self.map_unchecked_mut(|s| &mut *(s.0)) }; p.fsync(); } } impl Future for LlFut { type Output = std::io::Result<usize>; fn poll(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Self::Output> { let poll_result = self.0.poll_ready(cx); match poll_result { Poll::Pending => Poll::Pending, Poll::Ready(Err(e)) => Poll::Ready(Err(e)), Poll::Ready(Ok(ev)) => { // Clearing readiness makes the future non-idempotent; the // caller can't poll it repeatedly after it has already // returned Ready. But that's ok; most futures behave this // way. self.0.clear_ready(ev); let r = unsafe { libc::aio_return(self.0 .0.as_mut().get_unchecked_mut()) }; if r >= 0 { Poll::Ready(Ok(r as usize)) } else { Poll::Ready(Err(io::Error::last_os_error())) } } } } } #[tokio::test] async fn fsync() { let f = tempfile().unwrap(); let fd = f.as_fd(); let mode = AioFsyncMode::O_SYNC; let source = TokioSource(mio_aio::Fsync::fsync(fd, mode, 0)); let poll_aio = Aio::new_for_aio(source).unwrap(); let mut fut = pin!(FsyncFut(poll_aio)); fut.as_mut().submit().unwrap(); fut.await.unwrap(); } #[tokio::test] async fn ll_fsync() { let f = tempfile().unwrap(); let fd = f.as_raw_fd(); let mut aiocb: libc::aiocb = unsafe { mem::MaybeUninit::zeroed().assume_init() }; aiocb.aio_fildes = fd; let source = LlSource(Box::pin(aiocb)); let mut poll_aio = Aio::new_for_aio(source).unwrap(); let r = unsafe { let p = poll_aio.0.as_mut().get_unchecked_mut(); libc::aio_fsync(libc::O_SYNC, p) }; assert_eq!(0, r); let fut = LlFut(poll_aio); fut.await.unwrap(); } /// A suitably crafted future type can reuse an Aio object #[tokio::test] async fn reuse() { let f = tempfile().unwrap(); let fd = f.as_raw_fd(); let mut aiocb: libc::aiocb = unsafe { mem::MaybeUninit::zeroed().assume_init() }; aiocb.aio_fildes = fd; let source = LlSource(Box::pin(aiocb)); let poll_aio = Aio::new_for_aio(source).unwrap(); // Send the operation to the kernel the first time let mut fut = LlFut(poll_aio); { let mut pfut = Pin::new(&mut fut); pfut.as_mut().fsync(); pfut.as_mut().await.unwrap(); } // Check that readiness was cleared let mut ctx = Context::from_waker(futures::task::noop_waker_ref()); assert_pending!(fut.0.poll_ready(&mut ctx)); // and reuse the future and its Aio object { let mut pfut = Pin::new(&mut fut); pfut.as_mut().fsync(); pfut.as_mut().await.unwrap(); } } } mod lio { use super::*; /// Low-level source based on lio_listio /// /// An example demonstrating using AIO with `Interest::Lio`. mio_aio 0.8 /// doesn't include any bindings for lio_listio, so we've got to go /// low-level. struct LioSource<'a> { aiocb: Pin<&'a mut [&'a mut libc::aiocb; 1]>, sev: libc::sigevent, } impl<'a> LioSource<'a> { fn new(aiocb: Pin<&'a mut [&'a mut libc::aiocb; 1]>) -> Self { LioSource { aiocb, sev: unsafe { mem::zeroed() }, } } fn submit(mut self: Pin<&mut Self>) { let p: *const *mut libc::aiocb = unsafe { self.aiocb.as_mut().get_unchecked_mut() } as *const _ as *const *mut _; let r = unsafe { libc::lio_listio(libc::LIO_NOWAIT, p, 1, &mut self.sev) }; assert_eq!(r, 0); } } impl<'a> AioSource for LioSource<'a> { fn register(&mut self, kq: RawFd, token: usize) { let mut sev: libc::sigevent = unsafe { mem::MaybeUninit::zeroed().assume_init() }; sev.sigev_notify = libc::SIGEV_KEVENT; sev.sigev_signo = kq; sev.sigev_value = libc::sigval { sival_ptr: token as *mut libc::c_void, }; self.sev = sev; } fn deregister(&mut self) { unsafe { self.sev = mem::zeroed(); } } } struct LioFut<'a>(Aio<LioSource<'a>>); impl<'a> LioFut<'a> { pub fn submit(self: Pin<&mut Self>) { let p = unsafe { self.map_unchecked_mut(|s| &mut *(s.0)) }; p.submit(); } } impl<'a> Future for LioFut<'a> { type Output = std::io::Result<usize>; fn poll(self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Self::Output> { let poll_result = self.0.poll_ready(cx); match poll_result { Poll::Pending => Poll::Pending, Poll::Ready(Err(e)) => Poll::Ready(Err(e)), Poll::Ready(Ok(ev)) => { // Clearing readiness makes the future non-idempotent; the // caller can't poll it repeatedly after it has already // returned Ready. But that's ok; most futures behave this // way. Clearing readiness is especially useful for // lio_listio, because sometimes some operations will be // ready but not all. self.0.clear_ready(ev); let r = unsafe { let p1 = self.get_unchecked_mut(); let p2: &mut [&mut libc::aiocb; 1] = p1.0.aiocb.as_mut().get_unchecked_mut(); let p3: &mut libc::aiocb = p2[0]; libc::aio_return(p3) }; if r >= 0 { Poll::Ready(Ok(r as usize)) } else { Poll::Ready(Err(io::Error::last_os_error())) } } } } } /// An lio_listio operation with one fsync element #[tokio::test] async fn onewrite() { const WBUF: &[u8] = b"abcdef"; let f = tempfile().unwrap(); let mut aiocb: libc::aiocb = unsafe { mem::zeroed() }; aiocb.aio_fildes = f.as_raw_fd(); aiocb.aio_lio_opcode = libc::LIO_WRITE; aiocb.aio_nbytes = WBUF.len(); aiocb.aio_buf = WBUF.as_ptr() as *mut _; let aiocb = pin!([&mut aiocb]); let source = LioSource::new(aiocb); let poll_aio = Aio::new_for_lio(source).unwrap(); // Send the operation to the kernel let mut fut = pin!(LioFut(poll_aio)); fut.as_mut().submit(); fut.await.unwrap(); } /// A suitably crafted future type can reuse an Aio object #[tokio::test] async fn reuse() { const WBUF: &[u8] = b"abcdef"; let f = tempfile().unwrap(); let mut aiocb: libc::aiocb = unsafe { mem::zeroed() }; aiocb.aio_fildes = f.as_raw_fd(); aiocb.aio_lio_opcode = libc::LIO_WRITE; aiocb.aio_nbytes = WBUF.len(); aiocb.aio_buf = WBUF.as_ptr() as *mut _; let aiocb = pin!([&mut aiocb]); let source = LioSource::new(aiocb); let poll_aio = Aio::new_for_lio(source).unwrap(); // Send the operation to the kernel the first time let mut fut = LioFut(poll_aio); { let mut pfut = Pin::new(&mut fut); pfut.as_mut().submit(); pfut.as_mut().await.unwrap(); } // Check that readiness was cleared let mut ctx = Context::from_waker(futures::task::noop_waker_ref()); assert_pending!(fut.0.poll_ready(&mut ctx)); // and reuse the future and its Aio object { let mut pfut = Pin::new(&mut fut); pfut.as_mut().submit(); pfut.as_mut().await.unwrap(); } } }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/net_quickack.rs
tokio/tests/net_quickack.rs
#![warn(rust_2018_idioms)] #![cfg(all( feature = "net", any( target_os = "linux", target_os = "android", target_os = "fuchsia", target_os = "cygwin", ) ))] #![cfg(not(miri))] // No `socket` in miri. use tokio::io::{AsyncReadExt, AsyncWriteExt}; use tokio::net::TcpListener; use tokio::net::TcpStream; use tokio::sync::oneshot; #[tokio::test] async fn socket_works_with_quickack() { const MESSAGE: &str = "Hello, tokio!"; let (tx_port, rx_port) = oneshot::channel(); let server = tokio::spawn(async move { let listener = TcpListener::bind("127.0.0.1:0").await.unwrap(); let addr = listener.local_addr().unwrap(); tx_port.send(addr.port()).unwrap(); let (mut stream, _) = listener.accept().await.unwrap(); stream.set_quickack(true).unwrap(); assert!(stream.quickack().unwrap()); stream.write_all(MESSAGE.as_bytes()).await.unwrap(); let mut buf = vec![0; MESSAGE.len()]; stream.read_exact(&mut buf).await.unwrap(); assert_eq!(buf, MESSAGE.as_bytes()); // There is nothing special about setting quickack to false // at this point, we just want to test the `false` case. stream.set_quickack(false).unwrap(); assert!(!stream.quickack().unwrap()); stream.shutdown().await.unwrap(); }); let port = rx_port.await.unwrap(); let client = tokio::spawn(async move { let mut stream = TcpStream::connect(format!("127.0.0.1:{port}")) .await .unwrap(); stream.set_quickack(true).unwrap(); assert!(stream.quickack().unwrap()); let mut buf = vec![0; MESSAGE.len()]; stream.read_exact(&mut buf).await.unwrap(); assert_eq!(buf, MESSAGE.as_bytes()); stream.write_all(MESSAGE.as_bytes()).await.unwrap(); // There is nothing special about setting quickack to false // at this point, we just want to test the `false` case. stream.set_quickack(false).unwrap(); assert!(!stream.quickack().unwrap()); stream.shutdown().await.unwrap(); }); tokio::try_join!(server, client).unwrap(); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/uds_datagram.rs
tokio/tests/uds_datagram.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] #![cfg(unix)] use tokio::io::ReadBuf; use tokio::net::UnixDatagram; use tokio::try_join; use std::future::poll_fn; use std::io; use std::sync::Arc; async fn echo_server(socket: UnixDatagram) -> io::Result<()> { let mut recv_buf = vec![0u8; 1024]; loop { let (len, peer_addr) = socket.recv_from(&mut recv_buf[..]).await?; if let Some(path) = peer_addr.as_pathname() { socket.send_to(&recv_buf[..len], path).await?; } } } #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` on miri. async fn echo() -> io::Result<()> { let dir = tempfile::tempdir().unwrap(); let server_path = dir.path().join("server.sock"); let client_path = dir.path().join("client.sock"); let server_socket = UnixDatagram::bind(server_path.clone())?; tokio::spawn(async move { let _ = echo_server(server_socket).await; }); { let socket = UnixDatagram::bind(&client_path).unwrap(); socket.connect(server_path)?; socket.send(b"ECHO").await?; let mut recv_buf = [0u8; 16]; let len = socket.recv(&mut recv_buf[..]).await?; assert_eq!(&recv_buf[..len], b"ECHO"); } Ok(()) } #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` on miri. async fn echo_from() -> io::Result<()> { let dir = tempfile::tempdir().unwrap(); let server_path = dir.path().join("server.sock"); let client_path = dir.path().join("client.sock"); let server_socket = UnixDatagram::bind(server_path.clone())?; tokio::spawn(async move { let _ = echo_server(server_socket).await; }); { let socket = UnixDatagram::bind(&client_path).unwrap(); socket.connect(&server_path)?; socket.send(b"ECHO").await?; let mut recv_buf = [0u8; 16]; let (len, addr) = socket.recv_from(&mut recv_buf[..]).await?; assert_eq!(&recv_buf[..len], b"ECHO"); assert_eq!(addr.as_pathname(), Some(server_path.as_path())); } Ok(()) } // Even though we use sync non-blocking io we still need a reactor. #[tokio::test] #[cfg_attr(miri, ignore)] // No SOCK_DGRAM for `socketpair` in miri. async fn try_send_recv_never_block() -> io::Result<()> { let mut recv_buf = [0u8; 16]; let payload = b"PAYLOAD"; let mut count = 0; let (dgram1, dgram2) = UnixDatagram::pair()?; // Send until we hit the OS `net.unix.max_dgram_qlen`. loop { dgram1.writable().await.unwrap(); match dgram1.try_send(payload) { Err(err) => match (err.kind(), err.raw_os_error()) { (io::ErrorKind::WouldBlock, _) => break, (_, Some(libc::ENOBUFS)) => break, _ => { panic!("unexpected error {err:?}"); } }, Ok(len) => { assert_eq!(len, payload.len()); } } count += 1; } // Read every dgram we sent. while count > 0 { dgram2.readable().await.unwrap(); let len = dgram2.try_recv(&mut recv_buf[..])?; assert_eq!(len, payload.len()); assert_eq!(payload, &recv_buf[..len]); count -= 1; } let err = dgram2.try_recv(&mut recv_buf[..]).unwrap_err(); match err.kind() { io::ErrorKind::WouldBlock => (), _ => unreachable!("unexpected error {:?}", err), } Ok(()) } #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` on miri. async fn split() -> std::io::Result<()> { let dir = tempfile::tempdir().unwrap(); let path = dir.path().join("split.sock"); let s = Arc::new(UnixDatagram::bind(path.clone())?); let r = s.clone(); let msg = b"hello"; let ((), ()) = try_join! { async { s.send_to(msg, path).await?; io::Result::Ok(()) }, async { let mut recv_buf = [0u8; 32]; let (len, _) = r.recv_from(&mut recv_buf[..]).await?; assert_eq!(&recv_buf[..len], msg); Ok(()) }, }?; Ok(()) } #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` on miri. async fn send_to_recv_from_poll() -> std::io::Result<()> { let dir = tempfile::tempdir().unwrap(); let sender_path = dir.path().join("sender.sock"); let receiver_path = dir.path().join("receiver.sock"); let sender = UnixDatagram::bind(&sender_path)?; let receiver = UnixDatagram::bind(&receiver_path)?; let msg = b"hello"; poll_fn(|cx| sender.poll_send_to(cx, msg, &receiver_path)).await?; let mut recv_buf = [0u8; 32]; let mut read = ReadBuf::new(&mut recv_buf); let addr = poll_fn(|cx| receiver.poll_recv_from(cx, &mut read)).await?; assert_eq!(read.filled(), msg); assert_eq!(addr.as_pathname(), Some(sender_path.as_ref())); Ok(()) } #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` on miri. async fn send_recv_poll() -> std::io::Result<()> { let dir = tempfile::tempdir().unwrap(); let sender_path = dir.path().join("sender.sock"); let receiver_path = dir.path().join("receiver.sock"); let sender = UnixDatagram::bind(&sender_path)?; let receiver = UnixDatagram::bind(&receiver_path)?; sender.connect(&receiver_path)?; receiver.connect(&sender_path)?; let msg = b"hello"; poll_fn(|cx| sender.poll_send(cx, msg)).await?; let mut recv_buf = [0u8; 32]; let mut read = ReadBuf::new(&mut recv_buf); poll_fn(|cx| receiver.poll_recv(cx, &mut read)).await?; assert_eq!(read.filled(), msg); Ok(()) } #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` on miri. async fn try_send_to_recv_from() -> std::io::Result<()> { let dir = tempfile::tempdir().unwrap(); let server_path = dir.path().join("server.sock"); let client_path = dir.path().join("client.sock"); // Create listener let server = UnixDatagram::bind(&server_path)?; // Create socket pair let client = UnixDatagram::bind(&client_path)?; for _ in 0..5 { loop { client.writable().await?; match client.try_send_to(b"hello world", &server_path) { Ok(n) => { assert_eq!(n, 11); break; } Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("{e:?}"), } } loop { server.readable().await?; let mut buf = [0; 512]; match server.try_recv_from(&mut buf) { Ok((n, addr)) => { assert_eq!(n, 11); assert_eq!(addr.as_pathname(), Some(client_path.as_ref())); assert_eq!(&buf[0..11], &b"hello world"[..]); break; } Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("{e:?}"), } } } Ok(()) } #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` on miri. async fn try_recv_buf_from() -> std::io::Result<()> { let dir = tempfile::tempdir().unwrap(); let server_path = dir.path().join("server.sock"); let client_path = dir.path().join("client.sock"); // Create listener let server = UnixDatagram::bind(&server_path)?; // Create socket pair let client = UnixDatagram::bind(&client_path)?; for _ in 0..5 { loop { client.writable().await?; match client.try_send_to(b"hello world", &server_path) { Ok(n) => { assert_eq!(n, 11); break; } Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("{e:?}"), } } loop { server.readable().await?; let mut buf = Vec::with_capacity(512); match server.try_recv_buf_from(&mut buf) { Ok((n, addr)) => { assert_eq!(n, 11); assert_eq!(addr.as_pathname(), Some(client_path.as_ref())); assert_eq!(&buf[0..11], &b"hello world"[..]); break; } Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("{e:?}"), } } } Ok(()) } #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` on miri. async fn recv_buf_from() -> std::io::Result<()> { let tmp = tempfile::tempdir()?; // Bind each socket to a filesystem path let tx_path = tmp.path().join("tx"); let tx = UnixDatagram::bind(&tx_path)?; let rx_path = tmp.path().join("rx"); let rx = UnixDatagram::bind(&rx_path)?; let bytes = b"hello world"; tx.send_to(bytes, &rx_path).await?; let mut buf = Vec::with_capacity(24); let (size, addr) = rx.recv_buf_from(&mut buf).await?; let dgram = &buf[..size]; assert_eq!(dgram, bytes); assert_eq!(addr.as_pathname().unwrap(), &tx_path); Ok(()) } // Even though we use sync non-blocking io we still need a reactor. #[tokio::test] #[cfg_attr(miri, ignore)] // No SOCK_DGRAM for `socketpair` in miri. async fn try_recv_buf_never_block() -> io::Result<()> { let payload = b"PAYLOAD"; let mut count = 0; let (dgram1, dgram2) = UnixDatagram::pair()?; // Send until we hit the OS `net.unix.max_dgram_qlen`. loop { dgram1.writable().await.unwrap(); match dgram1.try_send(payload) { Err(err) => match (err.kind(), err.raw_os_error()) { (io::ErrorKind::WouldBlock, _) => break, (_, Some(libc::ENOBUFS)) => break, _ => { panic!("unexpected error {err:?}"); } }, Ok(len) => { assert_eq!(len, payload.len()); } } count += 1; } // Read every dgram we sent. while count > 0 { let mut recv_buf = Vec::with_capacity(16); dgram2.readable().await.unwrap(); let len = dgram2.try_recv_buf(&mut recv_buf)?; assert_eq!(len, payload.len()); assert_eq!(payload, &recv_buf[..len]); count -= 1; } let mut recv_buf = vec![0; 16]; let err = dgram2.try_recv_from(&mut recv_buf).unwrap_err(); match err.kind() { io::ErrorKind::WouldBlock => (), _ => unreachable!("unexpected error {:?}", err), } Ok(()) } #[tokio::test] #[cfg_attr(miri, ignore)] // No SOCK_DGRAM for `socketpair` in miri. async fn recv_buf() -> std::io::Result<()> { // Create the pair of sockets let (sock1, sock2) = UnixDatagram::pair()?; // Since the sockets are paired, the paired send/recv // functions can be used let bytes = b"hello world"; sock1.send(bytes).await?; let mut buff = Vec::with_capacity(24); let size = sock2.recv_buf(&mut buff).await?; let dgram = &buff[..size]; assert_eq!(dgram, bytes); Ok(()) } #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` on miri. async fn poll_ready() -> io::Result<()> { let dir = tempfile::tempdir().unwrap(); let server_path = dir.path().join("server.sock"); let client_path = dir.path().join("client.sock"); // Create listener let server = UnixDatagram::bind(&server_path)?; // Create socket pair let client = UnixDatagram::bind(&client_path)?; for _ in 0..5 { loop { poll_fn(|cx| client.poll_send_ready(cx)).await?; match client.try_send_to(b"hello world", &server_path) { Ok(n) => { assert_eq!(n, 11); break; } Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("{e:?}"), } } loop { poll_fn(|cx| server.poll_recv_ready(cx)).await?; let mut buf = Vec::with_capacity(512); match server.try_recv_buf_from(&mut buf) { Ok((n, addr)) => { assert_eq!(n, 11); assert_eq!(addr.as_pathname(), Some(client_path.as_ref())); assert_eq!(&buf[0..11], &b"hello world"[..]); break; } Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("{e:?}"), } } } Ok(()) }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/rt_poll_callbacks.rs
tokio/tests/rt_poll_callbacks.rs
#![allow(unknown_lints, unexpected_cfgs)] #![cfg(tokio_unstable)] use std::sync::{atomic::AtomicUsize, Arc, Mutex}; use tokio::task::yield_now; #[cfg(not(target_os = "wasi"))] #[test] fn callbacks_fire_multi_thread() { let poll_start_counter = Arc::new(AtomicUsize::new(0)); let poll_stop_counter = Arc::new(AtomicUsize::new(0)); let poll_start = poll_start_counter.clone(); let poll_stop = poll_stop_counter.clone(); let before_task_poll_callback_task_id: Arc<Mutex<Option<tokio::task::Id>>> = Arc::new(Mutex::new(None)); let after_task_poll_callback_task_id: Arc<Mutex<Option<tokio::task::Id>>> = Arc::new(Mutex::new(None)); let before_task_poll_callback_task_id_ref = Arc::clone(&before_task_poll_callback_task_id); let after_task_poll_callback_task_id_ref = Arc::clone(&after_task_poll_callback_task_id); let rt = tokio::runtime::Builder::new_multi_thread() .enable_all() .on_before_task_poll(move |task_meta| { before_task_poll_callback_task_id_ref .lock() .unwrap() .replace(task_meta.id()); poll_start_counter.fetch_add(1, std::sync::atomic::Ordering::Relaxed); }) .on_after_task_poll(move |task_meta| { after_task_poll_callback_task_id_ref .lock() .unwrap() .replace(task_meta.id()); poll_stop_counter.fetch_add(1, std::sync::atomic::Ordering::Relaxed); }) .build() .unwrap(); let task = rt.spawn(async { yield_now().await; yield_now().await; yield_now().await; }); let spawned_task_id = task.id(); rt.block_on(task).expect("task should succeed"); // We need to drop the runtime to guarantee the workers have exited (and thus called the callback) drop(rt); assert_eq!( before_task_poll_callback_task_id.lock().unwrap().unwrap(), spawned_task_id ); assert_eq!( after_task_poll_callback_task_id.lock().unwrap().unwrap(), spawned_task_id ); let actual_count = 4; assert_eq!( poll_start.load(std::sync::atomic::Ordering::Relaxed), actual_count, "unexpected number of poll starts" ); assert_eq!( poll_stop.load(std::sync::atomic::Ordering::Relaxed), actual_count, "unexpected number of poll stops" ); } #[test] fn callbacks_fire_current_thread() { let poll_start_counter = Arc::new(AtomicUsize::new(0)); let poll_stop_counter = Arc::new(AtomicUsize::new(0)); let poll_start = poll_start_counter.clone(); let poll_stop = poll_stop_counter.clone(); let before_task_poll_callback_task_id: Arc<Mutex<Option<tokio::task::Id>>> = Arc::new(Mutex::new(None)); let after_task_poll_callback_task_id: Arc<Mutex<Option<tokio::task::Id>>> = Arc::new(Mutex::new(None)); let before_task_poll_callback_task_id_ref = Arc::clone(&before_task_poll_callback_task_id); let after_task_poll_callback_task_id_ref = Arc::clone(&after_task_poll_callback_task_id); let rt = tokio::runtime::Builder::new_current_thread() .enable_all() .on_before_task_poll(move |task_meta| { before_task_poll_callback_task_id_ref .lock() .unwrap() .replace(task_meta.id()); poll_start_counter.fetch_add(1, std::sync::atomic::Ordering::Relaxed); }) .on_after_task_poll(move |task_meta| { after_task_poll_callback_task_id_ref .lock() .unwrap() .replace(task_meta.id()); poll_stop_counter.fetch_add(1, std::sync::atomic::Ordering::Relaxed); }) .build() .unwrap(); let task = rt.spawn(async { yield_now().await; yield_now().await; yield_now().await; }); let spawned_task_id = task.id(); let _ = rt.block_on(task); drop(rt); assert_eq!( before_task_poll_callback_task_id.lock().unwrap().unwrap(), spawned_task_id ); assert_eq!( after_task_poll_callback_task_id.lock().unwrap().unwrap(), spawned_task_id ); assert_eq!(poll_start.load(std::sync::atomic::Ordering::Relaxed), 4); assert_eq!(poll_stop.load(std::sync::atomic::Ordering::Relaxed), 4); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/fs_remove_dir_all.rs
tokio/tests/fs_remove_dir_all.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // WASI does not support all fs operations use tempfile::tempdir; use tokio::fs; #[tokio::test] async fn remove_dir_all() { let temp_dir = tempdir().unwrap(); let test_dir = temp_dir.path().join("test"); fs::create_dir(&test_dir).await.unwrap(); let file_path = test_dir.as_path().join("a.txt"); fs::write(&file_path, b"Hello File!").await.unwrap(); fs::remove_dir_all(test_dir.as_path()).await.unwrap(); // test dir should no longer exist match fs::try_exists(test_dir).await { Ok(exists) => assert!(!exists), Err(_) => println!("ignored try_exists error after remove_dir_all"), }; // contents should no longer exist match fs::try_exists(file_path).await { Ok(exists) => assert!(!exists), Err(_) => println!("ignored try_exists error after remove_dir_all"), }; }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/macros_rename_test.rs
tokio/tests/macros_rename_test.rs
#![cfg(all(feature = "full", not(target_os = "wasi")))] // Wasi doesn't support threading #[allow(unused_imports)] use std as tokio; use ::tokio as tokio1; mod test { pub use ::tokio; } async fn compute() -> usize { let join = tokio1::spawn(async { 1 }); join.await.unwrap() } #[tokio1::main(crate = "tokio1")] async fn compute_main() -> usize { compute().await } #[test] fn crate_rename_main() { assert_eq!(1, compute_main()); } #[tokio1::test(crate = "tokio1")] async fn crate_rename_test() { assert_eq!(1, compute().await); } #[test::tokio::test(crate = "test::tokio")] async fn crate_path_test() { assert_eq!(1, compute().await); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/test_clock.rs
tokio/tests/test_clock.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use tokio::time::{self, Duration, Instant}; #[tokio::test] async fn resume_lets_time_move_forward_instead_of_resetting_it() { let start = Instant::now(); time::pause(); time::advance(Duration::from_secs(10)).await; let advanced_by_ten_secs = Instant::now(); assert!(advanced_by_ten_secs - start > Duration::from_secs(10)); assert!(advanced_by_ten_secs - start < Duration::from_secs(11)); time::resume(); assert!(advanced_by_ten_secs < Instant::now()); assert!(Instant::now() - advanced_by_ten_secs < Duration::from_secs(1)); } #[tokio::test] async fn can_pause_after_resume() { let start = Instant::now(); time::pause(); time::advance(Duration::from_secs(10)).await; time::resume(); time::pause(); time::advance(Duration::from_secs(10)).await; assert!(Instant::now() - start > Duration::from_secs(20)); assert!(Instant::now() - start < Duration::from_secs(21)); } #[tokio::test] #[should_panic] async fn freezing_time_while_frozen_panics() { time::pause(); time::pause(); } #[tokio::test] #[should_panic] async fn advancing_time_when_time_is_not_frozen_panics() { time::advance(Duration::from_secs(1)).await; } #[tokio::test] #[should_panic] async fn resuming_time_when_not_frozen_panics() { time::pause(); time::resume(); time::resume(); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/task_builder.rs
tokio/tests/task_builder.rs
#![cfg(all(tokio_unstable, feature = "tracing"))] use std::rc::Rc; use tokio::task::{Builder, LocalSet}; #[tokio::test] async fn spawn_with_name() { let result = Builder::new() .name("name") .spawn(async { "task executed" }) .unwrap() .await; assert_eq!(result.unwrap(), "task executed"); } #[tokio::test(flavor = "local")] async fn spawn_local_on_local_runtime() { let result = Builder::new() .spawn_local(async { "task executed" }) .unwrap() .await; assert_eq!(result.unwrap(), "task executed"); } #[tokio::test] #[should_panic = "`spawn_local` called from outside of a `task::LocalSet` or `runtime::LocalRuntime`"] async fn spawn_local_panics_outside_local_set_or_local_runtime() { let _ = Builder::new() .spawn_local(async { "task executed" }) .unwrap() .await; } #[tokio::test(flavor = "multi_thread")] #[should_panic = "`spawn_local` called from outside of a `task::LocalSet` or `runtime::LocalRuntime`"] async fn spawn_local_panics_in_multi_thread_runtime() { let _ = Builder::new() .spawn_local(async { "task executed" }) .unwrap() .await; } #[tokio::test] async fn spawn_blocking_with_name() { let result = Builder::new() .name("name") .spawn_blocking(|| "task executed") .unwrap() .await; assert_eq!(result.unwrap(), "task executed"); } #[tokio::test] async fn spawn_local_with_name() { let unsend_data = Rc::new("task executed"); let result = LocalSet::new() .run_until(async move { Builder::new() .name("name") .spawn_local(async move { unsend_data }) .unwrap() .await }) .await; assert_eq!(*result.unwrap(), "task executed"); } #[tokio::test] async fn spawn_without_name() { let result = Builder::new() .spawn(async { "task executed" }) .unwrap() .await; assert_eq!(result.unwrap(), "task executed"); } #[tokio::test] async fn spawn_blocking_without_name() { let result = Builder::new() .spawn_blocking(|| "task executed") .unwrap() .await; assert_eq!(result.unwrap(), "task executed"); } #[tokio::test] async fn spawn_local_without_name() { let unsend_data = Rc::new("task executed"); let result = LocalSet::new() .run_until(async move { Builder::new() .spawn_local(async move { unsend_data }) .unwrap() .await }) .await; assert_eq!(*result.unwrap(), "task executed"); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/macros_test.rs
tokio/tests/macros_test.rs
#![cfg(all(feature = "full", not(target_os = "wasi")))] // Wasi doesn't support threading use tokio::test; #[test] async fn test_macro_can_be_used_via_use() { tokio::spawn(async {}).await.unwrap(); } #[tokio::test] async fn test_macro_is_resilient_to_shadowing() { tokio::spawn(async {}).await.unwrap(); } // https://github.com/tokio-rs/tokio/issues/3403 #[rustfmt::skip] // this `rustfmt::skip` is necessary because unused_braces does not warn if the block contains newline. #[tokio::main] pub async fn unused_braces_main() { println!("hello") } #[rustfmt::skip] // this `rustfmt::skip` is necessary because unused_braces does not warn if the block contains newline. #[tokio::test] async fn unused_braces_test() { assert_eq!(1 + 1, 2) } // https://github.com/tokio-rs/tokio/pull/3766#issuecomment-835508651 #[std::prelude::v1::test] fn trait_method() { trait A { fn f(self); fn g(self); } impl A for () { #[tokio::main] async fn f(self) { self.g() } fn g(self) {} } ().f() } // https://github.com/tokio-rs/tokio/issues/4175 #[tokio::main] pub async fn issue_4175_main_1() -> ! { panic!(); } #[tokio::main] pub async fn issue_4175_main_2() -> std::io::Result<()> { panic!(); } #[allow(unreachable_code)] #[tokio::test] pub async fn issue_4175_test() -> std::io::Result<()> { return Ok(()); panic!(); } // https://github.com/tokio-rs/tokio/issues/4175 #[allow(clippy::let_unit_value)] pub mod clippy_semicolon_if_nothing_returned { #![deny(clippy::semicolon_if_nothing_returned)] #[tokio::main] pub async fn local() { let _x = (); } #[tokio::main] pub async fn item() { fn _f() {} } #[tokio::main] pub async fn semi() { panic!(); } #[tokio::main] pub async fn empty() { // To trigger clippy::semicolon_if_nothing_returned lint, the block needs to contain newline. } } // https://github.com/tokio-rs/tokio/issues/5243 pub mod issue_5243 { macro_rules! mac { (async fn $name:ident() $b:block) => { #[::tokio::test] async fn $name() { $b } }; } mac!( async fn foo() {} ); } #[cfg(tokio_unstable)] pub mod macro_rt_arg_unhandled_panic { use tokio_test::assert_err; #[tokio::test(flavor = "current_thread", unhandled_panic = "shutdown_runtime")] #[should_panic] async fn unhandled_panic_shutdown_runtime() { let _ = tokio::spawn(async { panic!("This panic should shutdown the runtime."); }) .await; } #[tokio::test(flavor = "current_thread", unhandled_panic = "ignore")] async fn unhandled_panic_ignore() { let rt = tokio::spawn(async { panic!("This panic should be forwarded to rt as an error."); }) .await; assert_err!(rt); } }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/fs_try_exists.rs
tokio/tests/fs_try_exists.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // WASI does not support all fs operations use tempfile::tempdir; use tokio::fs; #[tokio::test] #[cfg_attr(miri, ignore)] // No `chmod` in miri. async fn try_exists() { let dir = tempdir().unwrap(); let existing_path = dir.path().join("foo.txt"); fs::write(&existing_path, b"Hello File!").await.unwrap(); let nonexisting_path = dir.path().join("bar.txt"); assert!(fs::try_exists(existing_path).await.unwrap()); assert!(!fs::try_exists(nonexisting_path).await.unwrap()); // FreeBSD root user always has permission to stat. #[cfg(all( unix, not(any(target_os = "freebsd", all(tokio_unstable, feature = "io-uring"))) ))] { use std::os::unix::prelude::PermissionsExt; let permission_denied_directory_path = dir.path().join("baz"); fs::create_dir(&permission_denied_directory_path) .await .unwrap(); let permission_denied_file_path = permission_denied_directory_path.join("baz.txt"); fs::write(&permission_denied_file_path, b"Hello File!") .await .unwrap(); let mut perms = tokio::fs::metadata(&permission_denied_directory_path) .await .unwrap() .permissions(); perms.set_mode(0o244); fs::set_permissions(&permission_denied_directory_path, perms) .await .unwrap(); let permission_denied_result = fs::try_exists(permission_denied_file_path).await; assert_eq!( permission_denied_result.err().unwrap().kind(), std::io::ErrorKind::PermissionDenied ); } }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_read_line.rs
tokio/tests/io_read_line.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use std::io::ErrorKind; use tokio::io::{AsyncBufReadExt, BufReader, Error}; use tokio_test::{assert_ok, io::Builder}; use std::io::Cursor; #[tokio::test] async fn read_line() { let mut buf = String::new(); let mut rd = Cursor::new(b"hello\nworld\n\n"); let n = assert_ok!(rd.read_line(&mut buf).await); assert_eq!(n, 6); assert_eq!(buf, "hello\n"); buf.clear(); let n = assert_ok!(rd.read_line(&mut buf).await); assert_eq!(n, 6); assert_eq!(buf, "world\n"); buf.clear(); let n = assert_ok!(rd.read_line(&mut buf).await); assert_eq!(n, 1); assert_eq!(buf, "\n"); buf.clear(); let n = assert_ok!(rd.read_line(&mut buf).await); assert_eq!(n, 0); assert_eq!(buf, ""); } #[tokio::test] async fn read_line_not_all_ready() { let mock = Builder::new() .read(b"Hello Wor") .read(b"ld\nFizzBuz") .read(b"z\n1\n2") .build(); let mut read = BufReader::new(mock); let mut line = "We say ".to_string(); let bytes = read.read_line(&mut line).await.unwrap(); assert_eq!(bytes, "Hello World\n".len()); assert_eq!(line.as_str(), "We say Hello World\n"); line = "I solve ".to_string(); let bytes = read.read_line(&mut line).await.unwrap(); assert_eq!(bytes, "FizzBuzz\n".len()); assert_eq!(line.as_str(), "I solve FizzBuzz\n"); line.clear(); let bytes = read.read_line(&mut line).await.unwrap(); assert_eq!(bytes, 2); assert_eq!(line.as_str(), "1\n"); line.clear(); let bytes = read.read_line(&mut line).await.unwrap(); assert_eq!(bytes, 1); assert_eq!(line.as_str(), "2"); } #[tokio::test] async fn read_line_invalid_utf8() { let mock = Builder::new().read(b"Hello Wor\xffld.\n").build(); let mut read = BufReader::new(mock); let mut line = "Foo".to_string(); let err = read.read_line(&mut line).await.expect_err("Should fail"); assert_eq!(err.kind(), ErrorKind::InvalidData); assert_eq!(err.to_string(), "stream did not contain valid UTF-8"); assert_eq!(line.as_str(), "Foo"); } #[tokio::test] async fn read_line_fail() { let mock = Builder::new() .read(b"Hello Wor") .read_error(Error::new(ErrorKind::Other, "The world has no end")) .build(); let mut read = BufReader::new(mock); let mut line = "Foo".to_string(); let err = read.read_line(&mut line).await.expect_err("Should fail"); assert_eq!(err.kind(), ErrorKind::Other); assert_eq!(err.to_string(), "The world has no end"); assert_eq!(line.as_str(), "FooHello Wor"); } #[tokio::test] async fn read_line_fail_and_utf8_fail() { let mock = Builder::new() .read(b"Hello Wor") .read(b"\xff\xff\xff") .read_error(Error::new(ErrorKind::Other, "The world has no end")) .build(); let mut read = BufReader::new(mock); let mut line = "Foo".to_string(); let err = read.read_line(&mut line).await.expect_err("Should fail"); assert_eq!(err.kind(), ErrorKind::Other); assert_eq!(err.to_string(), "The world has no end"); assert_eq!(line.as_str(), "Foo"); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/sync_broadcast_weak.rs
tokio/tests/sync_broadcast_weak.rs
#![allow(clippy::redundant_clone)] #![warn(rust_2018_idioms)] #![cfg(feature = "sync")] #[cfg(all(target_family = "wasm", not(target_os = "wasi")))] use wasm_bindgen_test::wasm_bindgen_test as test; use tokio::sync::broadcast::{self, channel}; #[tokio::test] async fn weak_sender() { let (tx, mut rx) = channel(11); let tx_weak = tokio::spawn(async move { let tx_weak = tx.clone().downgrade(); for i in 0..10 { if tx.send(i).is_err() { return None; } } let tx2 = tx_weak .upgrade() .expect("expected to be able to upgrade tx_weak"); let _ = tx2.send(20); let tx_weak = tx2.downgrade(); Some(tx_weak) }) .await .unwrap(); for i in 0..12 { let recvd = rx.recv().await; match recvd { Ok(msg) => { if i == 10 { assert_eq!(msg, 20); } } Err(_) => { assert_eq!(i, 11); break; } } } let tx_weak = tx_weak.unwrap(); let upgraded = tx_weak.upgrade(); assert!(upgraded.is_none()); } // Tests that a `WeakSender` fails to upgrade when no other `Sender` exists. #[test] fn downgrade_upgrade_sender_failure() { let (tx, _rx) = broadcast::channel::<i32>(1); let weak_tx = tx.downgrade(); drop(tx); assert!(weak_tx.upgrade().is_none()); } // Tests that a `WeakSender` cannot be upgraded after a `Sender` was dropped, // which existed at the time of the `downgrade` call. #[test] fn downgrade_drop_upgrade() { let (tx, _rx) = broadcast::channel::<i32>(1); // the cloned `Tx` is dropped right away let weak_tx = tx.clone().downgrade(); drop(tx); assert!(weak_tx.upgrade().is_none()); } // Tests that `downgrade` does not change the `strong_count` of the channel. #[test] fn test_tx_count_weak_sender() { let (tx, _rx) = broadcast::channel::<i32>(1); let tx_weak = tx.downgrade(); let tx_weak2 = tx.downgrade(); assert_eq!(tx.strong_count(), 1); assert_eq!(tx.weak_count(), 2); drop(tx); assert!(tx_weak.upgrade().is_none()); assert!(tx_weak2.upgrade().is_none()); assert_eq!(tx_weak.strong_count(), 0); assert_eq!(tx_weak.weak_count(), 2); } #[tokio::test] async fn test_rx_is_closed_when_dropping_all_senders_except_weak_senders() { let (tx, rx) = broadcast::channel::<()>(10); let weak_sender = tx.clone().downgrade(); drop(tx); // is_closed should return true after dropping all senders except for a weak sender. // The strong count should be 0 while the weak count should remain at 1. assert_eq!(weak_sender.strong_count(), 0); assert_eq!(weak_sender.weak_count(), 1); assert!(rx.is_closed()); } #[tokio::test] async fn sender_strong_count_when_cloned() { let (tx, rx) = broadcast::channel::<()>(1); let tx2 = tx.clone(); assert_eq!(tx.strong_count(), 2); assert_eq!(tx2.strong_count(), 2); assert_eq!(rx.sender_strong_count(), 2); } #[tokio::test] async fn sender_weak_count_when_downgraded() { let (tx, _rx) = broadcast::channel::<()>(1); let weak = tx.downgrade(); assert_eq!(tx.weak_count(), 1); assert_eq!(weak.weak_count(), 1); } #[tokio::test] async fn sender_strong_count_when_dropped() { let (tx, rx) = broadcast::channel::<()>(1); let tx2 = tx.clone(); drop(tx2); assert_eq!(tx.strong_count(), 1); assert_eq!(rx.sender_strong_count(), 1); } #[tokio::test] async fn sender_weak_count_when_dropped() { let (tx, rx) = broadcast::channel::<()>(1); let weak = tx.downgrade(); drop(weak); assert_eq!(tx.weak_count(), 0); assert_eq!(rx.sender_weak_count(), 0); } #[tokio::test] async fn sender_strong_and_weak_conut() { let (tx, rx) = broadcast::channel::<()>(1); let tx2 = tx.clone(); let weak = tx.downgrade(); let weak2 = tx2.downgrade(); assert_eq!(tx.strong_count(), 2); assert_eq!(tx2.strong_count(), 2); assert_eq!(weak.strong_count(), 2); assert_eq!(weak2.strong_count(), 2); assert_eq!(rx.sender_strong_count(), 2); assert_eq!(tx.weak_count(), 2); assert_eq!(tx2.weak_count(), 2); assert_eq!(weak.weak_count(), 2); assert_eq!(weak2.weak_count(), 2); assert_eq!(rx.sender_weak_count(), 2); drop(tx2); drop(weak2); assert_eq!(tx.strong_count(), 1); assert_eq!(weak.strong_count(), 1); assert_eq!(rx.sender_strong_count(), 1); assert_eq!(tx.weak_count(), 1); assert_eq!(weak.weak_count(), 1); assert_eq!(rx.sender_weak_count(), 1); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/rt_local.rs
tokio/tests/rt_local.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", tokio_unstable))] use std::panic; use tokio::runtime::LocalOptions; use tokio::task::spawn_local; use tokio::task::LocalSet; #[test] fn test_spawn_local_in_runtime() { let rt = rt(); let res = rt.block_on(async move { let (tx, rx) = tokio::sync::oneshot::channel(); spawn_local(async { tokio::task::yield_now().await; tx.send(5).unwrap(); }); rx.await.unwrap() }); assert_eq!(res, 5); } #[test] fn test_spawn_from_handle() { let rt = rt(); let (tx, rx) = tokio::sync::oneshot::channel(); rt.handle().spawn(async { tokio::task::yield_now().await; tx.send(5).unwrap(); }); let res = rt.block_on(async move { rx.await.unwrap() }); assert_eq!(res, 5); } #[test] fn test_spawn_local_on_runtime_object() { let rt = rt(); let (tx, rx) = tokio::sync::oneshot::channel(); rt.spawn_local(async { tokio::task::yield_now().await; tx.send(5).unwrap(); }); let res = rt.block_on(async move { rx.await.unwrap() }); assert_eq!(res, 5); } #[test] fn test_spawn_local_from_guard() { let rt = rt(); let (tx, rx) = tokio::sync::oneshot::channel(); let _guard = rt.enter(); spawn_local(async { tokio::task::yield_now().await; tx.send(5).unwrap(); }); let res = rt.block_on(async move { rx.await.unwrap() }); assert_eq!(res, 5); } #[test] #[cfg_attr(target_family = "wasm", ignore)] // threads not supported fn test_spawn_from_guard_other_thread() { let (tx, rx) = std::sync::mpsc::channel(); std::thread::spawn(move || { let rt = rt(); let handle = rt.handle().clone(); tx.send(handle).unwrap(); }); let handle = rx.recv().unwrap(); let _guard = handle.enter(); tokio::spawn(async {}); } #[test] #[should_panic = "Local tasks can only be spawned on a LocalRuntime from the thread the runtime was created on"] #[cfg_attr(target_family = "wasm", ignore)] // threads not supported fn test_spawn_local_from_guard_other_thread() { let (tx, rx) = std::sync::mpsc::channel(); std::thread::spawn(move || { let rt = rt(); let handle = rt.handle().clone(); tx.send(handle).unwrap(); }); let handle = rx.recv().unwrap(); let _guard = handle.enter(); spawn_local(async {}); } // This test guarantees that **`tokio::task::spawn_local` panics** when it is invoked // from a thread that is *not* running the `LocalRuntime` / `LocalSet` to which // the task would belong. // The test creates a `LocalRuntime` and `LocalSet`, drives the `LocalSet` on the `LocalRuntime`'s thread, // then spawns a **separate OS thread** and tries to call // `tokio::task::spawn_local` there. `std::panic::catch_unwind` is then used // to capture the panic and to assert that it indeed occurs. #[test] #[cfg_attr(target_family = "wasm", ignore)] // threads not supported fn test_spawn_local_panic() { let rt = rt(); let local = LocalSet::new(); rt.block_on(local.run_until(async { let thread_result = std::thread::spawn(|| { let panic_result = panic::catch_unwind(|| { let _jh = tokio::task::spawn_local(async { println!("you will never see this line"); }); }); assert!(panic_result.is_err(), "Expected panic, but none occurred"); }) .join(); assert!(thread_result.is_ok(), "Thread itself panicked unexpectedly"); })); } #[test] #[should_panic = "`spawn_local` called from outside of a `task::LocalSet` or `runtime::LocalRuntime`"] fn test_spawn_local_in_current_thread_runtime() { let rt = tokio::runtime::Builder::new_current_thread() .build() .unwrap(); rt.block_on(async move { spawn_local(async {}); }) } #[test] #[should_panic = "`spawn_local` called from outside of a `task::LocalSet` or `runtime::LocalRuntime`"] fn test_spawn_local_in_multi_thread_runtime() { let rt = tokio::runtime::Builder::new_multi_thread().build().unwrap(); rt.block_on(async move { spawn_local(async {}); }) } fn rt() -> tokio::runtime::LocalRuntime { tokio::runtime::Builder::new_current_thread() .enable_all() .build_local(LocalOptions::default()) .unwrap() }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/signal_multi_rt.rs
tokio/tests/signal_multi_rt.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] #![cfg(unix)] #![cfg(not(miri))] // No `sigaction` on Miri. mod support { pub mod signal; } use support::signal::send_signal; use tokio::runtime::Runtime; use tokio::signal::unix::{signal, SignalKind}; use std::sync::mpsc::channel; use std::thread; #[test] fn multi_loop() { // An "ordinary" (non-future) channel let (sender, receiver) = channel(); // Run multiple times, to make sure there are no race conditions for _ in 0..10 { // Run multiple event loops, each one in its own thread let threads: Vec<_> = (0..4) .map(|_| { let sender = sender.clone(); thread::spawn(move || { let rt = rt(); let _ = rt.block_on(async { let mut signal = signal(SignalKind::hangup()).unwrap(); sender.send(()).unwrap(); signal.recv().await }); }) }) .collect(); // Wait for them to declare they're ready for &_ in threads.iter() { receiver.recv().unwrap(); } // Send a signal send_signal(libc::SIGHUP); // Make sure the threads terminated correctly for t in threads { t.join().unwrap(); } } } fn rt() -> Runtime { tokio::runtime::Builder::new_current_thread() .enable_all() .build() .unwrap() }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_join.rs
tokio/tests/io_join.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use tokio::io::{join, AsyncRead, AsyncReadExt, AsyncWrite, AsyncWriteExt, Join, ReadBuf}; use std::io; use std::pin::Pin; use std::task::{Context, Poll}; struct R; impl AsyncRead for R { fn poll_read( self: Pin<&mut Self>, _cx: &mut Context<'_>, buf: &mut ReadBuf<'_>, ) -> Poll<io::Result<()>> { buf.put_slice(b"z"); Poll::Ready(Ok(())) } } struct W; impl AsyncWrite for W { fn poll_write( self: Pin<&mut Self>, _cx: &mut Context<'_>, _buf: &[u8], ) -> Poll<Result<usize, io::Error>> { Poll::Ready(Ok(1)) } fn poll_flush(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<Result<(), io::Error>> { Poll::Ready(Ok(())) } fn poll_shutdown(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<Result<(), io::Error>> { Poll::Ready(Ok(())) } fn poll_write_vectored( self: Pin<&mut Self>, _cx: &mut Context<'_>, _bufs: &[io::IoSlice<'_>], ) -> Poll<Result<usize, io::Error>> { Poll::Ready(Ok(2)) } fn is_write_vectored(&self) -> bool { true } } #[test] fn is_send_and_sync() { fn assert_bound<T: Send + Sync>() {} assert_bound::<Join<W, R>>(); } #[test] fn method_delegation() { let mut rw = join(R, W); let mut buf = [0; 1]; tokio_test::block_on(async move { assert_eq!(1, rw.read(&mut buf).await.unwrap()); assert_eq!(b'z', buf[0]); assert_eq!(1, rw.write(b"x").await.unwrap()); assert_eq!( 2, rw.write_vectored(&[io::IoSlice::new(b"x")]).await.unwrap() ); assert!(rw.is_write_vectored()); assert!(rw.flush().await.is_ok()); assert!(rw.shutdown().await.is_ok()); }); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_read.rs
tokio/tests/io_read.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // Wasi does not support panic recovery use tokio::io::{AsyncRead, AsyncReadExt, ReadBuf}; use tokio_test::assert_ok; use std::io; use std::pin::Pin; use std::task::{Context, Poll}; mod support { pub(crate) mod leaked_buffers; } use support::leaked_buffers::LeakedBuffers; #[tokio::test] async fn read() { #[derive(Default)] struct Rd { poll_cnt: usize, } impl AsyncRead for Rd { fn poll_read( mut self: Pin<&mut Self>, _cx: &mut Context<'_>, buf: &mut ReadBuf<'_>, ) -> Poll<io::Result<()>> { assert_eq!(0, self.poll_cnt); self.poll_cnt += 1; buf.put_slice(b"hello world"); Poll::Ready(Ok(())) } } let mut buf = Box::new([0; 11]); let mut rd = Rd::default(); let n = assert_ok!(rd.read(&mut buf[..]).await); assert_eq!(n, 11); assert_eq!(buf[..], b"hello world"[..]); } struct BadAsyncRead { leaked_buffers: LeakedBuffers, } impl BadAsyncRead { fn new() -> Self { Self { leaked_buffers: LeakedBuffers::new(), } } } impl AsyncRead for BadAsyncRead { fn poll_read( mut self: Pin<&mut Self>, _cx: &mut Context<'_>, buf: &mut ReadBuf<'_>, ) -> Poll<io::Result<()>> { *buf = ReadBuf::new(unsafe { self.leaked_buffers.create(buf.capacity()) }); buf.advance(buf.capacity()); Poll::Ready(Ok(())) } } #[tokio::test] #[should_panic] async fn read_buf_bad_async_read() { let mut buf = Vec::with_capacity(10); BadAsyncRead::new().read_buf(&mut buf).await.unwrap(); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/sync_once_cell.rs
tokio/tests/sync_once_cell.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use std::mem; use std::sync::atomic::{AtomicU32, Ordering}; use std::sync::Arc; use std::time::Duration; use tokio::runtime; use tokio::sync::OnceCell; use tokio::sync::SetError; use tokio::time; struct Foo { value: Arc<AtomicU32>, } impl Drop for Foo { fn drop(&mut self) { self.value.fetch_add(1, Ordering::Release); } } impl From<Arc<AtomicU32>> for Foo { fn from(value: Arc<AtomicU32>) -> Self { Foo { value } } } #[test] fn drop_cell() { let num_drops = Arc::new(AtomicU32::new(0)); { let once_cell = OnceCell::new(); let prev = once_cell.set(Foo::from(num_drops.clone())); assert!(prev.is_ok()) } assert!(num_drops.load(Ordering::Acquire) == 1); } #[test] fn drop_cell_new_with() { let num_drops = Arc::new(AtomicU32::new(0)); { let once_cell = OnceCell::new_with(Some(Foo::from(num_drops.clone()))); assert!(once_cell.initialized()); } assert!(num_drops.load(Ordering::Acquire) == 1); } #[test] fn drop_into_inner() { let num_drops = Arc::new(AtomicU32::new(0)); let once_cell = OnceCell::new(); assert!(once_cell.set(Foo::from(num_drops.clone())).is_ok()); let fooer = once_cell.into_inner(); let count = num_drops.load(Ordering::Acquire); assert!(count == 0); drop(fooer); let count = num_drops.load(Ordering::Acquire); assert!(count == 1); } #[test] fn drop_into_inner_new_with() { let num_drops = Arc::new(AtomicU32::new(0)); let fooer = Foo::from(num_drops.clone()); let once_cell = OnceCell::new_with(Some(fooer)); let fooer = once_cell.into_inner(); let count = num_drops.load(Ordering::Acquire); assert!(count == 0); mem::drop(fooer); let count = num_drops.load(Ordering::Acquire); assert!(count == 1); } #[test] fn from() { let cell = OnceCell::from(2); assert_eq!(*cell.get().unwrap(), 2); } async fn func1() -> u32 { 5 } async fn func2() -> u32 { time::sleep(Duration::from_millis(1)).await; 10 } async fn func_err() -> Result<u32, ()> { Err(()) } async fn func_ok() -> Result<u32, ()> { Ok(10) } async fn func_panic() -> u32 { time::sleep(Duration::from_millis(1)).await; panic!(); } async fn sleep_and_set() -> u32 { // Simulate sleep by pausing time and waiting for another thread to // resume clock when calling `set`, then finding the cell being initialized // by this call time::sleep(Duration::from_millis(2)).await; 5 } async fn advance_time_and_set(cell: &'static OnceCell<u32>, v: u32) -> Result<(), SetError<u32>> { time::advance(Duration::from_millis(1)).await; cell.set(v) } #[test] fn get_or_init() { let rt = runtime::Builder::new_current_thread() .enable_time() .start_paused(true) .build() .unwrap(); static ONCE: OnceCell<u32> = OnceCell::const_new(); rt.block_on(async { let handle1 = rt.spawn(async { ONCE.get_or_init(func1).await }); let handle2 = rt.spawn(async { ONCE.get_or_init(func2).await }); time::advance(Duration::from_millis(1)).await; time::resume(); let result1 = handle1.await.unwrap(); let result2 = handle2.await.unwrap(); assert_eq!(*result1, 5); assert_eq!(*result2, 5); }); } #[test] fn get_or_init_panic() { let rt = runtime::Builder::new_current_thread() .enable_time() .build() .unwrap(); static ONCE: OnceCell<u32> = OnceCell::const_new(); rt.block_on(async { time::pause(); let handle1 = rt.spawn(async { ONCE.get_or_init(func1).await }); let handle2 = rt.spawn(async { ONCE.get_or_init(func_panic).await }); time::advance(Duration::from_millis(1)).await; let result1 = handle1.await.unwrap(); let result2 = handle2.await.unwrap(); assert_eq!(*result1, 5); assert_eq!(*result2, 5); }); } #[test] fn set_and_get() { let rt = runtime::Builder::new_current_thread() .enable_time() .build() .unwrap(); static ONCE: OnceCell<u32> = OnceCell::const_new(); rt.block_on(async { let _ = rt.spawn(async { ONCE.set(5) }).await; let value = ONCE.get().unwrap(); assert_eq!(*value, 5); }); } #[test] fn get_uninit() { static ONCE: OnceCell<u32> = OnceCell::const_new(); let uninit = ONCE.get(); assert!(uninit.is_none()); } #[test] fn set_twice() { static ONCE: OnceCell<u32> = OnceCell::const_new(); let first = ONCE.set(5); assert_eq!(first, Ok(())); let second = ONCE.set(6); assert!(second.err().unwrap().is_already_init_err()); } #[test] fn set_while_initializing() { let rt = runtime::Builder::new_current_thread() .enable_time() .build() .unwrap(); static ONCE: OnceCell<u32> = OnceCell::const_new(); rt.block_on(async { time::pause(); let handle1 = rt.spawn(async { ONCE.get_or_init(sleep_and_set).await }); let handle2 = rt.spawn(async { advance_time_and_set(&ONCE, 10).await }); time::advance(Duration::from_millis(2)).await; let result1 = handle1.await.unwrap(); let result2 = handle2.await.unwrap(); assert_eq!(*result1, 5); assert!(result2.err().unwrap().is_initializing_err()); }); } #[test] fn get_or_try_init() { let rt = runtime::Builder::new_current_thread() .enable_time() .start_paused(true) .build() .unwrap(); static ONCE: OnceCell<u32> = OnceCell::const_new(); rt.block_on(async { let handle1 = rt.spawn(async { ONCE.get_or_try_init(func_err).await }); let handle2 = rt.spawn(async { ONCE.get_or_try_init(func_ok).await }); time::advance(Duration::from_millis(1)).await; time::resume(); let result1 = handle1.await.unwrap(); assert!(result1.is_err()); let result2 = handle2.await.unwrap(); assert_eq!(*result2.unwrap(), 10); }); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/sync_barrier.rs
tokio/tests/sync_barrier.rs
#![allow(clippy::unnecessary_operation)] #![warn(rust_2018_idioms)] #![cfg(feature = "sync")] #[cfg(all(target_family = "wasm", not(target_os = "wasi")))] use wasm_bindgen_test::wasm_bindgen_test as test; use tokio::sync::Barrier; use tokio_test::task::spawn; use tokio_test::{assert_pending, assert_ready}; struct IsSend<T: Send>(T); #[test] fn barrier_future_is_send() { let b = Barrier::new(0); IsSend(b.wait()); } #[test] fn zero_does_not_block() { let b = Barrier::new(0); { let mut w = spawn(b.wait()); let wr = assert_ready!(w.poll()); assert!(wr.is_leader()); } { let mut w = spawn(b.wait()); let wr = assert_ready!(w.poll()); assert!(wr.is_leader()); } } #[test] fn single() { let b = Barrier::new(1); { let mut w = spawn(b.wait()); let wr = assert_ready!(w.poll()); assert!(wr.is_leader()); } { let mut w = spawn(b.wait()); let wr = assert_ready!(w.poll()); assert!(wr.is_leader()); } { let mut w = spawn(b.wait()); let wr = assert_ready!(w.poll()); assert!(wr.is_leader()); } } #[test] fn tango() { let b = Barrier::new(2); let mut w1 = spawn(b.wait()); assert_pending!(w1.poll()); let mut w2 = spawn(b.wait()); let wr2 = assert_ready!(w2.poll()); let wr1 = assert_ready!(w1.poll()); assert!(wr1.is_leader() || wr2.is_leader()); assert!(!(wr1.is_leader() && wr2.is_leader())); } #[test] fn lots() { let b = Barrier::new(100); for _ in 0..10 { let mut wait = Vec::new(); for _ in 0..99 { let mut w = spawn(b.wait()); assert_pending!(w.poll()); wait.push(w); } for w in &mut wait { assert_pending!(w.poll()); } // pass the barrier let mut w = spawn(b.wait()); let mut found_leader = assert_ready!(w.poll()).is_leader(); for mut w in wait { let wr = assert_ready!(w.poll()); if wr.is_leader() { assert!(!found_leader); found_leader = true; } } assert!(found_leader); } }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/sync_mpsc_weak.rs
tokio/tests/sync_mpsc_weak.rs
#![allow(clippy::redundant_clone)] #![warn(rust_2018_idioms)] #![cfg(feature = "sync")] #[cfg(all(target_family = "wasm", not(target_os = "wasi")))] use wasm_bindgen_test::wasm_bindgen_test as test; use std::sync::atomic::AtomicUsize; use std::sync::atomic::Ordering::{Acquire, Release}; use tokio::sync::mpsc::{self, channel, unbounded_channel}; use tokio::sync::oneshot; #[tokio::test] async fn weak_sender() { let (tx, mut rx) = channel(11); let tx_weak = tokio::spawn(async move { let tx_weak = tx.clone().downgrade(); for i in 0..10 { if tx.send(i).await.is_err() { return None; } } let tx2 = tx_weak .upgrade() .expect("expected to be able to upgrade tx_weak"); let _ = tx2.send(20).await; let tx_weak = tx2.downgrade(); Some(tx_weak) }) .await .unwrap(); for i in 0..12 { let recvd = rx.recv().await; match recvd { Some(msg) => { if i == 10 { assert_eq!(msg, 20); } } None => { assert_eq!(i, 11); break; } } } let tx_weak = tx_weak.unwrap(); let upgraded = tx_weak.upgrade(); assert!(upgraded.is_none()); } #[tokio::test] async fn actor_weak_sender() { pub struct MyActor { receiver: mpsc::Receiver<ActorMessage>, sender: mpsc::WeakSender<ActorMessage>, next_id: u32, pub received_self_msg: bool, } enum ActorMessage { GetUniqueId { respond_to: oneshot::Sender<u32> }, SelfMessage {}, } impl MyActor { fn new( receiver: mpsc::Receiver<ActorMessage>, sender: mpsc::WeakSender<ActorMessage>, ) -> Self { MyActor { receiver, sender, next_id: 0, received_self_msg: false, } } fn handle_message(&mut self, msg: ActorMessage) { match msg { ActorMessage::GetUniqueId { respond_to } => { self.next_id += 1; // The `let _ =` ignores any errors when sending. // // This can happen if the `select!` macro is used // to cancel waiting for the response. let _ = respond_to.send(self.next_id); } ActorMessage::SelfMessage { .. } => { self.received_self_msg = true; } } } async fn send_message_to_self(&mut self) { let msg = ActorMessage::SelfMessage {}; let sender = self.sender.clone(); // cannot move self.sender here if let Some(sender) = sender.upgrade() { let _ = sender.send(msg).await; self.sender = sender.downgrade(); } } async fn run(&mut self) { let mut i = 0; while let Some(msg) = self.receiver.recv().await { self.handle_message(msg); if i == 0 { self.send_message_to_self().await; } i += 1 } assert!(self.received_self_msg); } } #[derive(Clone)] pub struct MyActorHandle { sender: mpsc::Sender<ActorMessage>, } impl MyActorHandle { pub fn new() -> (Self, MyActor) { let (sender, receiver) = mpsc::channel(8); let actor = MyActor::new(receiver, sender.clone().downgrade()); (Self { sender }, actor) } pub async fn get_unique_id(&self) -> u32 { let (send, recv) = oneshot::channel(); let msg = ActorMessage::GetUniqueId { respond_to: send }; // Ignore send errors. If this send fails, so does the // recv.await below. There's no reason to check the // failure twice. let _ = self.sender.send(msg).await; recv.await.expect("Actor task has been killed") } } let (handle, mut actor) = MyActorHandle::new(); let actor_handle = tokio::spawn(async move { actor.run().await }); let _ = tokio::spawn(async move { let _ = handle.get_unique_id().await; drop(handle); }) .await; let _ = actor_handle.await; } static NUM_DROPPED: AtomicUsize = AtomicUsize::new(0); #[derive(Debug)] struct Msg; impl Drop for Msg { fn drop(&mut self) { NUM_DROPPED.fetch_add(1, Release); } } // Tests that no pending messages are put onto the channel after `Rx` was // dropped. // // Note: After the introduction of `WeakSender`, which internally // used `Arc` and doesn't call a drop of the channel after the last strong // `Sender` was dropped while more than one `WeakSender` remains, we want to // ensure that no messages are kept in the channel, which were sent after // the receiver was dropped. #[tokio::test] async fn test_msgs_dropped_on_rx_drop() { let (tx, mut rx) = mpsc::channel(3); tx.send(Msg {}).await.unwrap(); tx.send(Msg {}).await.unwrap(); // This msg will be pending and should be dropped when `rx` is dropped let sent_fut = tx.send(Msg {}); let _ = rx.recv().await.unwrap(); let _ = rx.recv().await.unwrap(); sent_fut.await.unwrap(); drop(rx); assert_eq!(NUM_DROPPED.load(Acquire), 3); // This msg will not be put onto `Tx` list anymore, since `Rx` is closed. assert!(tx.send(Msg {}).await.is_err()); assert_eq!(NUM_DROPPED.load(Acquire), 4); } // Tests that a `WeakSender` is upgradeable when other `Sender`s exist. #[test] fn downgrade_upgrade_sender_success() { let (tx, _rx) = mpsc::channel::<i32>(1); let weak_tx = tx.downgrade(); assert!(weak_tx.upgrade().is_some()); } // Tests that a `WeakSender` fails to upgrade when no other `Sender` exists. #[test] fn downgrade_upgrade_sender_failure() { let (tx, _rx) = mpsc::channel::<i32>(1); let weak_tx = tx.downgrade(); drop(tx); assert!(weak_tx.upgrade().is_none()); } // Tests that a `WeakSender` cannot be upgraded after a `Sender` was dropped, // which existed at the time of the `downgrade` call. #[test] fn downgrade_drop_upgrade() { let (tx, _rx) = mpsc::channel::<i32>(1); // the cloned `Tx` is dropped right away let weak_tx = tx.clone().downgrade(); drop(tx); assert!(weak_tx.upgrade().is_none()); } // Tests that we can upgrade a weak sender with an outstanding permit // but no other strong senders. #[tokio::test] async fn downgrade_get_permit_upgrade_no_senders() { let (tx, _rx) = mpsc::channel::<i32>(1); let weak_tx = tx.downgrade(); let _permit = tx.reserve_owned().await.unwrap(); assert!(weak_tx.upgrade().is_some()); } // Tests that you can downgrade and upgrade a sender with an outstanding permit // but no other senders left. #[tokio::test] async fn downgrade_upgrade_get_permit_no_senders() { let (tx, _rx) = mpsc::channel::<i32>(1); let tx2 = tx.clone(); let _permit = tx.reserve_owned().await.unwrap(); let weak_tx = tx2.downgrade(); drop(tx2); assert!(weak_tx.upgrade().is_some()); } // Tests that `downgrade` does not change the `tx_count` of the channel. #[test] fn test_tx_count_weak_sender() { let (tx, _rx) = mpsc::channel::<i32>(1); let tx_weak = tx.downgrade(); let tx_weak2 = tx.downgrade(); drop(tx); assert!(tx_weak.upgrade().is_none() && tx_weak2.upgrade().is_none()); } #[tokio::test] async fn weak_unbounded_sender() { let (tx, mut rx) = unbounded_channel(); let tx_weak = tokio::spawn(async move { let tx_weak = tx.clone().downgrade(); for i in 0..10 { if tx.send(i).is_err() { return None; } } let tx2 = tx_weak .upgrade() .expect("expected to be able to upgrade tx_weak"); let _ = tx2.send(20); let tx_weak = tx2.downgrade(); Some(tx_weak) }) .await .unwrap(); for i in 0..12 { let recvd = rx.recv().await; match recvd { Some(msg) => { if i == 10 { assert_eq!(msg, 20); } } None => { assert_eq!(i, 11); break; } } } let tx_weak = tx_weak.unwrap(); let upgraded = tx_weak.upgrade(); assert!(upgraded.is_none()); } #[tokio::test] async fn actor_weak_unbounded_sender() { pub struct MyActor { receiver: mpsc::UnboundedReceiver<ActorMessage>, sender: mpsc::WeakUnboundedSender<ActorMessage>, next_id: u32, pub received_self_msg: bool, } enum ActorMessage { GetUniqueId { respond_to: oneshot::Sender<u32> }, SelfMessage {}, } impl MyActor { fn new( receiver: mpsc::UnboundedReceiver<ActorMessage>, sender: mpsc::WeakUnboundedSender<ActorMessage>, ) -> Self { MyActor { receiver, sender, next_id: 0, received_self_msg: false, } } fn handle_message(&mut self, msg: ActorMessage) { match msg { ActorMessage::GetUniqueId { respond_to } => { self.next_id += 1; // The `let _ =` ignores any errors when sending. // // This can happen if the `select!` macro is used // to cancel waiting for the response. let _ = respond_to.send(self.next_id); } ActorMessage::SelfMessage { .. } => { self.received_self_msg = true; } } } async fn send_message_to_self(&mut self) { let msg = ActorMessage::SelfMessage {}; let sender = self.sender.clone(); // cannot move self.sender here if let Some(sender) = sender.upgrade() { let _ = sender.send(msg); self.sender = sender.downgrade(); } } async fn run(&mut self) { let mut i = 0; while let Some(msg) = self.receiver.recv().await { self.handle_message(msg); if i == 0 { self.send_message_to_self().await; } i += 1 } assert!(self.received_self_msg); } } #[derive(Clone)] pub struct MyActorHandle { sender: mpsc::UnboundedSender<ActorMessage>, } impl MyActorHandle { pub fn new() -> (Self, MyActor) { let (sender, receiver) = mpsc::unbounded_channel(); let actor = MyActor::new(receiver, sender.clone().downgrade()); (Self { sender }, actor) } pub async fn get_unique_id(&self) -> u32 { let (send, recv) = oneshot::channel(); let msg = ActorMessage::GetUniqueId { respond_to: send }; // Ignore send errors. If this send fails, so does the // recv.await below. There's no reason to check the // failure twice. let _ = self.sender.send(msg); recv.await.expect("Actor task has been killed") } } let (handle, mut actor) = MyActorHandle::new(); let actor_handle = tokio::spawn(async move { actor.run().await }); let _ = tokio::spawn(async move { let _ = handle.get_unique_id().await; drop(handle); }) .await; let _ = actor_handle.await; } static NUM_DROPPED_UNBOUNDED: AtomicUsize = AtomicUsize::new(0); #[derive(Debug)] struct MsgUnbounded; impl Drop for MsgUnbounded { fn drop(&mut self) { NUM_DROPPED_UNBOUNDED.fetch_add(1, Release); } } // Tests that no pending messages are put onto the channel after `Rx` was // dropped. // // Note: After the introduction of `UnboundedWeakSender`, which internally // used `Arc` and doesn't call a drop of the channel after the last strong // `UnboundedSender` was dropped while more than one `UnboundedWeakSender` // remains, we want to ensure that no messages are kept in the channel, which // were sent after the receiver was dropped. #[tokio::test] async fn test_msgs_dropped_on_unbounded_rx_drop() { let (tx, mut rx) = mpsc::unbounded_channel(); tx.send(MsgUnbounded {}).unwrap(); tx.send(MsgUnbounded {}).unwrap(); // This msg will be pending and should be dropped when `rx` is dropped let sent = tx.send(MsgUnbounded {}); let _ = rx.recv().await.unwrap(); let _ = rx.recv().await.unwrap(); sent.unwrap(); drop(rx); assert_eq!(NUM_DROPPED_UNBOUNDED.load(Acquire), 3); // This msg will not be put onto `Tx` list anymore, since `Rx` is closed. assert!(tx.send(MsgUnbounded {}).is_err()); assert_eq!(NUM_DROPPED_UNBOUNDED.load(Acquire), 4); } // Tests that an `WeakUnboundedSender` is upgradeable when other // `UnboundedSender`s exist. #[test] fn downgrade_upgrade_unbounded_sender_success() { let (tx, _rx) = mpsc::unbounded_channel::<i32>(); let weak_tx = tx.downgrade(); assert!(weak_tx.upgrade().is_some()); } // Tests that a `WeakUnboundedSender` fails to upgrade when no other // `UnboundedSender` exists. #[test] fn downgrade_upgrade_unbounded_sender_failure() { let (tx, _rx) = mpsc::unbounded_channel::<i32>(); let weak_tx = tx.downgrade(); drop(tx); assert!(weak_tx.upgrade().is_none()); } // Tests that an `WeakUnboundedSender` cannot be upgraded after an // `UnboundedSender` was dropped, which existed at the time of the `downgrade` call. #[test] fn downgrade_drop_upgrade_unbounded() { let (tx, _rx) = mpsc::unbounded_channel::<i32>(); // the cloned `Tx` is dropped right away let weak_tx = tx.clone().downgrade(); drop(tx); assert!(weak_tx.upgrade().is_none()); } // Tests that `downgrade` does not change the `tx_count` of the channel. #[test] fn test_tx_count_weak_unbounded_sender() { let (tx, _rx) = mpsc::unbounded_channel::<i32>(); let tx_weak = tx.downgrade(); let tx_weak2 = tx.downgrade(); drop(tx); assert!(tx_weak.upgrade().is_none() && tx_weak2.upgrade().is_none()); } #[tokio::test] async fn test_rx_is_closed_when_dropping_all_senders_except_weak_senders() { // is_closed should return true after dropping all senders except for a weak sender let (tx, rx) = mpsc::channel::<()>(10); let _weak_sender = tx.clone().downgrade(); drop(tx); assert!(rx.is_closed()); } #[tokio::test] async fn test_rx_unbounded_is_closed_when_dropping_all_senders_except_weak_senders() { // is_closed should return true after dropping all senders except for a weak sender let (tx, rx) = mpsc::unbounded_channel::<()>(); let _weak_sender = tx.clone().downgrade(); drop(tx); assert!(rx.is_closed()); } #[tokio::test] async fn sender_strong_count_when_cloned() { let (tx, rx) = mpsc::channel::<()>(1); let tx2 = tx.clone(); assert_eq!(tx.strong_count(), 2); assert_eq!(tx2.strong_count(), 2); assert_eq!(rx.sender_strong_count(), 2); } #[tokio::test] async fn sender_weak_count_when_downgraded() { let (tx, _rx) = mpsc::channel::<()>(1); let weak = tx.downgrade(); assert_eq!(tx.weak_count(), 1); assert_eq!(weak.weak_count(), 1); } #[tokio::test] async fn sender_strong_count_when_dropped() { let (tx, rx) = mpsc::channel::<()>(1); let tx2 = tx.clone(); drop(tx2); assert_eq!(tx.strong_count(), 1); assert_eq!(rx.sender_strong_count(), 1); } #[tokio::test] async fn sender_weak_count_when_dropped() { let (tx, rx) = mpsc::channel::<()>(1); let weak = tx.downgrade(); drop(weak); assert_eq!(tx.weak_count(), 0); assert_eq!(rx.sender_weak_count(), 0); } #[tokio::test] async fn sender_strong_and_weak_conut() { let (tx, rx) = mpsc::channel::<()>(1); let tx2 = tx.clone(); let weak = tx.downgrade(); let weak2 = tx2.downgrade(); assert_eq!(tx.strong_count(), 2); assert_eq!(tx2.strong_count(), 2); assert_eq!(weak.strong_count(), 2); assert_eq!(weak2.strong_count(), 2); assert_eq!(rx.sender_strong_count(), 2); assert_eq!(tx.weak_count(), 2); assert_eq!(tx2.weak_count(), 2); assert_eq!(weak.weak_count(), 2); assert_eq!(weak2.weak_count(), 2); assert_eq!(rx.sender_weak_count(), 2); drop(tx2); drop(weak2); assert_eq!(tx.strong_count(), 1); assert_eq!(weak.strong_count(), 1); assert_eq!(rx.sender_strong_count(), 1); assert_eq!(tx.weak_count(), 1); assert_eq!(weak.weak_count(), 1); assert_eq!(rx.sender_weak_count(), 1); } #[tokio::test] async fn unbounded_sender_strong_count_when_cloned() { let (tx, rx) = mpsc::unbounded_channel::<()>(); let tx2 = tx.clone(); assert_eq!(tx.strong_count(), 2); assert_eq!(tx2.strong_count(), 2); assert_eq!(rx.sender_strong_count(), 2); } #[tokio::test] async fn unbounded_sender_weak_count_when_downgraded() { let (tx, rx) = mpsc::unbounded_channel::<()>(); let weak = tx.downgrade(); assert_eq!(tx.weak_count(), 1); assert_eq!(weak.weak_count(), 1); assert_eq!(rx.sender_weak_count(), 1); } #[tokio::test] async fn unbounded_sender_strong_count_when_dropped() { let (tx, rx) = mpsc::unbounded_channel::<()>(); let tx2 = tx.clone(); drop(tx2); assert_eq!(tx.strong_count(), 1); assert_eq!(rx.sender_strong_count(), 1); } #[tokio::test] async fn unbounded_sender_weak_count_when_dropped() { let (tx, rx) = mpsc::unbounded_channel::<()>(); let weak = tx.downgrade(); drop(weak); assert_eq!(tx.weak_count(), 0); assert_eq!(rx.sender_weak_count(), 0); } #[tokio::test] async fn unbounded_sender_strong_and_weak_conut() { let (tx, rx) = mpsc::unbounded_channel::<()>(); let tx2 = tx.clone(); let weak = tx.downgrade(); let weak2 = tx2.downgrade(); assert_eq!(tx.strong_count(), 2); assert_eq!(tx2.strong_count(), 2); assert_eq!(weak.strong_count(), 2); assert_eq!(weak2.strong_count(), 2); assert_eq!(rx.sender_strong_count(), 2); assert_eq!(tx.weak_count(), 2); assert_eq!(tx2.weak_count(), 2); assert_eq!(weak.weak_count(), 2); assert_eq!(weak2.weak_count(), 2); assert_eq!(rx.sender_weak_count(), 2); drop(tx2); drop(weak2); assert_eq!(tx.strong_count(), 1); assert_eq!(weak.strong_count(), 1); assert_eq!(rx.sender_strong_count(), 1); assert_eq!(tx.weak_count(), 1); assert_eq!(weak.weak_count(), 1); assert_eq!(rx.sender_weak_count(), 1); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/fs_copy.rs
tokio/tests/fs_copy.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // WASI does not support all fs operations use tempfile::tempdir; use tokio::fs; #[tokio::test] #[cfg_attr(miri, ignore)] // No `fchmod` in miri. async fn copy() { let dir = tempdir().unwrap(); let source_path = dir.path().join("foo.txt"); let dest_path = dir.path().join("bar.txt"); fs::write(&source_path, b"Hello File!").await.unwrap(); fs::copy(&source_path, &dest_path).await.unwrap(); let from = fs::read(&source_path).await.unwrap(); let to = fs::read(&dest_path).await.unwrap(); assert_eq!(from, to); } #[tokio::test] #[cfg_attr(miri, ignore)] // No `fchmod` in miri. async fn copy_permissions() { let dir = tempdir().unwrap(); let from_path = dir.path().join("foo.txt"); let to_path = dir.path().join("bar.txt"); let from = tokio::fs::File::create(&from_path).await.unwrap(); let mut from_perms = from.metadata().await.unwrap().permissions(); from_perms.set_readonly(true); from.set_permissions(from_perms.clone()).await.unwrap(); tokio::fs::copy(from_path, &to_path).await.unwrap(); let to_perms = tokio::fs::metadata(to_path).await.unwrap().permissions(); assert_eq!(from_perms, to_perms); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/macros_join.rs
tokio/tests/macros_join.rs
#![cfg(feature = "macros")] #![allow(clippy::disallowed_names)] use std::sync::Arc; #[cfg(all(target_family = "wasm", not(target_os = "wasi")))] #[cfg(target_pointer_width = "64")] use wasm_bindgen_test::wasm_bindgen_test as test; #[cfg(all(target_family = "wasm", not(target_os = "wasi")))] use wasm_bindgen_test::wasm_bindgen_test as maybe_tokio_test; #[cfg(not(all(target_family = "wasm", not(target_os = "wasi"))))] use tokio::test as maybe_tokio_test; use tokio::sync::{oneshot, Semaphore}; use tokio_test::{assert_pending, assert_ready, task}; #[maybe_tokio_test] async fn sync_one_lit_expr_comma() { let foo = tokio::join!(async { 1 },); assert_eq!(foo, (1,)); let foo = tokio::join!(biased; async { 1 },); assert_eq!(foo, (1,)); } #[maybe_tokio_test] async fn sync_one_lit_expr_no_comma() { let foo = tokio::join!(async { 1 }); assert_eq!(foo, (1,)); let foo = tokio::join!(biased; async { 1 }); assert_eq!(foo, (1,)); } #[maybe_tokio_test] async fn sync_two_lit_expr_comma() { let foo = tokio::join!(async { 1 }, async { 2 },); assert_eq!(foo, (1, 2)); let foo = tokio::join!(biased; async { 1 }, async { 2 },); assert_eq!(foo, (1, 2)); } #[maybe_tokio_test] async fn sync_two_lit_expr_no_comma() { let foo = tokio::join!(async { 1 }, async { 2 }); assert_eq!(foo, (1, 2)); let foo = tokio::join!(biased; async { 1 }, async { 2 }); assert_eq!(foo, (1, 2)); } #[maybe_tokio_test] async fn two_await() { let (tx1, rx1) = oneshot::channel::<&str>(); let (tx2, rx2) = oneshot::channel::<u32>(); let mut join = task::spawn(async { tokio::join!(async { rx1.await.unwrap() }, async { rx2.await.unwrap() }) }); assert_pending!(join.poll()); tx2.send(123).unwrap(); assert!(join.is_woken()); assert_pending!(join.poll()); tx1.send("hello").unwrap(); assert!(join.is_woken()); let res = assert_ready!(join.poll()); assert_eq!(("hello", 123), res); } #[test] #[cfg(target_pointer_width = "64")] fn join_size() { use futures::future; use std::mem; let fut = async { let ready = future::ready(0i32); tokio::join!(ready) }; assert_eq!(mem::size_of_val(&fut), 32); let fut = async { let ready1 = future::ready(0i32); let ready2 = future::ready(0i32); tokio::join!(ready1, ready2) }; assert_eq!(mem::size_of_val(&fut), 48); } async fn non_cooperative_task(permits: Arc<Semaphore>) -> usize { let mut exceeded_budget = 0; for _ in 0..5 { // Another task should run after this task uses its whole budget for _ in 0..128 { let _permit = permits.clone().acquire_owned().await.unwrap(); } exceeded_budget += 1; } exceeded_budget } async fn poor_little_task(permits: Arc<Semaphore>) -> usize { let mut how_many_times_i_got_to_run = 0; for _ in 0..5 { let _permit = permits.clone().acquire_owned().await.unwrap(); how_many_times_i_got_to_run += 1; } how_many_times_i_got_to_run } #[tokio::test] async fn join_does_not_allow_tasks_to_starve() { let permits = Arc::new(Semaphore::new(1)); // non_cooperative_task should yield after its budget is exceeded and then poor_little_task should run. let (non_cooperative_result, little_task_result) = tokio::join!( non_cooperative_task(Arc::clone(&permits)), poor_little_task(permits) ); assert_eq!(5, non_cooperative_result); assert_eq!(5, little_task_result); } #[tokio::test] async fn a_different_future_is_polled_first_every_time_poll_fn_is_polled() { let poll_order = Arc::new(std::sync::Mutex::new(vec![])); let fut = |x, poll_order: Arc<std::sync::Mutex<Vec<i32>>>| async move { for _ in 0..4 { { let mut guard = poll_order.lock().unwrap(); guard.push(x); } tokio::task::yield_now().await; } }; tokio::join!( fut(1, Arc::clone(&poll_order)), fut(2, Arc::clone(&poll_order)), fut(3, Arc::clone(&poll_order)), ); // Each time the future created by join! is polled, it should start // by polling a different future first. assert_eq!( vec![1, 2, 3, 2, 3, 1, 3, 1, 2, 1, 2, 3], *poll_order.lock().unwrap() ); } #[tokio::test] async fn futures_are_polled_in_order_in_biased_mode() { let poll_order = Arc::new(std::sync::Mutex::new(vec![])); let fut = |x, poll_order: Arc<std::sync::Mutex<Vec<i32>>>| async move { for _ in 0..4 { { let mut guard = poll_order.lock().unwrap(); guard.push(x); } tokio::task::yield_now().await; } }; tokio::join!( biased; fut(1, Arc::clone(&poll_order)), fut(2, Arc::clone(&poll_order)), fut(3, Arc::clone(&poll_order)), ); // Each time the future created by join! is polled, it should start // by polling in the order as declared in the macro inputs. assert_eq!( vec![1, 2, 3, 1, 2, 3, 1, 2, 3, 1, 2, 3], *poll_order.lock().unwrap() ); } #[test] #[cfg(target_pointer_width = "64")] fn join_size_biased() { use futures::future; use std::mem; let fut = async { let ready = future::ready(0i32); tokio::join!(biased; ready) }; assert_eq!(mem::size_of_val(&fut), 24); let fut = async { let ready1 = future::ready(0i32); let ready2 = future::ready(0i32); tokio::join!(biased; ready1, ready2) }; assert_eq!(mem::size_of_val(&fut), 40); } #[tokio::test] #[allow(clippy::unit_cmp)] async fn empty_join() { assert_eq!(tokio::join!(), ()); assert_eq!(tokio::join!(biased;), ()); } #[tokio::test] async fn join_into_future() { struct NotAFuture; impl std::future::IntoFuture for NotAFuture { type Output = (); type IntoFuture = std::future::Ready<()>; fn into_future(self) -> Self::IntoFuture { std::future::ready(()) } } tokio::join!(NotAFuture); } // Regression test for: https://github.com/tokio-rs/tokio/issues/7637 // We want to make sure that the `const COUNT: u32` declaration // inside the macro body doesn't leak to the caller to cause compiler failures // or variable shadowing. #[tokio::test] async fn caller_names_const_count() { let (tx, rx) = oneshot::channel::<u32>(); const COUNT: u32 = 2; let mut join = task::spawn(async { tokio::join!(async { tx.send(COUNT).unwrap() }) }); assert_ready!(join.poll()); let res = rx.await.unwrap(); // This passing demonstrates that the const in the macro is // not shadowing the caller-specified COUNT value assert_eq!(2, res); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/fs_symlink_dir_windows.rs
tokio/tests/fs_symlink_dir_windows.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // WASI does not support all fs operations #![cfg(windows)] use tempfile::tempdir; use tokio::fs; #[tokio::test] async fn symlink_file_windows() { const FILE_NAME: &str = "abc.txt"; let temp_dir = tempdir().unwrap(); let dir1 = temp_dir.path().join("a"); fs::create_dir(&dir1).await.unwrap(); let file1 = dir1.as_path().join(FILE_NAME); fs::write(&file1, b"Hello File!").await.unwrap(); let dir2 = temp_dir.path().join("b"); fs::symlink_dir(&dir1, &dir2).await.unwrap(); fs::write(&file1, b"new data!").await.unwrap(); let file2 = dir2.as_path().join(FILE_NAME); let from = fs::read(&file1).await.unwrap(); let to = fs::read(&file2).await.unwrap(); assert_eq!(from, to); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/uds_socket.rs
tokio/tests/uds_socket.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] #![cfg(unix)] use futures::future::try_join; use std::io; use tokio::{ io::{AsyncReadExt, AsyncWriteExt}, net::UnixSocket, }; #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` in miri. async fn datagram_echo_server() -> io::Result<()> { let dir = tempfile::tempdir().unwrap(); let server_path = dir.path().join("server.sock"); let client_path = dir.path().join("client.sock"); let server_socket = { let socket = UnixSocket::new_datagram()?; socket.bind(&server_path)?; socket.datagram()? }; tokio::spawn(async move { let mut recv_buf = vec![0u8; 1024]; loop { let (len, peer_addr) = server_socket.recv_from(&mut recv_buf[..]).await?; if let Some(path) = peer_addr.as_pathname() { server_socket.send_to(&recv_buf[..len], path).await?; } } #[allow(unreachable_code)] Ok::<(), io::Error>(()) }); { let socket = UnixSocket::new_datagram()?; socket.bind(&client_path).unwrap(); let socket = socket.datagram()?; socket.connect(server_path)?; socket.send(b"ECHO").await?; let mut recv_buf = [0u8; 16]; let len = socket.recv(&mut recv_buf[..]).await?; assert_eq!(&recv_buf[..len], b"ECHO"); } Ok(()) } #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` in miri. async fn listen_and_stream() -> std::io::Result<()> { let dir = tempfile::Builder::new().tempdir().unwrap(); let sock_path = dir.path().join("connect.sock"); let peer_path = dir.path().join("peer.sock"); let listener = { let sock = UnixSocket::new_stream()?; sock.bind(&sock_path)?; sock.listen(1024)? }; let accept = listener.accept(); let connect = { let sock = UnixSocket::new_stream()?; sock.bind(&peer_path)?; sock.connect(&sock_path) }; let ((mut server, _), mut client) = try_join(accept, connect).await?; assert_eq!( server.peer_addr().unwrap().as_pathname().unwrap(), &peer_path ); // Write to the client. client.write_all(b"hello").await?; drop(client); // Read from the server. let mut buf = vec![]; server.read_to_end(&mut buf).await?; assert_eq!(&buf, b"hello"); let len = server.read(&mut buf).await?; assert_eq!(len, 0); Ok(()) } #[tokio::test] #[cfg_attr(miri, ignore)] // No `socket` in miri. async fn assert_usage() -> std::io::Result<()> { let datagram_socket = UnixSocket::new_datagram()?; let result = datagram_socket .connect(std::path::PathBuf::new().join("invalid.sock")) .await; assert_eq!( result.unwrap_err().to_string(), "connect cannot be called on a datagram socket" ); let datagram_socket = UnixSocket::new_datagram()?; let result = datagram_socket.listen(1024); assert_eq!( result.unwrap_err().to_string(), "listen cannot be called on a datagram socket" ); let stream_socket = UnixSocket::new_stream()?; let result = stream_socket.datagram(); assert_eq!( result.unwrap_err().to_string(), "datagram cannot be called on a stream socket" ); Ok(()) }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_fill_buf.rs
tokio/tests/io_fill_buf.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // Wasi does not support file operations use tempfile::NamedTempFile; use tokio::fs::File; use tokio::io::{AsyncBufReadExt, BufReader}; use tokio_test::assert_ok; #[tokio::test] async fn fill_buf_file() { let file = NamedTempFile::new().unwrap(); assert_ok!(std::fs::write(file.path(), b"hello")); let file = assert_ok!(File::open(file.path()).await); let mut file = BufReader::new(file); let mut contents = Vec::new(); loop { let consumed = { let buffer = assert_ok!(file.fill_buf().await); if buffer.is_empty() { break; } contents.extend_from_slice(buffer); buffer.len() }; file.consume(consumed); } assert_eq!(contents, b"hello"); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/uds_stream.rs
tokio/tests/uds_stream.rs
#![cfg(feature = "full")] #![warn(rust_2018_idioms)] #![cfg(unix)] #![cfg(not(miri))] // No socket in miri. use std::io; #[cfg(target_os = "android")] use std::os::android::net::SocketAddrExt; #[cfg(target_os = "linux")] use std::os::linux::net::SocketAddrExt; use std::task::Poll; use tokio::io::{AsyncReadExt, AsyncWriteExt, Interest}; use tokio::net::{UnixListener, UnixStream}; use tokio_test::{assert_ok, assert_pending, assert_ready_ok, task}; use futures::future::{poll_fn, try_join}; #[tokio::test] async fn accept_read_write() -> std::io::Result<()> { let dir = tempfile::Builder::new() .prefix("tokio-uds-tests") .tempdir() .unwrap(); let sock_path = dir.path().join("connect.sock"); let listener = UnixListener::bind(&sock_path)?; let accept = listener.accept(); let connect = UnixStream::connect(&sock_path); let ((mut server, _), mut client) = try_join(accept, connect).await?; // Write to the client. client.write_all(b"hello").await?; drop(client); // Read from the server. let mut buf = vec![]; server.read_to_end(&mut buf).await?; assert_eq!(&buf, b"hello"); let len = server.read(&mut buf).await?; assert_eq!(len, 0); Ok(()) } #[tokio::test] async fn shutdown() -> std::io::Result<()> { let dir = tempfile::Builder::new() .prefix("tokio-uds-tests") .tempdir() .unwrap(); let sock_path = dir.path().join("connect.sock"); let listener = UnixListener::bind(&sock_path)?; let accept = listener.accept(); let connect = UnixStream::connect(&sock_path); let ((mut server, _), mut client) = try_join(accept, connect).await?; // Shut down the client AsyncWriteExt::shutdown(&mut client).await?; // Read from the server should return 0 to indicate the channel has been closed. let mut buf = [0u8; 1]; let n = server.read(&mut buf).await?; assert_eq!(n, 0); Ok(()) } #[tokio::test] async fn try_read_write() -> std::io::Result<()> { let msg = b"hello world"; let dir = tempfile::tempdir()?; let bind_path = dir.path().join("bind.sock"); // Create listener let listener = UnixListener::bind(&bind_path)?; // Create socket pair let client = UnixStream::connect(&bind_path).await?; let (server, _) = listener.accept().await?; let mut written = msg.to_vec(); // Track the server receiving data let mut readable = task::spawn(server.readable()); assert_pending!(readable.poll()); // Write data. client.writable().await?; assert_eq!(msg.len(), client.try_write(msg)?); // The task should be notified while !readable.is_woken() { tokio::task::yield_now().await; } // Fill the write buffer using non-vectored I/O loop { // Still ready let mut writable = task::spawn(client.writable()); assert_ready_ok!(writable.poll()); match client.try_write(msg) { Ok(n) => written.extend(&msg[..n]), Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => { break; } Err(e) => panic!("error = {e:?}"), } } { // Write buffer full let mut writable = task::spawn(client.writable()); assert_pending!(writable.poll()); // Drain the socket from the server end using non-vectored I/O let mut read = vec![0; written.len()]; let mut i = 0; while i < read.len() { server.readable().await?; match server.try_read(&mut read[i..]) { Ok(n) => i += n, Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("error = {e:?}"), } } assert_eq!(read, written); } written.clear(); client.writable().await.unwrap(); // Fill the write buffer using vectored I/O let msg_bufs: Vec<_> = msg.chunks(3).map(io::IoSlice::new).collect(); loop { // Still ready let mut writable = task::spawn(client.writable()); assert_ready_ok!(writable.poll()); match client.try_write_vectored(&msg_bufs) { Ok(n) => written.extend(&msg[..n]), Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => { break; } Err(e) => panic!("error = {e:?}"), } } { // Write buffer full let mut writable = task::spawn(client.writable()); assert_pending!(writable.poll()); // Drain the socket from the server end using vectored I/O let mut read = vec![0; written.len()]; let mut i = 0; while i < read.len() { server.readable().await?; let mut bufs: Vec<_> = read[i..] .chunks_mut(0x10000) .map(io::IoSliceMut::new) .collect(); match server.try_read_vectored(&mut bufs) { Ok(n) => i += n, Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("error = {e:?}"), } } assert_eq!(read, written); } // Now, we listen for shutdown drop(client); loop { let ready = server.ready(Interest::READABLE).await?; if ready.is_read_closed() { break; } else { tokio::task::yield_now().await; } } Ok(()) } async fn create_pair() -> (UnixStream, UnixStream) { let dir = assert_ok!(tempfile::tempdir()); let bind_path = dir.path().join("bind.sock"); let listener = assert_ok!(UnixListener::bind(&bind_path)); let accept = listener.accept(); let connect = UnixStream::connect(&bind_path); let ((server, _), client) = assert_ok!(try_join(accept, connect).await); (client, server) } macro_rules! assert_readable_by_polling { ($stream:expr) => { assert_ok!(poll_fn(|cx| $stream.poll_read_ready(cx)).await); }; } macro_rules! assert_not_readable_by_polling { ($stream:expr) => { poll_fn(|cx| { assert_pending!($stream.poll_read_ready(cx)); Poll::Ready(()) }) .await; }; } macro_rules! assert_writable_by_polling { ($stream:expr) => { assert_ok!(poll_fn(|cx| $stream.poll_write_ready(cx)).await); }; } macro_rules! assert_not_writable_by_polling { ($stream:expr) => { poll_fn(|cx| { assert_pending!($stream.poll_write_ready(cx)); Poll::Ready(()) }) .await; }; } #[tokio::test] async fn poll_read_ready() { let (mut client, mut server) = create_pair().await; // Initial state - not readable. assert_not_readable_by_polling!(server); // There is data in the buffer - readable. assert_ok!(client.write_all(b"ping").await); assert_readable_by_polling!(server); // Readable until calls to `poll_read` return `Poll::Pending`. let mut buf = [0u8; 4]; assert_ok!(server.read_exact(&mut buf).await); assert_readable_by_polling!(server); read_until_pending(&mut server); assert_not_readable_by_polling!(server); // Detect the client disconnect. drop(client); assert_readable_by_polling!(server); } #[tokio::test] async fn poll_write_ready() { let (mut client, server) = create_pair().await; // Initial state - writable. assert_writable_by_polling!(client); // No space to write - not writable. write_until_pending(&mut client); assert_not_writable_by_polling!(client); // Detect the server disconnect. drop(server); assert_writable_by_polling!(client); } fn read_until_pending(stream: &mut UnixStream) { let mut buf = vec![0u8; 1024 * 1024]; loop { match stream.try_read(&mut buf) { Ok(_) => (), Err(err) => { assert_eq!(err.kind(), io::ErrorKind::WouldBlock); break; } } } } fn write_until_pending(stream: &mut UnixStream) { let buf = vec![0u8; 1024 * 1024]; loop { match stream.try_write(&buf) { Ok(_) => (), Err(err) => { assert_eq!(err.kind(), io::ErrorKind::WouldBlock); break; } } } } #[tokio::test] async fn try_read_buf() -> std::io::Result<()> { let msg = b"hello world"; let dir = tempfile::tempdir()?; let bind_path = dir.path().join("bind.sock"); // Create listener let listener = UnixListener::bind(&bind_path)?; // Create socket pair let client = UnixStream::connect(&bind_path).await?; let (server, _) = listener.accept().await?; let mut written = msg.to_vec(); // Track the server receiving data let mut readable = task::spawn(server.readable()); assert_pending!(readable.poll()); // Write data. client.writable().await?; assert_eq!(msg.len(), client.try_write(msg)?); // The task should be notified while !readable.is_woken() { tokio::task::yield_now().await; } // Fill the write buffer loop { // Still ready let mut writable = task::spawn(client.writable()); assert_ready_ok!(writable.poll()); match client.try_write(msg) { Ok(n) => written.extend(&msg[..n]), Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => { break; } Err(e) => panic!("error = {e:?}"), } } { // Write buffer full let mut writable = task::spawn(client.writable()); assert_pending!(writable.poll()); // Drain the socket from the server end let mut read = Vec::with_capacity(written.len()); let mut i = 0; while i < read.capacity() { server.readable().await?; match server.try_read_buf(&mut read) { Ok(n) => i += n, Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => continue, Err(e) => panic!("error = {e:?}"), } } assert_eq!(read, written); } // Now, we listen for shutdown drop(client); loop { let ready = server.ready(Interest::READABLE).await?; if ready.is_read_closed() { break; } else { tokio::task::yield_now().await; } } Ok(()) } // https://github.com/tokio-rs/tokio/issues/3879 #[tokio::test] #[cfg(not(target_os = "macos"))] async fn epollhup() -> io::Result<()> { let dir = tempfile::Builder::new() .prefix("tokio-uds-tests") .tempdir() .unwrap(); let sock_path = dir.path().join("connect.sock"); let listener = UnixListener::bind(&sock_path)?; let connect = UnixStream::connect(&sock_path); tokio::pin!(connect); // Poll `connect` once. poll_fn(|cx| { use std::future::Future; assert_pending!(connect.as_mut().poll(cx)); Poll::Ready(()) }) .await; drop(listener); let err = connect.await.unwrap_err(); let errno = err.kind(); assert!( // As far as I can tell, whether we see ECONNREFUSED or ECONNRESET here // seems relatively inconsistent, at least on non-Linux operating // systems. The difference in meaning between these errnos is not // particularly well-defined, so let's just accept either. matches!( errno, io::ErrorKind::ConnectionRefused | io::ErrorKind::ConnectionReset ), "unexpected error kind: {errno:?} (expected ConnectionRefused or ConnectionReset)" ); Ok(()) } // test for https://github.com/tokio-rs/tokio/issues/6767 #[tokio::test] #[cfg(any(target_os = "linux", target_os = "android"))] async fn abstract_socket_name() { let socket_path = "\0aaa"; let listener = UnixListener::bind(socket_path).unwrap(); let accept = listener.accept(); let connect = UnixStream::connect(&socket_path); let ((stream, _), _) = try_join(accept, connect).await.unwrap(); let local_addr = stream.into_std().unwrap().local_addr().unwrap(); let abstract_path_name = local_addr.as_abstract_name().unwrap(); // `as_abstract_name` removes leading zero bytes assert_eq!(abstract_path_name, b"aaa"); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/duplex_stream.rs
tokio/tests/duplex_stream.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use std::io::IoSlice; use tokio::io::{AsyncReadExt, AsyncWriteExt}; const HELLO: &[u8] = b"hello world..."; #[tokio::test] async fn write_vectored() { let (mut client, mut server) = tokio::io::duplex(64); let ret = client .write_vectored(&[IoSlice::new(HELLO), IoSlice::new(HELLO)]) .await .unwrap(); assert_eq!(ret, HELLO.len() * 2); client.flush().await.unwrap(); drop(client); let mut buf = Vec::with_capacity(HELLO.len() * 2); let bytes_read = server.read_to_end(&mut buf).await.unwrap(); assert_eq!(bytes_read, HELLO.len() * 2); assert_eq!(buf, [HELLO, HELLO].concat()); } #[tokio::test] async fn write_vectored_and_shutdown() { let (mut client, mut server) = tokio::io::duplex(64); let ret = client .write_vectored(&[IoSlice::new(HELLO), IoSlice::new(HELLO)]) .await .unwrap(); assert_eq!(ret, HELLO.len() * 2); client.shutdown().await.unwrap(); drop(client); let mut buf = Vec::with_capacity(HELLO.len() * 2); let bytes_read = server.read_to_end(&mut buf).await.unwrap(); assert_eq!(bytes_read, HELLO.len() * 2); assert_eq!(buf, [HELLO, HELLO].concat()); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/signal_drop_recv.rs
tokio/tests/signal_drop_recv.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] #![cfg(unix)] #![cfg(not(miri))] // No `sigaction` in Miri. mod support { pub mod signal; } use support::signal::send_signal; use tokio::signal::unix::{signal, SignalKind}; #[tokio::test] async fn drop_then_get_a_signal() { let kind = SignalKind::user_defined1(); let sig = signal(kind).expect("failed to create first signal"); drop(sig); send_signal(libc::SIGUSR1); let mut sig = signal(kind).expect("failed to create second signal"); let _ = sig.recv().await; }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/process_issue_2174.rs
tokio/tests/process_issue_2174.rs
#![cfg(feature = "process")] #![warn(rust_2018_idioms)] // This test reveals a difference in behavior of kqueue on FreeBSD. When the // reader disconnects, there does not seem to be an `EVFILT_WRITE` filter that // is returned. // // It is expected that `EVFILT_WRITE` would be returned with either the // `EV_EOF` or `EV_ERROR` flag set. If either flag is set a write would be // attempted, but that does not seem to occur. #![cfg(all(unix, not(target_os = "freebsd"), not(miri)))] use std::process::Stdio; use std::time::Duration; use tokio::io::AsyncWriteExt; use tokio::process::Command; use tokio::time; use tokio_test::assert_err; #[tokio::test] #[cfg_attr(panic = "abort", ignore)] async fn issue_2174() { let mut child = Command::new("sleep") .arg("2") .stdin(Stdio::piped()) .stdout(Stdio::null()) .spawn() .unwrap(); let mut input = child.stdin.take().unwrap(); // Writes will buffer up to 65_636. This *should* loop at least 8 times // and then register interest. let handle = tokio::spawn(async move { let data = [0u8; 8192]; loop { input.write_all(&data).await.unwrap(); } }); // Sleep enough time so that the child process's stdin's buffer fills. time::sleep(Duration::from_secs(1)).await; // Kill the child process. child.kill().await.unwrap(); assert_err!(handle.await); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_chain.rs
tokio/tests/io_chain.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use tokio::io::AsyncReadExt; use tokio_test::assert_ok; #[tokio::test] async fn chain() { let mut buf = Vec::new(); let rd1: &[u8] = b"hello "; let rd2: &[u8] = b"world"; let mut rd = rd1.chain(rd2); assert_ok!(rd.read_to_end(&mut buf).await); assert_eq!(buf, b"hello world"); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/tcp_split.rs
tokio/tests/tcp_split.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi"), not(miri)))] // Wasi doesn't support bind // No `socket` on miri. use std::io::Result; use std::io::{Read, Write}; use std::{net, thread}; use tokio::io::{AsyncReadExt, AsyncWriteExt}; use tokio::net::TcpStream; #[tokio::test] async fn split() -> Result<()> { const MSG: &[u8] = b"split"; let listener = net::TcpListener::bind("127.0.0.1:0")?; let addr = listener.local_addr()?; let handle = thread::spawn(move || { let (mut stream, _) = listener.accept().unwrap(); stream.write_all(MSG).unwrap(); let mut read_buf = [0u8; 32]; let read_len = stream.read(&mut read_buf).unwrap(); assert_eq!(&read_buf[..read_len], MSG); }); let mut stream = TcpStream::connect(&addr).await?; let (mut read_half, mut write_half) = stream.split(); let mut read_buf = [0u8; 32]; let peek_len1 = read_half.peek(&mut read_buf[..]).await?; let peek_len2 = read_half.peek(&mut read_buf[..]).await?; assert_eq!(peek_len1, peek_len2); let read_len = read_half.read(&mut read_buf[..]).await?; assert_eq!(peek_len1, read_len); assert_eq!(&read_buf[..read_len], MSG); assert_eq!(write_half.write(MSG).await?, MSG.len()); handle.join().unwrap(); Ok(()) }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/signal_notify_both.rs
tokio/tests/signal_notify_both.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] #![cfg(unix)] #![cfg(not(miri))] // No `sigaction` on Miri. mod support { pub mod signal; } use support::signal::send_signal; use tokio::signal::unix::{signal, SignalKind}; #[tokio::test] async fn notify_both() { let kind = SignalKind::user_defined2(); let mut signal1 = signal(kind).expect("failed to create signal1"); let mut signal2 = signal(kind).expect("failed to create signal2"); send_signal(libc::SIGUSR2); signal1.recv().await; signal2.recv().await; }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/sync_set_once.rs
tokio/tests/sync_set_once.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use std::sync::{ atomic::{AtomicU32, Ordering}, Arc, }; use tokio::sync::SetOnce; #[derive(Clone)] struct DropCounter { drops: Arc<AtomicU32>, } impl DropCounter { fn new() -> Self { DropCounter { drops: Arc::new(AtomicU32::new(0)), } } fn assert_num_drops(&self, value: u32) { assert_eq!(value, self.drops.load(Ordering::Relaxed)); } } impl Drop for DropCounter { fn drop(&mut self) { self.drops.fetch_add(1, Ordering::Relaxed); } } #[test] fn drop_cell() { let fooer = DropCounter::new(); let fooer_cl = fooer.clone(); { let once_cell = SetOnce::new(); let prev = once_cell.set(fooer_cl); assert!(prev.is_ok()) } fooer.assert_num_drops(1); } #[test] fn drop_cell_new_with() { let fooer = DropCounter::new(); { let once_cell = SetOnce::new_with(Some(fooer.clone())); assert!(once_cell.initialized()); } fooer.assert_num_drops(1); } #[test] fn drop_into_inner() { let fooer = DropCounter::new(); let once_cell = SetOnce::new(); assert!(once_cell.set(fooer.clone()).is_ok()); let val = once_cell.into_inner(); fooer.assert_num_drops(0); drop(val); fooer.assert_num_drops(1); } #[test] fn drop_into_inner_new_with() { let fooer = DropCounter::new(); let once_cell = SetOnce::new_with(Some(fooer.clone())); let val = once_cell.into_inner(); fooer.assert_num_drops(0); drop(val); fooer.assert_num_drops(1); } #[test] fn from() { let cell = SetOnce::from(2); assert_eq!(*cell.get().unwrap(), 2); } #[test] fn set_and_get() { static ONCE: SetOnce<u32> = SetOnce::const_new(); ONCE.set(5).unwrap(); let value = ONCE.get().unwrap(); assert_eq!(*value, 5); } #[tokio::test] async fn set_and_wait() { static ONCE: SetOnce<u32> = SetOnce::const_new(); tokio::spawn(async { ONCE.set(5) }); let value = ONCE.wait().await; assert_eq!(*value, 5); } #[test] #[cfg_attr(target_family = "wasm", ignore)] fn set_and_wait_multiple_threads() { static ONCE: SetOnce<u32> = SetOnce::const_new(); let res1 = std::thread::spawn(|| ONCE.set(4)); let res2 = std::thread::spawn(|| ONCE.set(3)); let result_first = res1.join().unwrap().is_err(); let result_two = res2.join().unwrap().is_err(); assert!(result_first != result_two); } #[tokio::test] #[cfg_attr(target_family = "wasm", ignore)] async fn set_and_wait_threads() { static ONCE: SetOnce<u32> = SetOnce::const_new(); let thread = std::thread::spawn(|| { ONCE.set(4).unwrap(); }); let value = ONCE.wait().await; thread.join().unwrap(); assert_eq!(*value, 4); } #[test] fn get_uninit() { static ONCE: SetOnce<u32> = SetOnce::const_new(); let uninit = ONCE.get(); assert!(uninit.is_none()); } #[test] fn set_twice() { static ONCE: SetOnce<u32> = SetOnce::const_new(); let first = ONCE.set(5); assert_eq!(first, Ok(())); let second = ONCE.set(6); assert!(second.is_err()); } #[test] fn is_none_initializing() { static ONCE: SetOnce<u32> = SetOnce::const_new(); assert_eq!(ONCE.get(), None); ONCE.set(20).unwrap(); assert!(ONCE.set(10).is_err()); } #[tokio::test] async fn is_some_initializing() { static ONCE: SetOnce<u32> = SetOnce::const_new(); tokio::spawn(async { ONCE.set(20) }); assert_eq!(*ONCE.wait().await, 20); } #[test] fn into_inner_int_empty_setonce() { let once = SetOnce::<u32>::new(); let val = once.into_inner(); assert!(val.is_none()); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/uds_cred.rs
tokio/tests/uds_cred.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] #![cfg(all(unix, not(target_os = "dragonfly"), not(miri)))] // No `getsockopt` on miri. use tokio::net::UnixStream; use libc::getegid; use libc::geteuid; #[tokio::test] #[cfg_attr( target_os = "netbsd", ignore = "NetBSD does not support getpeereid() for sockets created by socketpair()" )] async fn test_socket_pair() { let (a, b) = UnixStream::pair().unwrap(); let cred_a = a.peer_cred().unwrap(); let cred_b = b.peer_cred().unwrap(); assert_eq!(cred_a, cred_b); let uid = unsafe { geteuid() }; let gid = unsafe { getegid() }; assert_eq!(cred_a.uid(), uid); assert_eq!(cred_a.gid(), gid); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/tcp_echo.rs
tokio/tests/tcp_echo.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi"), not(miri)))] // Wasi doesn't support bind // No socket on miri. use tokio::io::{self, AsyncReadExt, AsyncWriteExt}; use tokio::net::{TcpListener, TcpStream}; use tokio::sync::oneshot; use tokio_test::assert_ok; #[tokio::test] async fn echo_server() { const ITER: usize = 1024; let (tx, rx) = oneshot::channel(); let srv = assert_ok!(TcpListener::bind("127.0.0.1:0").await); let addr = assert_ok!(srv.local_addr()); let msg = "foo bar baz"; tokio::spawn(async move { let mut stream = assert_ok!(TcpStream::connect(&addr).await); for _ in 0..ITER { // write assert_ok!(stream.write_all(msg.as_bytes()).await); // read let mut buf = [0; 11]; assert_ok!(stream.read_exact(&mut buf).await); assert_eq!(&buf[..], msg.as_bytes()); } assert_ok!(tx.send(())); }); let (mut stream, _) = assert_ok!(srv.accept().await); let (mut rd, mut wr) = stream.split(); let n = assert_ok!(io::copy(&mut rd, &mut wr).await); assert_eq!(n, (ITER * msg.len()) as u64); assert_ok!(rx.await); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/signal_panic.rs
tokio/tests/signal_panic.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] #![cfg(unix)] #![cfg(panic = "unwind")] #![cfg(not(miri))] // No `sigaction` on Miri. use std::error::Error; use tokio::runtime::Builder; use tokio::signal::unix::{signal, SignalKind}; mod support { pub mod panic; } use support::panic::test_panic; #[test] fn signal_panic_caller() -> Result<(), Box<dyn Error>> { let panic_location_file = test_panic(|| { let rt = Builder::new_current_thread().build().unwrap(); rt.block_on(async { let kind = SignalKind::from_raw(-1); let _ = signal(kind); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/uds_split.rs
tokio/tests/uds_split.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] #![cfg(unix)] #![cfg(not(miri))] // No `socket` in miri. use tokio::io::{AsyncRead, AsyncReadExt, AsyncWrite, AsyncWriteExt}; use tokio::net::UnixStream; /// Checks that `UnixStream` can be split into a read half and a write half using /// `UnixStream::split` and `UnixStream::split_mut`. /// /// Verifies that the implementation of `AsyncWrite::poll_shutdown` shutdowns the stream for /// writing by reading to the end of stream on the other side of the connection. #[tokio::test] async fn split() -> std::io::Result<()> { let (mut a, mut b) = UnixStream::pair()?; let (mut a_read, mut a_write) = a.split(); let (mut b_read, mut b_write) = b.split(); let (a_response, b_response) = futures::future::try_join( send_recv_all(&mut a_read, &mut a_write, b"A"), send_recv_all(&mut b_read, &mut b_write, b"B"), ) .await?; assert_eq!(a_response, b"B"); assert_eq!(b_response, b"A"); Ok(()) } async fn send_recv_all( read: &mut (dyn AsyncRead + Unpin), write: &mut (dyn AsyncWrite + Unpin), input: &[u8], ) -> std::io::Result<Vec<u8>> { write.write_all(input).await?; write.shutdown().await?; let mut output = Vec::new(); read.read_to_end(&mut output).await?; Ok(output) }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/time_wasm.rs
tokio/tests/time_wasm.rs
#![warn(rust_2018_idioms)] #![cfg(all(target_arch = "wasm32", not(target_os = "wasi")))] use wasm_bindgen_test::wasm_bindgen_test; #[wasm_bindgen_test] #[should_panic] fn instant_now_panics() { let _ = tokio::time::Instant::now(); } #[cfg(all(feature = "rt", not(feature = "time")))] #[wasm_bindgen_test] fn runtime_without_time_does_not_panic() { let rt = tokio::runtime::Builder::new_current_thread() .build() .unwrap(); rt.block_on(async {}); } #[cfg(all(feature = "rt", feature = "time"))] #[wasm_bindgen_test] #[should_panic] // should remove this once time is supported fn runtime_with_time_does_not_panic() { let rt = tokio::runtime::Builder::new_current_thread() .build() .unwrap(); rt.block_on(async {}); } #[cfg(all(feature = "rt", feature = "time"))] #[wasm_bindgen_test] #[should_panic] fn sleep_panics_on_unknown_unknown() { let rt = tokio::runtime::Builder::new_current_thread() .enable_time() .build() .unwrap(); rt.block_on(async { tokio::time::sleep(core::time::Duration::from_millis(1)).await }); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/fs_canonicalize_dir.rs
tokio/tests/fs_canonicalize_dir.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi")))] // WASI does not support all fs operations use tokio::fs; #[tokio::test] #[cfg(unix)] async fn canonicalize_root_dir_unix() { assert_eq!(fs::canonicalize("/.").await.unwrap().to_str().unwrap(), "/"); } #[tokio::test] #[cfg(windows)] async fn canonicalize_root_dir_windows() { // 2-step let bindings due to Rust memory semantics let dir_path = fs::canonicalize("C:\\.\\").await.unwrap(); let dir_name = dir_path.to_str().unwrap(); assert!(dir_name.starts_with("\\\\")); assert!(dir_name.ends_with("C:\\")); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_read_to_string.rs
tokio/tests/io_read_to_string.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use std::io; use tokio::io::AsyncReadExt; use tokio_test::assert_ok; use tokio_test::io::Builder; #[tokio::test] async fn read_to_string() { let mut buf = String::new(); let mut rd: &[u8] = b"hello world"; let n = assert_ok!(rd.read_to_string(&mut buf).await); assert_eq!(n, 11); assert_eq!(buf[..], "hello world"[..]); } #[tokio::test] async fn to_string_does_not_truncate_on_utf8_error() { let data = vec![0xff, 0xff, 0xff]; let mut s = "abc".to_string(); match AsyncReadExt::read_to_string(&mut data.as_slice(), &mut s).await { Ok(len) => panic!("Should fail: {len} bytes."), Err(err) if err.to_string() == "stream did not contain valid UTF-8" => {} Err(err) => panic!("Fail: {err}."), } assert_eq!(s, "abc"); } #[tokio::test] async fn to_string_does_not_truncate_on_io_error() { let mut mock = Builder::new() .read(b"def") .read_error(io::Error::new(io::ErrorKind::Other, "whoops")) .build(); let mut s = "abc".to_string(); match AsyncReadExt::read_to_string(&mut mock, &mut s).await { Ok(len) => panic!("Should fail: {len} bytes."), Err(err) if err.to_string() == "whoops" => {} Err(err) => panic!("Fail: {err}."), } assert_eq!(s, "abc"); } #[tokio::test] async fn to_string_appends() { let data = b"def".to_vec(); let mut s = "abc".to_string(); let len = AsyncReadExt::read_to_string(&mut data.as_slice(), &mut s) .await .unwrap(); assert_eq!(len, 3); assert_eq!(s, "abcdef"); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/task_id.rs
tokio/tests/task_id.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use std::error::Error; use std::future::Future; use std::pin::Pin; use std::task::{Context, Poll}; use tokio::runtime::Runtime; use tokio::sync::oneshot; use tokio::task::{self, Id, LocalSet}; mod support { pub mod panic; } use support::panic::test_panic; #[tokio::test(flavor = "current_thread")] async fn task_id_spawn() { tokio::spawn(async { println!("task id: {}", task::id()) }) .await .unwrap(); } #[cfg(not(target_os = "wasi"))] #[tokio::test(flavor = "current_thread")] async fn task_id_spawn_blocking() { task::spawn_blocking(|| println!("task id: {}", task::id())) .await .unwrap(); } #[tokio::test(flavor = "current_thread")] async fn task_id_collision_current_thread() { let handle1 = tokio::spawn(async { task::id() }); let handle2 = tokio::spawn(async { task::id() }); let (id1, id2) = tokio::join!(handle1, handle2); assert_ne!(id1.unwrap(), id2.unwrap()); } #[cfg(not(target_os = "wasi"))] #[tokio::test(flavor = "multi_thread")] async fn task_id_collision_multi_thread() { let handle1 = tokio::spawn(async { task::id() }); let handle2 = tokio::spawn(async { task::id() }); let (id1, id2) = tokio::join!(handle1, handle2); assert_ne!(id1.unwrap(), id2.unwrap()); } #[tokio::test(flavor = "current_thread")] async fn task_ids_match_current_thread() { let (tx, rx) = oneshot::channel(); let handle = tokio::spawn(async { let id = rx.await.unwrap(); assert_eq!(id, task::id()); }); tx.send(handle.id()).unwrap(); handle.await.unwrap(); } #[cfg(not(target_os = "wasi"))] #[tokio::test(flavor = "multi_thread")] async fn task_ids_match_multi_thread() { let (tx, rx) = oneshot::channel(); let handle = tokio::spawn(async { let id = rx.await.unwrap(); assert_eq!(id, task::id()); }); tx.send(handle.id()).unwrap(); handle.await.unwrap(); } #[cfg(not(target_os = "wasi"))] #[tokio::test(flavor = "multi_thread")] async fn task_id_future_destructor_completion() { struct MyFuture { tx: Option<oneshot::Sender<Id>>, } impl Future for MyFuture { type Output = (); fn poll(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<()> { Poll::Ready(()) } } impl Drop for MyFuture { fn drop(&mut self) { let _ = self.tx.take().unwrap().send(task::id()); } } let (tx, rx) = oneshot::channel(); let handle = tokio::spawn(MyFuture { tx: Some(tx) }); let id = handle.id(); handle.await.unwrap(); assert_eq!(rx.await.unwrap(), id); } #[cfg(not(target_os = "wasi"))] #[tokio::test(flavor = "multi_thread")] async fn task_id_future_destructor_abort() { struct MyFuture { tx: Option<oneshot::Sender<Id>>, } impl Future for MyFuture { type Output = (); fn poll(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<()> { Poll::Pending } } impl Drop for MyFuture { fn drop(&mut self) { let _ = self.tx.take().unwrap().send(task::id()); } } let (tx, rx) = oneshot::channel(); let handle = tokio::spawn(MyFuture { tx: Some(tx) }); let id = handle.id(); handle.abort(); assert!(handle.await.unwrap_err().is_cancelled()); assert_eq!(rx.await.unwrap(), id); } #[tokio::test(flavor = "current_thread")] async fn task_id_output_destructor_handle_dropped_before_completion() { struct MyOutput { tx: Option<oneshot::Sender<Id>>, } impl Drop for MyOutput { fn drop(&mut self) { let _ = self.tx.take().unwrap().send(task::id()); } } struct MyFuture { tx: Option<oneshot::Sender<Id>>, } impl Future for MyFuture { type Output = MyOutput; fn poll(mut self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<Self::Output> { Poll::Ready(MyOutput { tx: self.tx.take() }) } } let (tx, mut rx) = oneshot::channel(); let handle = tokio::spawn(MyFuture { tx: Some(tx) }); let id = handle.id(); drop(handle); assert!(rx.try_recv().is_err()); assert_eq!(rx.await.unwrap(), id); } #[tokio::test(flavor = "current_thread")] async fn task_id_output_destructor_handle_dropped_after_completion() { struct MyOutput { tx: Option<oneshot::Sender<Id>>, } impl Drop for MyOutput { fn drop(&mut self) { let _ = self.tx.take().unwrap().send(task::id()); } } struct MyFuture { tx_output: Option<oneshot::Sender<Id>>, tx_future: Option<oneshot::Sender<()>>, } impl Future for MyFuture { type Output = MyOutput; fn poll(mut self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<Self::Output> { let _ = self.tx_future.take().unwrap().send(()); Poll::Ready(MyOutput { tx: self.tx_output.take(), }) } } let (tx_output, mut rx_output) = oneshot::channel(); let (tx_future, rx_future) = oneshot::channel(); let handle = tokio::spawn(MyFuture { tx_output: Some(tx_output), tx_future: Some(tx_future), }); let id = handle.id(); rx_future.await.unwrap(); assert!(rx_output.try_recv().is_err()); drop(handle); assert_eq!(rx_output.await.unwrap(), id); } #[test] fn task_try_id_outside_task() { assert_eq!(None, task::try_id()); } #[cfg(not(target_os = "wasi"))] #[test] fn task_try_id_inside_block_on() { let rt = Runtime::new().unwrap(); rt.block_on(async { assert_eq!(None, task::try_id()); }); } #[tokio::test(flavor = "current_thread")] async fn task_id_spawn_local() { LocalSet::new() .run_until(async { task::spawn_local(async { println!("task id: {}", task::id()) }) .await .unwrap(); }) .await } #[tokio::test(flavor = "current_thread")] async fn task_id_nested_spawn_local() { LocalSet::new() .run_until(async { task::spawn_local(async { let parent_id = task::id(); LocalSet::new() .run_until(async { task::spawn_local(async move { assert_ne!(parent_id, task::id()); }) .await .unwrap(); }) .await; assert_eq!(parent_id, task::id()); }) .await .unwrap(); }) .await; } #[cfg(not(target_os = "wasi"))] #[tokio::test(flavor = "multi_thread")] async fn task_id_block_in_place_block_on_spawn() { use tokio::runtime::Builder; task::spawn(async { let parent_id = task::id(); task::block_in_place(move || { let rt = Builder::new_current_thread().build().unwrap(); rt.block_on(rt.spawn(async move { assert_ne!(parent_id, task::id()); })) .unwrap(); }); assert_eq!(parent_id, task::id()); }) .await .unwrap(); } #[test] #[cfg_attr(not(panic = "unwind"), ignore)] fn task_id_outside_task_panic_caller() -> Result<(), Box<dyn Error>> { let panic_location_file = test_panic(|| { let _ = task::id(); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) } #[test] #[cfg_attr(not(panic = "unwind"), ignore)] fn task_id_inside_block_on_panic_caller() -> Result<(), Box<dyn Error>> { let panic_location_file = test_panic(|| { let rt = Runtime::new().unwrap(); rt.block_on(async { task::id(); }); }); // The panic location should be in this file assert_eq!(&panic_location_file.unwrap(), file!()); Ok(()) }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/tcp_into_std.rs
tokio/tests/tcp_into_std.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi"), not(miri)))] // Wasi doesn't support bind // No socket on `miri`. use std::io::Read; use std::io::Result; use tokio::io::{AsyncReadExt, AsyncWriteExt}; use tokio::net::TcpListener; use tokio::net::TcpStream; #[tokio::test] async fn tcp_into_std() -> Result<()> { let mut data = [0u8; 12]; let listener = TcpListener::bind("127.0.0.1:0").await?; let addr = listener.local_addr().unwrap().to_string(); let handle = tokio::spawn(async { let stream: TcpStream = TcpStream::connect(addr).await.unwrap(); stream }); let (tokio_tcp_stream, _) = listener.accept().await?; let mut std_tcp_stream = tokio_tcp_stream.into_std()?; std_tcp_stream .set_nonblocking(false) .expect("set_nonblocking call failed"); let mut client = handle.await.expect("The task being joined has panicked"); client.write_all(b"Hello world!").await?; std_tcp_stream .read_exact(&mut data) .expect("std TcpStream read failed!"); assert_eq!(b"Hello world!", &data); // test back to tokio stream std_tcp_stream .set_nonblocking(true) .expect("set_nonblocking call failed"); let mut tokio_tcp_stream = TcpStream::from_std(std_tcp_stream)?; client.write_all(b"Hello tokio!").await?; let _size = tokio_tcp_stream.read_exact(&mut data).await?; assert_eq!(b"Hello tokio!", &data); Ok(()) }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/task_blocking.rs
tokio/tests/task_blocking.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi"), not(miri)))] // Wasi doesn't support threads use tokio::{runtime, task, time}; use tokio_test::assert_ok; use std::thread; use std::time::Duration; mod support { pub(crate) mod mpsc_stream; } #[tokio::test] async fn basic_blocking() { // Run a few times for _ in 0..100 { let out = assert_ok!( tokio::spawn(async { assert_ok!( task::spawn_blocking(|| { thread::sleep(Duration::from_millis(5)); "hello" }) .await ) }) .await ); assert_eq!(out, "hello"); } } #[tokio::test(flavor = "multi_thread")] async fn block_in_blocking() { // Run a few times for _ in 0..100 { let out = assert_ok!( tokio::spawn(async { assert_ok!( task::spawn_blocking(|| { task::block_in_place(|| { thread::sleep(Duration::from_millis(5)); }); "hello" }) .await ) }) .await ); assert_eq!(out, "hello"); } } #[tokio::test(flavor = "multi_thread")] async fn block_in_block() { // Run a few times for _ in 0..100 { let out = assert_ok!( tokio::spawn(async { task::block_in_place(|| { task::block_in_place(|| { thread::sleep(Duration::from_millis(5)); }); "hello" }) }) .await ); assert_eq!(out, "hello"); } } #[tokio::test(flavor = "current_thread")] #[should_panic] async fn no_block_in_current_thread_scheduler() { task::block_in_place(|| {}); } #[test] fn yes_block_in_threaded_block_on() { let rt = runtime::Runtime::new().unwrap(); rt.block_on(async { task::block_in_place(|| {}); }); } #[test] #[should_panic] fn no_block_in_current_thread_block_on() { let rt = runtime::Builder::new_current_thread().build().unwrap(); rt.block_on(async { task::block_in_place(|| {}); }); } #[test] fn can_enter_current_thread_rt_from_within_block_in_place() { let outer = tokio::runtime::Runtime::new().unwrap(); outer.block_on(async { tokio::task::block_in_place(|| { let inner = tokio::runtime::Builder::new_current_thread() .build() .unwrap(); inner.block_on(async {}) }) }); } #[test] #[cfg(panic = "unwind")] fn useful_panic_message_when_dropping_rt_in_rt() { use std::panic::{catch_unwind, AssertUnwindSafe}; let outer = tokio::runtime::Runtime::new().unwrap(); let result = catch_unwind(AssertUnwindSafe(|| { outer.block_on(async { let _ = tokio::runtime::Builder::new_current_thread() .build() .unwrap(); }); })); assert!(result.is_err()); let err = result.unwrap_err(); let err: &'static str = err.downcast_ref::<&'static str>().unwrap(); assert!( err.contains("Cannot drop a runtime"), "Wrong panic message: {err:?}" ); } #[test] fn can_shutdown_with_zero_timeout_in_runtime() { let outer = tokio::runtime::Runtime::new().unwrap(); outer.block_on(async { let rt = tokio::runtime::Builder::new_current_thread() .build() .unwrap(); rt.shutdown_timeout(Duration::from_nanos(0)); }); } #[test] fn can_shutdown_now_in_runtime() { let outer = tokio::runtime::Runtime::new().unwrap(); outer.block_on(async { let rt = tokio::runtime::Builder::new_current_thread() .build() .unwrap(); rt.shutdown_background(); }); } #[test] fn coop_disabled_in_block_in_place() { let outer = tokio::runtime::Builder::new_multi_thread() .enable_time() .build() .unwrap(); let (tx, rx) = support::mpsc_stream::unbounded_channel_stream(); for i in 0..200 { tx.send(i).unwrap(); } drop(tx); outer.block_on(async move { let jh = tokio::spawn(async move { tokio::task::block_in_place(move || { futures::executor::block_on(async move { use tokio_stream::StreamExt; assert_eq!(rx.fold(0, |n, _| n + 1).await, 200); }) }) }); tokio::time::timeout(Duration::from_secs(1), jh) .await .expect("timed out (probably hanging)") .unwrap() }); } #[test] fn coop_disabled_in_block_in_place_in_block_on() { let (done_tx, done_rx) = std::sync::mpsc::channel(); let done = done_tx.clone(); thread::spawn(move || { let outer = tokio::runtime::Runtime::new().unwrap(); let (tx, rx) = support::mpsc_stream::unbounded_channel_stream(); for i in 0..200 { tx.send(i).unwrap(); } drop(tx); outer.block_on(async move { tokio::task::block_in_place(move || { futures::executor::block_on(async move { use tokio_stream::StreamExt; assert_eq!(rx.fold(0, |n, _| n + 1).await, 200); }) }) }); let _ = done.send(Ok(())); }); thread::spawn(move || { thread::sleep(Duration::from_secs(1)); let _ = done_tx.send(Err("timed out (probably hanging)")); }); done_rx.recv().unwrap().unwrap(); } #[cfg(feature = "test-util")] #[tokio::test(start_paused = true)] async fn blocking_when_paused() { // Do not auto-advance time when we have started a blocking task that has // not yet finished. time::timeout( Duration::from_secs(3), task::spawn_blocking(|| thread::sleep(Duration::from_millis(1))), ) .await .expect("timeout should not trigger") .expect("blocking task should finish"); // Really: Do not auto-advance time, even if the timeout is short and the // blocking task runs for longer than that. It doesn't matter: Tokio time // is paused; system time is not. time::timeout( Duration::from_millis(1), task::spawn_blocking(|| thread::sleep(Duration::from_millis(50))), ) .await .expect("timeout should not trigger") .expect("blocking task should finish"); } #[cfg(feature = "test-util")] #[tokio::test(start_paused = true)] async fn blocking_task_wakes_paused_runtime() { let t0 = std::time::Instant::now(); time::timeout( Duration::from_secs(15), task::spawn_blocking(|| thread::sleep(Duration::from_millis(1))), ) .await .expect("timeout should not trigger") .expect("blocking task should finish"); assert!( t0.elapsed() < Duration::from_secs(10), "completing a spawn_blocking should wake the scheduler if it's parked while time is paused" ); } #[cfg(feature = "test-util")] #[tokio::test(start_paused = true)] async fn unawaited_blocking_task_wakes_paused_runtime() { let t0 = std::time::Instant::now(); // When this task finishes, time should auto-advance, even though the // JoinHandle has not been awaited yet. let a = task::spawn_blocking(|| { thread::sleep(Duration::from_millis(1)); }); crate::time::sleep(Duration::from_secs(15)).await; a.await.expect("blocking task should finish"); assert!( t0.elapsed() < Duration::from_secs(10), "completing a spawn_blocking should wake the scheduler if it's parked while time is paused" ); } #[cfg(panic = "unwind")] #[cfg(feature = "test-util")] #[tokio::test(start_paused = true)] async fn panicking_blocking_task_wakes_paused_runtime() { let t0 = std::time::Instant::now(); let result = time::timeout( Duration::from_secs(15), task::spawn_blocking(|| { thread::sleep(Duration::from_millis(1)); panic!("blocking task panicked"); }), ) .await .expect("timeout should not trigger"); assert!(result.is_err(), "blocking task should have panicked"); assert!( t0.elapsed() < Duration::from_secs(10), "completing a spawn_blocking should wake the scheduler if it's parked while time is paused" ); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/tcp_into_split.rs
tokio/tests/tcp_into_split.rs
#![warn(rust_2018_idioms)] #![cfg(all(feature = "full", not(target_os = "wasi"), not(miri)))] // Wasi doesn't support bind // No `socket` on miri. use std::io::{Error, ErrorKind, Result}; use std::io::{Read, Write}; use std::{net, thread}; use tokio::io::{AsyncReadExt, AsyncWriteExt}; use tokio::net::{TcpListener, TcpStream}; use tokio::try_join; #[tokio::test] async fn split() -> Result<()> { const MSG: &[u8] = b"split"; let listener = TcpListener::bind("127.0.0.1:0").await?; let addr = listener.local_addr()?; let (stream1, (mut stream2, _)) = try_join! { TcpStream::connect(&addr), listener.accept(), }?; let (mut read_half, mut write_half) = stream1.into_split(); let ((), (), ()) = try_join! { async { let len = stream2.write(MSG).await?; assert_eq!(len, MSG.len()); let mut read_buf = vec![0u8; 32]; let read_len = stream2.read(&mut read_buf).await?; assert_eq!(&read_buf[..read_len], MSG); Result::Ok(()) }, async { let len = write_half.write(MSG).await?; assert_eq!(len, MSG.len()); Ok(()) }, async { let mut read_buf = [0u8; 32]; let peek_len1 = read_half.peek(&mut read_buf[..]).await?; let peek_len2 = read_half.peek(&mut read_buf[..]).await?; assert_eq!(peek_len1, peek_len2); let read_len = read_half.read(&mut read_buf[..]).await?; assert_eq!(peek_len1, read_len); assert_eq!(&read_buf[..read_len], MSG); Ok(()) }, }?; Ok(()) } #[tokio::test] async fn reunite() -> Result<()> { let listener = net::TcpListener::bind("127.0.0.1:0")?; let addr = listener.local_addr()?; let handle = thread::spawn(move || { drop(listener.accept().unwrap()); drop(listener.accept().unwrap()); }); let stream1 = TcpStream::connect(&addr).await?; let (read1, write1) = stream1.into_split(); let stream2 = TcpStream::connect(&addr).await?; let (_, write2) = stream2.into_split(); let read1 = match read1.reunite(write2) { Ok(_) => panic!("Reunite should not succeed"), Err(err) => err.0, }; read1.reunite(write1).expect("Reunite should succeed"); handle.join().unwrap(); Ok(()) } /// Test that dropping the write half actually closes the stream. #[tokio::test] async fn drop_write() -> Result<()> { const MSG: &[u8] = b"split"; let listener = net::TcpListener::bind("127.0.0.1:0")?; let addr = listener.local_addr()?; let handle = thread::spawn(move || { let (mut stream, _) = listener.accept().unwrap(); stream.write_all(MSG).unwrap(); let mut read_buf = [0u8; 32]; let res = match stream.read(&mut read_buf) { Ok(0) => Ok(()), Ok(len) => Err(Error::new( ErrorKind::Other, format!("Unexpected read: {len} bytes."), )), Err(err) => Err(err), }; drop(stream); res }); let stream = TcpStream::connect(&addr).await?; let (mut read_half, write_half) = stream.into_split(); let mut read_buf = [0u8; 32]; let read_len = read_half.read(&mut read_buf[..]).await?; assert_eq!(&read_buf[..read_len], MSG); // drop it while the read is in progress std::thread::spawn(move || { thread::sleep(std::time::Duration::from_millis(10)); drop(write_half); }); match read_half.read(&mut read_buf[..]).await { Ok(0) => {} Ok(len) => panic!("Unexpected read: {len} bytes."), Err(err) => panic!("Unexpected error: {err}."), } handle.join().unwrap().unwrap(); Ok(()) }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/sync_mutex_owned.rs
tokio/tests/sync_mutex_owned.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "sync")] #[cfg(all(target_family = "wasm", not(target_os = "wasi")))] use wasm_bindgen_test::wasm_bindgen_test as test; #[cfg(all(target_family = "wasm", not(target_os = "wasi")))] use wasm_bindgen_test::wasm_bindgen_test as maybe_tokio_test; #[cfg(not(all(target_family = "wasm", not(target_os = "wasi"))))] use tokio::test as maybe_tokio_test; use tokio::sync::Mutex; use tokio_test::task::spawn; use tokio_test::{assert_pending, assert_ready}; use std::sync::Arc; #[test] fn straight_execution() { let l = Arc::new(Mutex::new(100)); { let mut t = spawn(l.clone().lock_owned()); let mut g = assert_ready!(t.poll()); assert_eq!(&*g, &100); *g = 99; } { let mut t = spawn(l.clone().lock_owned()); let mut g = assert_ready!(t.poll()); assert_eq!(&*g, &99); *g = 98; } { let mut t = spawn(l.lock_owned()); let g = assert_ready!(t.poll()); assert_eq!(&*g, &98); } } #[test] fn readiness() { let l = Arc::new(Mutex::new(100)); let mut t1 = spawn(l.clone().lock_owned()); let mut t2 = spawn(l.lock_owned()); let g = assert_ready!(t1.poll()); // We can't now acquire the lease since it's already held in g assert_pending!(t2.poll()); // But once g unlocks, we can acquire it drop(g); assert!(t2.is_woken()); assert_ready!(t2.poll()); } /// Ensure a mutex is unlocked if a future holding the lock /// is aborted prematurely. #[tokio::test] #[cfg(feature = "full")] #[cfg_attr(miri, ignore)] async fn aborted_future_1() { use std::time::Duration; use tokio::time::{interval, timeout}; let m1: Arc<Mutex<usize>> = Arc::new(Mutex::new(0)); { let m2 = m1.clone(); // Try to lock mutex in a future that is aborted prematurely timeout(Duration::from_millis(1u64), async move { let iv = interval(Duration::from_millis(1000)); tokio::pin!(iv); m2.lock_owned().await; iv.as_mut().tick().await; iv.as_mut().tick().await; }) .await .unwrap_err(); } // This should succeed as there is no lock left for the mutex. timeout(Duration::from_millis(1u64), async move { m1.lock_owned().await; }) .await .expect("Mutex is locked"); } /// This test is similar to `aborted_future_1` but this time the /// aborted future is waiting for the lock. #[tokio::test] #[cfg(feature = "full")] async fn aborted_future_2() { use std::time::Duration; use tokio::time::timeout; let m1: Arc<Mutex<usize>> = Arc::new(Mutex::new(0)); { // Lock mutex let _lock = m1.clone().lock_owned().await; { let m2 = m1.clone(); // Try to lock mutex in a future that is aborted prematurely timeout(Duration::from_millis(1u64), async move { m2.lock_owned().await; }) .await .unwrap_err(); } } // This should succeed as there is no lock left for the mutex. timeout(Duration::from_millis(1u64), async move { m1.lock_owned().await; }) .await .expect("Mutex is locked"); } #[test] fn try_lock_owned() { let m: Arc<Mutex<usize>> = Arc::new(Mutex::new(0)); { let g1 = m.clone().try_lock_owned(); assert!(g1.is_ok()); let g2 = m.clone().try_lock_owned(); assert!(g2.is_err()); } let g3 = m.try_lock_owned(); assert!(g3.is_ok()); } #[maybe_tokio_test] async fn debug_format() { let s = "debug"; let m = Arc::new(Mutex::new(s.to_string())); assert_eq!(format!("{s:?}"), format!("{:?}", m.lock_owned().await)); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/task_trace_self.rs
tokio/tests/task_trace_self.rs
#![allow(unknown_lints, unexpected_cfgs)] #![cfg(all( tokio_unstable, feature = "taskdump", target_os = "linux", any(target_arch = "aarch64", target_arch = "x86", target_arch = "x86_64") ))] use std::future::Future; use std::pin::Pin; use std::sync::{Arc, Mutex}; use std::task::{Context, Poll}; use std::time::{Duration, Instant}; use tokio::runtime::dump::{Root, Trace}; pin_project_lite::pin_project! { pub struct PrettyFuture<F: Future> { #[pin] f: Root<F>, t_last: State, logs: Arc<Mutex<Vec<Trace>>>, } } enum State { NotStarted, Running { since: Instant }, Alerted, } impl<F: Future> PrettyFuture<F> { pub fn pretty(f: F, logs: Arc<Mutex<Vec<Trace>>>) -> Self { PrettyFuture { f: Trace::root(f), t_last: State::NotStarted, logs, } } } impl<F: Future> Future for PrettyFuture<F> { type Output = F::Output; fn poll(self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<F::Output> { let mut this = self.project(); let now = Instant::now(); let t_last = match this.t_last { State::Running { since } => Some(*since), State::NotStarted => { *this.t_last = State::Running { since: now }; None } State::Alerted => { // don't double-alert for the same future None } }; if t_last.is_some_and(|t_last| now.duration_since(t_last) > Duration::from_millis(500)) { let (res, trace) = tokio::runtime::dump::Trace::capture(|| this.f.as_mut().poll(cx)); this.logs.lock().unwrap().push(trace); *this.t_last = State::Alerted; return res; } this.f.poll(cx) } } #[tokio::test] async fn task_trace_self() { let log = Arc::new(Mutex::new(vec![])); let log2 = Arc::new(Mutex::new(vec![])); let mut good_line = vec![]; let mut bad_line = vec![]; PrettyFuture::pretty( PrettyFuture::pretty( async { bad_line.push(line!() + 1); tokio::task::yield_now().await; bad_line.push(line!() + 1); tokio::time::sleep(Duration::from_millis(1)).await; for _ in 0..100 { good_line.push(line!() + 1); tokio::time::sleep(Duration::from_millis(10)).await; } }, log.clone(), ), log2.clone(), ) .await; for line in good_line { let s = format!("{}:{}:", file!(), line); assert!(log.lock().unwrap().iter().any(|x| { eprintln!("{x}"); format!("{x}").contains(&s) })); } for line in bad_line { let s = format!("{}:{}:", file!(), line); assert!(!log .lock() .unwrap() .iter() .any(|x| format!("{x}").contains(&s))); } }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false
tokio-rs/tokio
https://github.com/tokio-rs/tokio/blob/41d1877689f8669902b003a6affce60bdfeb3025/tokio/tests/io_write_all.rs
tokio/tests/io_write_all.rs
#![warn(rust_2018_idioms)] #![cfg(feature = "full")] use tokio::io::{AsyncWrite, AsyncWriteExt}; use tokio_test::assert_ok; use bytes::BytesMut; use std::cmp; use std::io; use std::pin::Pin; use std::task::{Context, Poll}; #[tokio::test] async fn write_all() { struct Wr { buf: BytesMut, cnt: usize, } impl AsyncWrite for Wr { fn poll_write( mut self: Pin<&mut Self>, _cx: &mut Context<'_>, buf: &[u8], ) -> Poll<io::Result<usize>> { let n = cmp::min(4, buf.len()); let buf = &buf[0..n]; self.cnt += 1; self.buf.extend(buf); Ok(buf.len()).into() } fn poll_flush(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<io::Result<()>> { Ok(()).into() } fn poll_shutdown(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<io::Result<()>> { Ok(()).into() } } let mut wr = Wr { buf: BytesMut::with_capacity(64), cnt: 0, }; assert_ok!(wr.write_all(b"hello world").await); assert_eq!(wr.buf, b"hello world"[..]); assert_eq!(wr.cnt, 3); }
rust
MIT
41d1877689f8669902b003a6affce60bdfeb3025
2026-01-04T15:33:40.250594Z
false