// Copyright 2017 Parity Technologies (UK) Ltd. // // Permission is hereby granted, free of charge, to any person obtaining a // copy of this software and associated documentation files (the "Software"), // to deal in the Software without restriction, including without limitation // the rights to use, copy, modify, merge, publish, distribute, sublicense, // and/or sell copies of the Software, and to permit persons to whom the // Software is furnished to do so, subject to the following conditions: // // The above copyright notice and this permission notice shall be included in // all copies or substantial portions of the Software. // // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS // OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, // FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE // AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER // LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING // FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER // DEALINGS IN THE SOFTWARE. //! Implementation of the libp2p [`libp2p_core::Transport`] trait for TCP/IP. //! //! # Usage //! //! This crate provides a [`async_io::Transport`] and [`tokio::Transport`], depending on //! the enabled features, which implement the [`libp2p_core::Transport`] trait for use as a //! transport with `libp2p-core` or `libp2p-swarm`. #![cfg_attr(docsrs, feature(doc_cfg, doc_auto_cfg))] mod provider; #[cfg(feature = "async-io")] pub use provider::async_io; #[cfg(feature = "tokio")] pub use provider::tokio; use futures::{future::Ready, prelude::*, stream::SelectAll}; use futures_timer::Delay; use if_watch::IfEvent; use libp2p_core::{ multiaddr::{Multiaddr, Protocol}, transport::{DialOpts, ListenerId, PortUse, TransportError, TransportEvent}, }; use provider::{Incoming, Provider}; use socket2::{Domain, Socket, Type}; use std::{ collections::{HashSet, VecDeque}, io, net::{IpAddr, Ipv4Addr, Ipv6Addr, SocketAddr, TcpListener}, pin::Pin, sync::{Arc, RwLock}, task::{Context, Poll, Waker}, time::Duration, }; /// The configuration for a TCP/IP transport capability for libp2p. #[derive(Clone, Debug)] pub struct Config { /// TTL to set for opened sockets, or `None` to keep default. ttl: Option, /// `TCP_NODELAY` to set for opened sockets, or `None` to keep default. nodelay: Option, /// Size of the listen backlog for listen sockets. backlog: u32, } type Port = u16; /// The configuration for port reuse of listening sockets. #[derive(Debug, Clone, Default)] struct PortReuse { /// The addresses and ports of the listening sockets /// registered as eligible for port reuse when dialing listen_addrs: Arc>>, } impl PortReuse { /// Registers a socket address for port reuse. /// /// Has no effect if port reuse is disabled. fn register(&mut self, ip: IpAddr, port: Port) { tracing::trace!(%ip, %port, "Registering for port reuse"); self.listen_addrs .write() .expect("`register()` and `unregister()` never panic while holding the lock") .insert((ip, port)); } /// Unregisters a socket address for port reuse. /// /// Has no effect if port reuse is disabled. fn unregister(&mut self, ip: IpAddr, port: Port) { tracing::trace!(%ip, %port, "Unregistering for port reuse"); self.listen_addrs .write() .expect("`register()` and `unregister()` never panic while holding the lock") .remove(&(ip, port)); } /// Selects a listening socket address suitable for use /// as the local socket address when dialing. /// /// If multiple listening sockets are registered for port /// reuse, one is chosen whose IP protocol version and /// loopback status is the same as that of `remote_ip`. /// /// Returns `None` if port reuse is disabled or no suitable /// listening socket address is found. fn local_dial_addr(&self, remote_ip: &IpAddr) -> Option { for (ip, port) in self .listen_addrs .read() .expect("`local_dial_addr` never panic while holding the lock") .iter() { if ip.is_ipv4() == remote_ip.is_ipv4() && ip.is_loopback() == remote_ip.is_loopback() { if remote_ip.is_ipv4() { return Some(SocketAddr::new(IpAddr::V4(Ipv4Addr::UNSPECIFIED), *port)); } else { return Some(SocketAddr::new(IpAddr::V6(Ipv6Addr::UNSPECIFIED), *port)); } } } None } } impl Config { /// Creates a new configuration for a TCP/IP transport: /// /// * Nagle's algorithm, i.e. `TCP_NODELAY`, is _enabled_. /// See [`Config::nodelay`]. /// * Reuse of listening ports is _disabled_. /// See [`Config::port_reuse`]. /// * No custom `IP_TTL` is set. The default of the OS TCP stack applies. /// See [`Config::ttl`]. /// * The size of the listen backlog for new listening sockets is `1024`. /// See [`Config::listen_backlog`]. pub fn new() -> Self { Self { ttl: None, nodelay: Some(false), // Disable Nagle's algorithm by default backlog: 1024, } } /// Configures the `IP_TTL` option for new sockets. pub fn ttl(mut self, value: u32) -> Self { self.ttl = Some(value); self } /// Configures the `TCP_NODELAY` option for new sockets. pub fn nodelay(mut self, value: bool) -> Self { self.nodelay = Some(value); self } /// Configures the listen backlog for new listen sockets. pub fn listen_backlog(mut self, backlog: u32) -> Self { self.backlog = backlog; self } /// Configures port reuse for local sockets, which implies /// reuse of listening ports for outgoing connections to /// enhance NAT traversal capabilities. /// /// # Deprecation Notice /// /// The new implementation works on a per-connaction basis, defined by the behaviour. This /// removes the necessaity to configure the transport for port reuse, instead the behaviour /// requiring this behaviour can decide whether to use port reuse or not. /// /// The API to configure port reuse is part of [`Transport`] and the option can be found in /// [`libp2p_core::transport::DialOpts`]. /// /// If [`PortUse::Reuse`] is enabled, the transport will try to reuse the local port of the /// listener. If that's not possible, i.e. there is no listener or the transport doesn't allow /// a direct control over ports, a new port (or the default behaviour) is used. If port reuse /// is enabled for a connection, this option will be treated on a best-effor basis. #[deprecated( since = "0.42.0", note = "This option does nothing now, since the port reuse policy is now decided on a per-connection basis by the behaviour. The function will be removed in a future release." )] pub fn port_reuse(self, _port_reuse: bool) -> Self { self } fn create_socket(&self, socket_addr: SocketAddr, port_use: PortUse) -> io::Result { let socket = Socket::new( Domain::for_address(socket_addr), Type::STREAM, Some(socket2::Protocol::TCP), )?; if socket_addr.is_ipv6() { socket.set_only_v6(true)?; } if let Some(ttl) = self.ttl { socket.set_ttl(ttl)?; } if let Some(nodelay) = self.nodelay { socket.set_nodelay(nodelay)?; } socket.set_reuse_address(true)?; #[cfg(all(unix, not(any(target_os = "solaris", target_os = "illumos"))))] if port_use == PortUse::Reuse { socket.set_reuse_port(true)?; } #[cfg(not(all(unix, not(any(target_os = "solaris", target_os = "illumos")))))] let _ = port_use; // silence the unused warning on non-unix platforms (i.e. Windows) socket.set_nonblocking(true)?; Ok(socket) } } impl Default for Config { fn default() -> Self { Self::new() } } /// An abstract [`libp2p_core::Transport`] implementation. /// /// You shouldn't need to use this type directly. Use one of the following instead: /// /// - [`tokio::Transport`] /// - [`async_io::Transport`] pub struct Transport where T: Provider + Send, { config: Config, /// The configuration of port reuse when dialing. port_reuse: PortReuse, /// All the active listeners. /// The [`ListenStream`] struct contains a stream that we want to be pinned. Since the `VecDeque` /// can be resized, the only way is to use a `Pin>`. listeners: SelectAll>, /// Pending transport events to return from [`libp2p_core::Transport::poll`]. pending_events: VecDeque::ListenerUpgrade, io::Error>>, } impl Transport where T: Provider + Send, { /// Create a new instance of [`Transport`]. /// /// If you don't want to specify a [`Config`], use [`Transport::default`]. /// /// It is best to call this function through one of the type-aliases of this type: /// /// - [`tokio::Transport::new`] /// - [`async_io::Transport::new`] pub fn new(config: Config) -> Self { Transport { config, ..Default::default() } } fn do_listen( &mut self, id: ListenerId, socket_addr: SocketAddr, ) -> io::Result> { let socket = self.config.create_socket(socket_addr, PortUse::Reuse)?; socket.bind(&socket_addr.into())?; socket.listen(self.config.backlog as _)?; socket.set_nonblocking(true)?; let listener: TcpListener = socket.into(); let local_addr = listener.local_addr()?; if local_addr.ip().is_unspecified() { return ListenStream::::new( id, listener, Some(T::new_if_watcher()?), self.port_reuse.clone(), ); } self.port_reuse.register(local_addr.ip(), local_addr.port()); let listen_addr = ip_to_multiaddr(local_addr.ip(), local_addr.port()); self.pending_events.push_back(TransportEvent::NewAddress { listener_id: id, listen_addr, }); ListenStream::::new(id, listener, None, self.port_reuse.clone()) } } impl Default for Transport where T: Provider + Send, { /// Creates a [`Transport`] with reasonable defaults. /// /// This transport will have port-reuse disabled. fn default() -> Self { Transport { port_reuse: PortReuse::default(), config: Config::default(), listeners: SelectAll::new(), pending_events: VecDeque::new(), } } } impl libp2p_core::Transport for Transport where T: Provider + Send + 'static, T::Listener: Unpin, T::Stream: Unpin, { type Output = T::Stream; type Error = io::Error; type Dial = Pin> + Send>>; type ListenerUpgrade = Ready>; fn listen_on( &mut self, id: ListenerId, addr: Multiaddr, ) -> Result<(), TransportError> { let socket_addr = multiaddr_to_socketaddr(addr.clone()) .map_err(|_| TransportError::MultiaddrNotSupported(addr))?; tracing::debug!("listening on {}", socket_addr); let listener = self .do_listen(id, socket_addr) .map_err(TransportError::Other)?; self.listeners.push(listener); Ok(()) } fn remove_listener(&mut self, id: ListenerId) -> bool { if let Some(listener) = self.listeners.iter_mut().find(|l| l.listener_id == id) { listener.close(Ok(())); true } else { false } } fn dial( &mut self, addr: Multiaddr, opts: DialOpts, ) -> Result> { let socket_addr = if let Ok(socket_addr) = multiaddr_to_socketaddr(addr.clone()) { if socket_addr.port() == 0 || socket_addr.ip().is_unspecified() { return Err(TransportError::MultiaddrNotSupported(addr)); } socket_addr } else { return Err(TransportError::MultiaddrNotSupported(addr)); }; tracing::debug!(address=%socket_addr, "dialing address"); let socket = self .config .create_socket(socket_addr, opts.port_use) .map_err(TransportError::Other)?; let bind_addr = match self.port_reuse.local_dial_addr(&socket_addr.ip()) { Some(socket_addr) if opts.port_use == PortUse::Reuse => { tracing::trace!(address=%addr, "Binding dial socket to listen socket address"); Some(socket_addr) } _ => None, }; let local_config = self.config.clone(); Ok(async move { if let Some(bind_addr) = bind_addr { socket.bind(&bind_addr.into())?; } // [`Transport::dial`] should do no work unless the returned [`Future`] is polled. Thus // do the `connect` call within the [`Future`]. let socket = match (socket.connect(&socket_addr.into()), bind_addr) { (Ok(()), _) => socket, (Err(err), _) if err.raw_os_error() == Some(libc::EINPROGRESS) => socket, (Err(err), _) if err.kind() == io::ErrorKind::WouldBlock => socket, (Err(err), Some(bind_addr)) if err.kind() == io::ErrorKind::AddrNotAvailable => { // The socket was bound to a local address that is no longer available. // Retry without binding. tracing::debug!(connect_addr = %socket_addr, ?bind_addr, "Failed to connect using existing socket because we already have a connection, re-dialing with new port"); std::mem::drop(socket); let socket = local_config.create_socket(socket_addr, PortUse::New)?; match socket.connect(&socket_addr.into()) { Ok(()) => socket, Err(err) if err.raw_os_error() == Some(libc::EINPROGRESS) => socket, Err(err) if err.kind() == io::ErrorKind::WouldBlock => socket, Err(err) => return Err(err), } } (Err(err), _) => return Err(err), }; let stream = T::new_stream(socket.into()).await?; Ok(stream) } .boxed()) } /// Poll all listeners. #[tracing::instrument(level = "trace", name = "Transport::poll", skip(self, cx))] fn poll( mut self: Pin<&mut Self>, cx: &mut Context<'_>, ) -> Poll> { // Return pending events from closed listeners. if let Some(event) = self.pending_events.pop_front() { return Poll::Ready(event); } match self.listeners.poll_next_unpin(cx) { Poll::Ready(Some(transport_event)) => Poll::Ready(transport_event), _ => Poll::Pending, } } } /// A stream of incoming connections on one or more interfaces. struct ListenStream where T: Provider, { /// The ID of this listener. listener_id: ListenerId, /// The socket address that the listening socket is bound to, /// which may be a "wildcard address" like `INADDR_ANY` or `IN6ADDR_ANY` /// when listening on all interfaces for IPv4 respectively IPv6 connections. listen_addr: SocketAddr, /// The async listening socket for incoming connections. listener: T::Listener, /// Watcher for network interface changes. /// Reports [`IfEvent`]s for new / deleted ip-addresses when interfaces /// become or stop being available. /// /// `None` if the socket is only listening on a single interface. if_watcher: Option, /// The port reuse configuration for outgoing connections. /// /// If enabled, all IP addresses on which this listening stream /// is accepting connections (`in_addr`) are registered for reuse /// as local addresses for the sockets of outgoing connections. They are /// unregistered when the stream encounters an error or is dropped. port_reuse: PortReuse, /// How long to sleep after a (non-fatal) error while trying /// to accept a new connection. sleep_on_error: Duration, /// The current pause, if any. pause: Option, /// Pending event to reported. pending_event: Option<::Item>, /// The listener can be manually closed with [`Transport::remove_listener`](libp2p_core::Transport::remove_listener). is_closed: bool, /// The stream must be awaken after it has been closed to deliver the last event. close_listener_waker: Option, } impl ListenStream where T: Provider, { /// Constructs a [`ListenStream`] for incoming connections around /// the given [`TcpListener`]. fn new( listener_id: ListenerId, listener: TcpListener, if_watcher: Option, port_reuse: PortReuse, ) -> io::Result { let listen_addr = listener.local_addr()?; let listener = T::new_listener(listener)?; Ok(ListenStream { port_reuse, listener, listener_id, listen_addr, if_watcher, pause: None, sleep_on_error: Duration::from_millis(100), pending_event: None, is_closed: false, close_listener_waker: None, }) } /// Disables port reuse for any listen address of this stream. /// /// This is done when the [`ListenStream`] encounters a fatal /// error (for the stream) or is dropped. /// /// Has no effect if port reuse is disabled. fn disable_port_reuse(&mut self) { match &self.if_watcher { Some(if_watcher) => { for ip_net in T::addrs(if_watcher) { self.port_reuse .unregister(ip_net.addr(), self.listen_addr.port()); } } None => self .port_reuse .unregister(self.listen_addr.ip(), self.listen_addr.port()), } } /// Close the listener. /// /// This will create a [`TransportEvent::ListenerClosed`] and /// terminate the stream once the event has been reported. fn close(&mut self, reason: Result<(), io::Error>) { if self.is_closed { return; } self.pending_event = Some(TransportEvent::ListenerClosed { listener_id: self.listener_id, reason, }); self.is_closed = true; // Wake the stream to deliver the last event. if let Some(waker) = self.close_listener_waker.take() { waker.wake(); } } /// Poll for a next If Event. fn poll_if_addr(&mut self, cx: &mut Context<'_>) -> Poll<::Item> { let Some(if_watcher) = self.if_watcher.as_mut() else { return Poll::Pending; }; let my_listen_addr_port = self.listen_addr.port(); while let Poll::Ready(Some(event)) = if_watcher.poll_next_unpin(cx) { match event { Ok(IfEvent::Up(inet)) => { let ip = inet.addr(); if self.listen_addr.is_ipv4() == ip.is_ipv4() { let ma = ip_to_multiaddr(ip, my_listen_addr_port); tracing::debug!(address=%ma, "New listen address"); self.port_reuse.register(ip, my_listen_addr_port); return Poll::Ready(TransportEvent::NewAddress { listener_id: self.listener_id, listen_addr: ma, }); } } Ok(IfEvent::Down(inet)) => { let ip = inet.addr(); if self.listen_addr.is_ipv4() == ip.is_ipv4() { let ma = ip_to_multiaddr(ip, my_listen_addr_port); tracing::debug!(address=%ma, "Expired listen address"); self.port_reuse.unregister(ip, my_listen_addr_port); return Poll::Ready(TransportEvent::AddressExpired { listener_id: self.listener_id, listen_addr: ma, }); } } Err(error) => { self.pause = Some(Delay::new(self.sleep_on_error)); return Poll::Ready(TransportEvent::ListenerError { listener_id: self.listener_id, error, }); } } } Poll::Pending } } impl Drop for ListenStream where T: Provider, { fn drop(&mut self) { self.disable_port_reuse(); } } impl Stream for ListenStream where T: Provider, T::Listener: Unpin, T::Stream: Unpin, { type Item = TransportEvent>, io::Error>; fn poll_next(mut self: Pin<&mut Self>, cx: &mut Context) -> Poll> { if let Some(mut pause) = self.pause.take() { match pause.poll_unpin(cx) { Poll::Ready(_) => {} Poll::Pending => { self.pause = Some(pause); return Poll::Pending; } } } if let Some(event) = self.pending_event.take() { return Poll::Ready(Some(event)); } if self.is_closed { // Terminate the stream if the listener closed and all remaining events have been reported. return Poll::Ready(None); } if let Poll::Ready(event) = self.poll_if_addr(cx) { return Poll::Ready(Some(event)); } // Take the pending connection from the backlog. match T::poll_accept(&mut self.listener, cx) { Poll::Ready(Ok(Incoming { local_addr, remote_addr, stream, })) => { let local_addr = ip_to_multiaddr(local_addr.ip(), local_addr.port()); let remote_addr = ip_to_multiaddr(remote_addr.ip(), remote_addr.port()); tracing::debug!( remote_address=%remote_addr, local_address=%local_addr, "Incoming connection from remote at local" ); return Poll::Ready(Some(TransportEvent::Incoming { listener_id: self.listener_id, upgrade: future::ok(stream), local_addr, send_back_addr: remote_addr, })); } Poll::Ready(Err(error)) => { // These errors are non-fatal for the listener stream. self.pause = Some(Delay::new(self.sleep_on_error)); return Poll::Ready(Some(TransportEvent::ListenerError { listener_id: self.listener_id, error, })); } Poll::Pending => {} } self.close_listener_waker = Some(cx.waker().clone()); Poll::Pending } } /// Extracts a `SocketAddr` from a given `Multiaddr`. /// /// Fails if the given `Multiaddr` does not begin with an IP /// protocol encapsulating a TCP port. fn multiaddr_to_socketaddr(mut addr: Multiaddr) -> Result { // "Pop" the IP address and TCP port from the end of the address, // ignoring a `/p2p/...` suffix as well as any prefix of possibly // outer protocols, if present. let mut port = None; while let Some(proto) = addr.pop() { match proto { Protocol::Ip4(ipv4) => match port { Some(port) => return Ok(SocketAddr::new(ipv4.into(), port)), None => return Err(()), }, Protocol::Ip6(ipv6) => match port { Some(port) => return Ok(SocketAddr::new(ipv6.into(), port)), None => return Err(()), }, Protocol::Tcp(portnum) => match port { Some(_) => return Err(()), None => port = Some(portnum), }, Protocol::P2p(_) => {} _ => return Err(()), } } Err(()) } // Create a [`Multiaddr`] from the given IP address and port number. fn ip_to_multiaddr(ip: IpAddr, port: u16) -> Multiaddr { Multiaddr::empty().with(ip.into()).with(Protocol::Tcp(port)) } #[cfg(test)] mod tests { use super::*; use futures::{ channel::{mpsc, oneshot}, future::poll_fn, }; use libp2p_core::Endpoint; use libp2p_core::Transport as _; #[test] fn multiaddr_to_tcp_conversion() { use std::net::{IpAddr, Ipv4Addr, Ipv6Addr}; assert!( multiaddr_to_socketaddr("/ip4/127.0.0.1/udp/1234".parse::().unwrap()) .is_err() ); assert_eq!( multiaddr_to_socketaddr("/ip4/127.0.0.1/tcp/12345".parse::().unwrap()), Ok(SocketAddr::new( IpAddr::V4(Ipv4Addr::new(127, 0, 0, 1)), 12345, )) ); assert_eq!( multiaddr_to_socketaddr( "/ip4/255.255.255.255/tcp/8080" .parse::() .unwrap() ), Ok(SocketAddr::new( IpAddr::V4(Ipv4Addr::new(255, 255, 255, 255)), 8080, )) ); assert_eq!( multiaddr_to_socketaddr("/ip6/::1/tcp/12345".parse::().unwrap()), Ok(SocketAddr::new( IpAddr::V6(Ipv6Addr::new(0, 0, 0, 0, 0, 0, 0, 1)), 12345, )) ); assert_eq!( multiaddr_to_socketaddr( "/ip6/ffff:ffff:ffff:ffff:ffff:ffff:ffff:ffff/tcp/8080" .parse::() .unwrap() ), Ok(SocketAddr::new( IpAddr::V6(Ipv6Addr::new( 65535, 65535, 65535, 65535, 65535, 65535, 65535, 65535, )), 8080, )) ); } #[test] fn communicating_between_dialer_and_listener() { let _ = tracing_subscriber::fmt() .with_env_filter(tracing_subscriber::EnvFilter::from_default_env()) .try_init(); async fn listener(addr: Multiaddr, mut ready_tx: mpsc::Sender) { let mut tcp = Transport::::default().boxed(); tcp.listen_on(ListenerId::next(), addr).unwrap(); loop { match tcp.select_next_some().await { TransportEvent::NewAddress { listen_addr, .. } => { ready_tx.send(listen_addr).await.unwrap(); } TransportEvent::Incoming { upgrade, .. } => { let mut upgrade = upgrade.await.unwrap(); let mut buf = [0u8; 3]; upgrade.read_exact(&mut buf).await.unwrap(); assert_eq!(buf, [1, 2, 3]); upgrade.write_all(&[4, 5, 6]).await.unwrap(); return; } e => panic!("Unexpected transport event: {e:?}"), } } } async fn dialer(mut ready_rx: mpsc::Receiver) { let addr = ready_rx.next().await.unwrap(); let mut tcp = Transport::::default(); // Obtain a future socket through dialing let mut socket = tcp .dial( addr.clone(), DialOpts { role: Endpoint::Dialer, port_use: PortUse::Reuse, }, ) .unwrap() .await .unwrap(); socket.write_all(&[0x1, 0x2, 0x3]).await.unwrap(); let mut buf = [0u8; 3]; socket.read_exact(&mut buf).await.unwrap(); assert_eq!(buf, [4, 5, 6]); } fn test(addr: Multiaddr) { #[cfg(feature = "async-io")] { let (ready_tx, ready_rx) = mpsc::channel(1); let listener = listener::(addr.clone(), ready_tx); let dialer = dialer::(ready_rx); let listener = async_std::task::spawn(listener); async_std::task::block_on(dialer); async_std::task::block_on(listener); } #[cfg(feature = "tokio")] { let (ready_tx, ready_rx) = mpsc::channel(1); let listener = listener::(addr, ready_tx); let dialer = dialer::(ready_rx); let rt = ::tokio::runtime::Builder::new_current_thread() .enable_io() .build() .unwrap(); let tasks = ::tokio::task::LocalSet::new(); let listener = tasks.spawn_local(listener); tasks.block_on(&rt, dialer); tasks.block_on(&rt, listener).unwrap(); } } test("/ip4/127.0.0.1/tcp/0".parse().unwrap()); test("/ip6/::1/tcp/0".parse().unwrap()); } #[test] fn wildcard_expansion() { let _ = tracing_subscriber::fmt() .with_env_filter(tracing_subscriber::EnvFilter::from_default_env()) .try_init(); async fn listener(addr: Multiaddr, mut ready_tx: mpsc::Sender) { let mut tcp = Transport::::default().boxed(); tcp.listen_on(ListenerId::next(), addr).unwrap(); loop { match tcp.select_next_some().await { TransportEvent::NewAddress { listen_addr, .. } => { let mut iter = listen_addr.iter(); match iter.next().expect("ip address") { Protocol::Ip4(ip) => assert!(!ip.is_unspecified()), Protocol::Ip6(ip) => assert!(!ip.is_unspecified()), other => panic!("Unexpected protocol: {other}"), } if let Protocol::Tcp(port) = iter.next().expect("port") { assert_ne!(0, port) } else { panic!("No TCP port in address: {listen_addr}") } ready_tx.send(listen_addr).await.ok(); } TransportEvent::Incoming { .. } => { return; } _ => {} } } } async fn dialer(mut ready_rx: mpsc::Receiver) { let dest_addr = ready_rx.next().await.unwrap(); let mut tcp = Transport::::default(); tcp.dial( dest_addr, DialOpts { role: Endpoint::Dialer, port_use: PortUse::New, }, ) .unwrap() .await .unwrap(); } fn test(addr: Multiaddr) { #[cfg(feature = "async-io")] { let (ready_tx, ready_rx) = mpsc::channel(1); let listener = listener::(addr.clone(), ready_tx); let dialer = dialer::(ready_rx); let listener = async_std::task::spawn(listener); async_std::task::block_on(dialer); async_std::task::block_on(listener); } #[cfg(feature = "tokio")] { let (ready_tx, ready_rx) = mpsc::channel(1); let listener = listener::(addr, ready_tx); let dialer = dialer::(ready_rx); let rt = ::tokio::runtime::Builder::new_current_thread() .enable_io() .build() .unwrap(); let tasks = ::tokio::task::LocalSet::new(); let listener = tasks.spawn_local(listener); tasks.block_on(&rt, dialer); tasks.block_on(&rt, listener).unwrap(); } } test("/ip4/0.0.0.0/tcp/0".parse().unwrap()); test("/ip6/::1/tcp/0".parse().unwrap()); } #[test] fn port_reuse_dialing() { let _ = tracing_subscriber::fmt() .with_env_filter(tracing_subscriber::EnvFilter::from_default_env()) .try_init(); async fn listener( addr: Multiaddr, mut ready_tx: mpsc::Sender, port_reuse_rx: oneshot::Receiver>, ) { let mut tcp = Transport::::new(Config::new()).boxed(); tcp.listen_on(ListenerId::next(), addr).unwrap(); loop { match tcp.select_next_some().await { TransportEvent::NewAddress { listen_addr, .. } => { ready_tx.send(listen_addr).await.ok(); } TransportEvent::Incoming { upgrade, mut send_back_addr, .. } => { // Receive the dialer tcp port reuse let remote_port_reuse = port_reuse_rx.await.unwrap(); // And check it is the same as the remote port used for upgrade assert_eq!(send_back_addr.pop().unwrap(), remote_port_reuse); let mut upgrade = upgrade.await.unwrap(); let mut buf = [0u8; 3]; upgrade.read_exact(&mut buf).await.unwrap(); assert_eq!(buf, [1, 2, 3]); upgrade.write_all(&[4, 5, 6]).await.unwrap(); return; } e => panic!("Unexpected event: {e:?}"), } } } async fn dialer( addr: Multiaddr, mut ready_rx: mpsc::Receiver, port_reuse_tx: oneshot::Sender>, ) { let dest_addr = ready_rx.next().await.unwrap(); let mut tcp = Transport::::new(Config::new()); tcp.listen_on(ListenerId::next(), addr).unwrap(); match poll_fn(|cx| Pin::new(&mut tcp).poll(cx)).await { TransportEvent::NewAddress { .. } => { // Check that tcp and listener share the same port reuse SocketAddr let listener = tcp.listeners.iter().next().unwrap(); let port_reuse_tcp = tcp.port_reuse.local_dial_addr(&listener.listen_addr.ip()); let port_reuse_listener = listener .port_reuse .local_dial_addr(&listener.listen_addr.ip()); assert!(port_reuse_tcp.is_some()); assert_eq!(port_reuse_tcp, port_reuse_listener); // Send the dialer tcp port reuse to the listener port_reuse_tx .send(Protocol::Tcp(port_reuse_tcp.unwrap().port())) .ok(); // Obtain a future socket through dialing let mut socket = tcp .dial( dest_addr, DialOpts { role: Endpoint::Dialer, port_use: PortUse::Reuse, }, ) .unwrap() .await .unwrap(); socket.write_all(&[0x1, 0x2, 0x3]).await.unwrap(); // socket.flush().await; let mut buf = [0u8; 3]; socket.read_exact(&mut buf).await.unwrap(); assert_eq!(buf, [4, 5, 6]); } e => panic!("Unexpected transport event: {e:?}"), } } fn test(addr: Multiaddr) { #[cfg(feature = "async-io")] { let (ready_tx, ready_rx) = mpsc::channel(1); let (port_reuse_tx, port_reuse_rx) = oneshot::channel(); let listener = listener::(addr.clone(), ready_tx, port_reuse_rx); let dialer = dialer::(addr.clone(), ready_rx, port_reuse_tx); let listener = async_std::task::spawn(listener); async_std::task::block_on(dialer); async_std::task::block_on(listener); } #[cfg(feature = "tokio")] { let (ready_tx, ready_rx) = mpsc::channel(1); let (port_reuse_tx, port_reuse_rx) = oneshot::channel(); let listener = listener::(addr.clone(), ready_tx, port_reuse_rx); let dialer = dialer::(addr, ready_rx, port_reuse_tx); let rt = ::tokio::runtime::Builder::new_current_thread() .enable_io() .build() .unwrap(); let tasks = ::tokio::task::LocalSet::new(); let listener = tasks.spawn_local(listener); tasks.block_on(&rt, dialer); tasks.block_on(&rt, listener).unwrap(); } } test("/ip4/127.0.0.1/tcp/0".parse().unwrap()); test("/ip6/::1/tcp/0".parse().unwrap()); } #[test] fn port_reuse_listening() { let _ = tracing_subscriber::fmt() .with_env_filter(tracing_subscriber::EnvFilter::from_default_env()) .try_init(); async fn listen_twice(addr: Multiaddr) { let mut tcp = Transport::::new(Config::new()); tcp.listen_on(ListenerId::next(), addr).unwrap(); match poll_fn(|cx| Pin::new(&mut tcp).poll(cx)).await { TransportEvent::NewAddress { listen_addr: addr1, .. } => { let listener1 = tcp.listeners.iter().next().unwrap(); let port_reuse_tcp = tcp.port_reuse.local_dial_addr(&listener1.listen_addr.ip()); let port_reuse_listener1 = listener1 .port_reuse .local_dial_addr(&listener1.listen_addr.ip()); assert!(port_reuse_tcp.is_some()); assert_eq!(port_reuse_tcp, port_reuse_listener1); // Listen on the same address a second time. tcp.listen_on(ListenerId::next(), addr1.clone()).unwrap(); match poll_fn(|cx| Pin::new(&mut tcp).poll(cx)).await { TransportEvent::NewAddress { listen_addr: addr2, .. } => assert_eq!(addr1, addr2), e => panic!("Unexpected transport event: {e:?}"), } } e => panic!("Unexpected transport event: {e:?}"), } } fn test(addr: Multiaddr) { #[cfg(feature = "async-io")] { let listener = listen_twice::(addr.clone()); async_std::task::block_on(listener); } #[cfg(feature = "tokio")] { let listener = listen_twice::(addr); let rt = ::tokio::runtime::Builder::new_current_thread() .enable_io() .build() .unwrap(); rt.block_on(listener); } } test("/ip4/127.0.0.1/tcp/0".parse().unwrap()); } #[test] fn listen_port_0() { let _ = tracing_subscriber::fmt() .with_env_filter(tracing_subscriber::EnvFilter::from_default_env()) .try_init(); async fn listen(addr: Multiaddr) -> Multiaddr { let mut tcp = Transport::::default().boxed(); tcp.listen_on(ListenerId::next(), addr).unwrap(); tcp.select_next_some() .await .into_new_address() .expect("listen address") } fn test(addr: Multiaddr) { #[cfg(feature = "async-io")] { let new_addr = async_std::task::block_on(listen::(addr.clone())); assert!(!new_addr.to_string().contains("tcp/0")); } #[cfg(feature = "tokio")] { let rt = ::tokio::runtime::Builder::new_current_thread() .enable_io() .build() .unwrap(); let new_addr = rt.block_on(listen::(addr)); assert!(!new_addr.to_string().contains("tcp/0")); } } test("/ip6/::1/tcp/0".parse().unwrap()); test("/ip4/127.0.0.1/tcp/0".parse().unwrap()); } #[test] fn listen_invalid_addr() { let _ = tracing_subscriber::fmt() .with_env_filter(tracing_subscriber::EnvFilter::from_default_env()) .try_init(); fn test(addr: Multiaddr) { #[cfg(feature = "async-io")] { let mut tcp = async_io::Transport::default(); assert!(tcp.listen_on(ListenerId::next(), addr.clone()).is_err()); } #[cfg(feature = "tokio")] { let mut tcp = tokio::Transport::default(); assert!(tcp.listen_on(ListenerId::next(), addr).is_err()); } } test("/ip4/127.0.0.1/tcp/12345/tcp/12345".parse().unwrap()); } #[test] fn test_remove_listener() { let _ = tracing_subscriber::fmt() .with_env_filter(tracing_subscriber::EnvFilter::from_default_env()) .try_init(); async fn cycle_listeners() -> bool { let mut tcp = Transport::::default().boxed(); let listener_id = ListenerId::next(); tcp.listen_on(listener_id, "/ip4/127.0.0.1/tcp/0".parse().unwrap()) .unwrap(); tcp.remove_listener(listener_id) } #[cfg(feature = "async-io")] { assert!(async_std::task::block_on(cycle_listeners::())); } #[cfg(feature = "tokio")] { let rt = ::tokio::runtime::Builder::new_current_thread() .enable_io() .build() .unwrap(); assert!(rt.block_on(cycle_listeners::())); } } #[test] fn test_listens_ipv4_ipv6_separately() { fn test() { let port = { let listener = TcpListener::bind("127.0.0.1:0").unwrap(); listener.local_addr().unwrap().port() }; let mut tcp = Transport::::default().boxed(); let listener_id = ListenerId::next(); tcp.listen_on( listener_id, format!("/ip4/0.0.0.0/tcp/{port}").parse().unwrap(), ) .unwrap(); tcp.listen_on( ListenerId::next(), format!("/ip6/::/tcp/{port}").parse().unwrap(), ) .unwrap(); } #[cfg(feature = "async-io")] { async_std::task::block_on(async { test::(); }) } #[cfg(feature = "tokio")] { let rt = ::tokio::runtime::Builder::new_current_thread() .enable_io() .build() .unwrap(); rt.block_on(async { test::(); }); } } }