// Copyright 2018 Parity Technologies (UK) Ltd. // // Permission is hereby granted, free of charge, to any person obtaining a // copy of this software and associated documentation files (the "Software"), // to deal in the Software without restriction, including without limitation // the rights to use, copy, modify, merge, publish, distribute, sublicense, // and/or sell copies of the Software, and to permit persons to whom the // Software is furnished to do so, subject to the following conditions: // // The above copyright notice and this permission notice shall be included in // all copies or substantial portions of the Software. // // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS // OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, // FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE // AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER // LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING // FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER // DEALINGS IN THE SOFTWARE. use crate::behaviour::Mode; use crate::protocol::{ KadInStreamSink, KadOutStreamSink, KadPeer, KadRequestMsg, KadResponseMsg, ProtocolConfig, }; use crate::record::{self, Record}; use crate::QueryId; use either::Either; use futures::channel::oneshot; use futures::prelude::*; use futures::stream::SelectAll; use libp2p_core::{upgrade, ConnectedPoint}; use libp2p_identity::PeerId; use libp2p_swarm::handler::{ConnectionEvent, FullyNegotiatedInbound, FullyNegotiatedOutbound}; use libp2p_swarm::{ ConnectionHandler, ConnectionHandlerEvent, Stream, StreamUpgradeError, SubstreamProtocol, SupportedProtocols, }; use std::collections::VecDeque; use std::task::Waker; use std::time::Duration; use std::{error, fmt, io, marker::PhantomData, pin::Pin, task::Context, task::Poll}; const MAX_NUM_STREAMS: usize = 32; /// Protocol handler that manages substreams for the Kademlia protocol /// on a single connection with a peer. /// /// The handler will automatically open a Kademlia substream with the remote for each request we /// make. /// /// It also handles requests made by the remote. pub struct Handler { /// Configuration of the wire protocol. protocol_config: ProtocolConfig, /// In client mode, we don't accept inbound substreams. mode: Mode, /// Next unique ID of a connection. next_connec_unique_id: UniqueConnecId, /// List of active outbound streams. outbound_substreams: futures_bounded::FuturesTupleSet>, QueryId>, /// Contains one [`oneshot::Sender`] per outbound stream that we have requested. pending_streams: VecDeque, StreamUpgradeError>>>, /// List of outbound substreams that are waiting to become active next. /// Contains the request we want to send, and the user data if we expect an answer. pending_messages: VecDeque<(KadRequestMsg, QueryId)>, /// List of active inbound substreams with the state they are in. inbound_substreams: SelectAll, /// The connected endpoint of the connection that the handler /// is associated with. endpoint: ConnectedPoint, /// The [`PeerId`] of the remote. remote_peer_id: PeerId, /// The current state of protocol confirmation. protocol_status: Option, remote_supported_protocols: SupportedProtocols, } /// The states of protocol confirmation that a connection /// handler transitions through. #[derive(Debug, Copy, Clone, PartialEq)] struct ProtocolStatus { /// Whether the remote node supports one of our kademlia protocols. supported: bool, /// Whether we reported the state to the behaviour. reported: bool, } /// State of an active inbound substream. enum InboundSubstreamState { /// Waiting for a request from the remote. WaitingMessage { /// Whether it is the first message to be awaited on this stream. first: bool, connection_id: UniqueConnecId, substream: KadInStreamSink, }, /// Waiting for the behaviour to send a [`HandlerIn`] event containing the response. WaitingBehaviour(UniqueConnecId, KadInStreamSink, Option), /// Waiting to send an answer back to the remote. PendingSend(UniqueConnecId, KadInStreamSink, KadResponseMsg), /// Waiting to flush an answer back to the remote. PendingFlush(UniqueConnecId, KadInStreamSink), /// The substream is being closed. Closing(KadInStreamSink), /// The substream was cancelled in favor of a new one. Cancelled, Poisoned { phantom: PhantomData, }, } impl InboundSubstreamState { fn try_answer_with( &mut self, id: RequestId, msg: KadResponseMsg, ) -> Result<(), KadResponseMsg> { match std::mem::replace( self, InboundSubstreamState::Poisoned { phantom: PhantomData, }, ) { InboundSubstreamState::WaitingBehaviour(conn_id, substream, mut waker) if conn_id == id.connec_unique_id => { *self = InboundSubstreamState::PendingSend(conn_id, substream, msg); if let Some(waker) = waker.take() { waker.wake(); } Ok(()) } other => { *self = other; Err(msg) } } } fn close(&mut self) { match std::mem::replace( self, InboundSubstreamState::Poisoned { phantom: PhantomData, }, ) { InboundSubstreamState::WaitingMessage { substream, .. } | InboundSubstreamState::WaitingBehaviour(_, substream, _) | InboundSubstreamState::PendingSend(_, substream, _) | InboundSubstreamState::PendingFlush(_, substream) | InboundSubstreamState::Closing(substream) => { *self = InboundSubstreamState::Closing(substream); } InboundSubstreamState::Cancelled => { *self = InboundSubstreamState::Cancelled; } InboundSubstreamState::Poisoned { .. } => unreachable!(), } } } /// Event produced by the Kademlia handler. #[derive(Debug)] pub enum HandlerEvent { /// The configured protocol name has been confirmed by the peer through /// a successfully negotiated substream or by learning the supported protocols of the remote. ProtocolConfirmed { endpoint: ConnectedPoint }, /// The configured protocol name(s) are not or no longer supported by the peer on the provided /// connection and it should be removed from the routing table. ProtocolNotSupported { endpoint: ConnectedPoint }, /// Request for the list of nodes whose IDs are the closest to `key`. The number of nodes /// returned is not specified, but should be around 20. FindNodeReq { /// The key for which to locate the closest nodes. key: Vec, /// Identifier of the request. Needs to be passed back when answering. request_id: RequestId, }, /// Response to an `HandlerIn::FindNodeReq`. FindNodeRes { /// Results of the request. closer_peers: Vec, /// The user data passed to the `FindNodeReq`. query_id: QueryId, }, /// Same as `FindNodeReq`, but should also return the entries of the local providers list for /// this key. GetProvidersReq { /// The key for which providers are requested. key: record::Key, /// Identifier of the request. Needs to be passed back when answering. request_id: RequestId, }, /// Response to an `HandlerIn::GetProvidersReq`. GetProvidersRes { /// Nodes closest to the key. closer_peers: Vec, /// Known providers for this key. provider_peers: Vec, /// The user data passed to the `GetProvidersReq`. query_id: QueryId, }, /// An error happened when performing a query. QueryError { /// The error that happened. error: HandlerQueryErr, /// The user data passed to the query. query_id: QueryId, }, /// The peer announced itself as a provider of a key. AddProvider { /// The key for which the peer is a provider of the associated value. key: record::Key, /// The peer that is the provider of the value for `key`. provider: KadPeer, }, /// Request to get a value from the dht records GetRecord { /// Key for which we should look in the dht key: record::Key, /// Identifier of the request. Needs to be passed back when answering. request_id: RequestId, }, /// Response to a `HandlerIn::GetRecord`. GetRecordRes { /// The result is present if the key has been found record: Option, /// Nodes closest to the key. closer_peers: Vec, /// The user data passed to the `GetValue`. query_id: QueryId, }, /// Request to put a value in the dht records PutRecord { record: Record, /// Identifier of the request. Needs to be passed back when answering. request_id: RequestId, }, /// Response to a request to store a record. PutRecordRes { /// The key of the stored record. key: record::Key, /// The value of the stored record. value: Vec, /// The user data passed to the `PutValue`. query_id: QueryId, }, } /// Error that can happen when requesting an RPC query. #[derive(Debug)] pub enum HandlerQueryErr { /// Received an answer that doesn't correspond to the request. UnexpectedMessage, /// I/O error in the substream. Io(io::Error), } impl fmt::Display for HandlerQueryErr { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { match self { HandlerQueryErr::UnexpectedMessage => { write!( f, "Remote answered our Kademlia RPC query with the wrong message type" ) } HandlerQueryErr::Io(err) => { write!(f, "I/O error during a Kademlia RPC query: {err}") } } } } impl error::Error for HandlerQueryErr { fn source(&self) -> Option<&(dyn error::Error + 'static)> { match self { HandlerQueryErr::UnexpectedMessage => None, HandlerQueryErr::Io(err) => Some(err), } } } /// Event to send to the handler. #[derive(Debug)] pub enum HandlerIn { /// Resets the (sub)stream associated with the given request ID, /// thus signaling an error to the remote. /// /// Explicitly resetting the (sub)stream associated with a request /// can be used as an alternative to letting requests simply time /// out on the remote peer, thus potentially avoiding some delay /// for the query on the remote. Reset(RequestId), /// Change the connection to the specified mode. ReconfigureMode { new_mode: Mode }, /// Request for the list of nodes whose IDs are the closest to `key`. The number of nodes /// returned is not specified, but should be around 20. FindNodeReq { /// Identifier of the node. key: Vec, /// ID of the query that generated this request. query_id: QueryId, }, /// Response to a `FindNodeReq`. FindNodeRes { /// Results of the request. closer_peers: Vec, /// Identifier of the request that was made by the remote. /// /// It is a logic error to use an id of the handler of a different node. request_id: RequestId, }, /// Same as `FindNodeReq`, but should also return the entries of the local providers list for /// this key. GetProvidersReq { /// Identifier being searched. key: record::Key, /// ID of the query that generated this request. query_id: QueryId, }, /// Response to a `GetProvidersReq`. GetProvidersRes { /// Nodes closest to the key. closer_peers: Vec, /// Known providers for this key. provider_peers: Vec, /// Identifier of the request that was made by the remote. /// /// It is a logic error to use an id of the handler of a different node. request_id: RequestId, }, /// Indicates that this provider is known for this key. /// /// The API of the handler doesn't expose any event that allows you to know whether this /// succeeded. AddProvider { /// Key for which we should add providers. key: record::Key, /// Known provider for this key. provider: KadPeer, /// ID of the query that generated this request. query_id: QueryId, }, /// Request to retrieve a record from the DHT. GetRecord { /// The key of the record. key: record::Key, /// ID of the query that generated this request. query_id: QueryId, }, /// Response to a `GetRecord` request. GetRecordRes { /// The value that might have been found in our storage. record: Option, /// Nodes that are closer to the key we were searching for. closer_peers: Vec, /// Identifier of the request that was made by the remote. request_id: RequestId, }, /// Put a value into the dht records. PutRecord { record: Record, /// ID of the query that generated this request. query_id: QueryId, }, /// Response to a `PutRecord`. PutRecordRes { /// Key of the value that was put. key: record::Key, /// Value that was put. value: Vec, /// Identifier of the request that was made by the remote. request_id: RequestId, }, } /// Unique identifier for a request. Must be passed back in order to answer a request from /// the remote. #[derive(Debug, PartialEq, Eq, Copy, Clone)] pub struct RequestId { /// Unique identifier for an incoming connection. connec_unique_id: UniqueConnecId, } /// Unique identifier for a connection. #[derive(Debug, Copy, Clone, PartialEq, Eq)] struct UniqueConnecId(u64); impl Handler { pub fn new( protocol_config: ProtocolConfig, endpoint: ConnectedPoint, remote_peer_id: PeerId, mode: Mode, ) -> Self { match &endpoint { ConnectedPoint::Dialer { .. } => { tracing::debug!( peer=%remote_peer_id, mode=%mode, "New outbound connection" ); } ConnectedPoint::Listener { .. } => { tracing::debug!( peer=%remote_peer_id, mode=%mode, "New inbound connection" ); } } Handler { protocol_config, mode, endpoint, remote_peer_id, next_connec_unique_id: UniqueConnecId(0), inbound_substreams: Default::default(), outbound_substreams: futures_bounded::FuturesTupleSet::new( Duration::from_secs(10), MAX_NUM_STREAMS, ), pending_streams: Default::default(), pending_messages: Default::default(), protocol_status: None, remote_supported_protocols: Default::default(), } } fn on_fully_negotiated_outbound( &mut self, FullyNegotiatedOutbound { protocol: stream, info: (), }: FullyNegotiatedOutbound< ::OutboundProtocol, ::OutboundOpenInfo, >, ) { if let Some(sender) = self.pending_streams.pop_front() { let _ = sender.send(Ok(stream)); } if self.protocol_status.is_none() { // Upon the first successfully negotiated substream, we know that the // remote is configured with the same protocol name and we want // the behaviour to add this peer to the routing table, if possible. self.protocol_status = Some(ProtocolStatus { supported: true, reported: false, }); } } fn on_fully_negotiated_inbound( &mut self, FullyNegotiatedInbound { protocol, .. }: FullyNegotiatedInbound< ::InboundProtocol, ::InboundOpenInfo, >, ) { // If `self.allow_listening` is false, then we produced a `DeniedUpgrade` and `protocol` // is a `Void`. let protocol = match protocol { future::Either::Left(p) => p, future::Either::Right(p) => void::unreachable(p), }; if self.protocol_status.is_none() { // Upon the first successfully negotiated substream, we know that the // remote is configured with the same protocol name and we want // the behaviour to add this peer to the routing table, if possible. self.protocol_status = Some(ProtocolStatus { supported: true, reported: false, }); } if self.inbound_substreams.len() == MAX_NUM_STREAMS { if let Some(s) = self.inbound_substreams.iter_mut().find(|s| { matches!( s, // An inbound substream waiting to be reused. InboundSubstreamState::WaitingMessage { first: false, .. } ) }) { *s = InboundSubstreamState::Cancelled; tracing::debug!( peer=?self.remote_peer_id, "New inbound substream to peer exceeds inbound substream limit. \ Removed older substream waiting to be reused." ) } else { tracing::warn!( peer=?self.remote_peer_id, "New inbound substream to peer exceeds inbound substream limit. \ No older substream waiting to be reused. Dropping new substream." ); return; } } let connec_unique_id = self.next_connec_unique_id; self.next_connec_unique_id.0 += 1; self.inbound_substreams .push(InboundSubstreamState::WaitingMessage { first: true, connection_id: connec_unique_id, substream: protocol, }); } /// Takes the given [`KadRequestMsg`] and composes it into an outbound request-response protocol handshake using a [`oneshot::channel`]. fn queue_new_stream(&mut self, id: QueryId, msg: KadRequestMsg) { let (sender, receiver) = oneshot::channel(); self.pending_streams.push_back(sender); let result = self.outbound_substreams.try_push( async move { let mut stream = receiver .await .map_err(|_| io::Error::from(io::ErrorKind::BrokenPipe))? .map_err(|e| match e { StreamUpgradeError::Timeout => io::ErrorKind::TimedOut.into(), StreamUpgradeError::Apply(e) => e, StreamUpgradeError::NegotiationFailed => io::Error::new( io::ErrorKind::ConnectionRefused, "protocol not supported", ), StreamUpgradeError::Io(e) => e, })?; let has_answer = !matches!(msg, KadRequestMsg::AddProvider { .. }); stream.send(msg).await?; stream.close().await?; if !has_answer { return Ok(None); } let msg = stream.next().await.ok_or(io::ErrorKind::UnexpectedEof)??; Ok(Some(msg)) }, id, ); debug_assert!( result.is_ok(), "Expected to not create more streams than allowed" ); } } impl ConnectionHandler for Handler { type FromBehaviour = HandlerIn; type ToBehaviour = HandlerEvent; type InboundProtocol = Either; type OutboundProtocol = ProtocolConfig; type OutboundOpenInfo = (); type InboundOpenInfo = (); fn listen_protocol(&self) -> SubstreamProtocol { match self.mode { Mode::Server => SubstreamProtocol::new(Either::Left(self.protocol_config.clone()), ()), Mode::Client => SubstreamProtocol::new(Either::Right(upgrade::DeniedUpgrade), ()), } } fn on_behaviour_event(&mut self, message: HandlerIn) { match message { HandlerIn::Reset(request_id) => { if let Some(state) = self .inbound_substreams .iter_mut() .find(|state| match state { InboundSubstreamState::WaitingBehaviour(conn_id, _, _) => { conn_id == &request_id.connec_unique_id } _ => false, }) { state.close(); } } HandlerIn::FindNodeReq { key, query_id } => { let msg = KadRequestMsg::FindNode { key }; self.pending_messages.push_back((msg, query_id)); } HandlerIn::FindNodeRes { closer_peers, request_id, } => self.answer_pending_request(request_id, KadResponseMsg::FindNode { closer_peers }), HandlerIn::GetProvidersReq { key, query_id } => { let msg = KadRequestMsg::GetProviders { key }; self.pending_messages.push_back((msg, query_id)); } HandlerIn::GetProvidersRes { closer_peers, provider_peers, request_id, } => self.answer_pending_request( request_id, KadResponseMsg::GetProviders { closer_peers, provider_peers, }, ), HandlerIn::AddProvider { key, provider, query_id, } => { let msg = KadRequestMsg::AddProvider { key, provider }; self.pending_messages.push_back((msg, query_id)); } HandlerIn::GetRecord { key, query_id } => { let msg = KadRequestMsg::GetValue { key }; self.pending_messages.push_back((msg, query_id)); } HandlerIn::PutRecord { record, query_id } => { let msg = KadRequestMsg::PutValue { record }; self.pending_messages.push_back((msg, query_id)); } HandlerIn::GetRecordRes { record, closer_peers, request_id, } => { self.answer_pending_request( request_id, KadResponseMsg::GetValue { record, closer_peers, }, ); } HandlerIn::PutRecordRes { key, request_id, value, } => { self.answer_pending_request(request_id, KadResponseMsg::PutValue { key, value }); } HandlerIn::ReconfigureMode { new_mode } => { let peer = self.remote_peer_id; match &self.endpoint { ConnectedPoint::Dialer { .. } => { tracing::debug!( %peer, mode=%new_mode, "Changed mode on outbound connection" ) } ConnectedPoint::Listener { local_addr, .. } => { tracing::debug!( %peer, mode=%new_mode, local_address=%local_addr, "Changed mode on inbound connection assuming that one of our external addresses routes to the local address") } } self.mode = new_mode; } } } #[tracing::instrument(level = "trace", name = "ConnectionHandler::poll", skip(self, cx))] fn poll( &mut self, cx: &mut Context<'_>, ) -> Poll< ConnectionHandlerEvent, > { loop { match &mut self.protocol_status { Some(status) if !status.reported => { status.reported = true; let event = if status.supported { HandlerEvent::ProtocolConfirmed { endpoint: self.endpoint.clone(), } } else { HandlerEvent::ProtocolNotSupported { endpoint: self.endpoint.clone(), } }; return Poll::Ready(ConnectionHandlerEvent::NotifyBehaviour(event)); } _ => {} } match self.outbound_substreams.poll_unpin(cx) { Poll::Ready((Ok(Ok(Some(response))), query_id)) => { return Poll::Ready(ConnectionHandlerEvent::NotifyBehaviour( process_kad_response(response, query_id), )) } Poll::Ready((Ok(Ok(None)), _)) => { continue; } Poll::Ready((Ok(Err(e)), query_id)) => { return Poll::Ready(ConnectionHandlerEvent::NotifyBehaviour( HandlerEvent::QueryError { error: HandlerQueryErr::Io(e), query_id, }, )) } Poll::Ready((Err(_timeout), query_id)) => { return Poll::Ready(ConnectionHandlerEvent::NotifyBehaviour( HandlerEvent::QueryError { error: HandlerQueryErr::Io(io::ErrorKind::TimedOut.into()), query_id, }, )) } Poll::Pending => {} } if let Poll::Ready(Some(event)) = self.inbound_substreams.poll_next_unpin(cx) { return Poll::Ready(event); } if self.outbound_substreams.len() < MAX_NUM_STREAMS { if let Some((msg, id)) = self.pending_messages.pop_front() { self.queue_new_stream(id, msg); return Poll::Ready(ConnectionHandlerEvent::OutboundSubstreamRequest { protocol: SubstreamProtocol::new(self.protocol_config.clone(), ()), }); } } return Poll::Pending; } } fn on_connection_event( &mut self, event: ConnectionEvent< Self::InboundProtocol, Self::OutboundProtocol, Self::InboundOpenInfo, Self::OutboundOpenInfo, >, ) { match event { ConnectionEvent::FullyNegotiatedOutbound(fully_negotiated_outbound) => { self.on_fully_negotiated_outbound(fully_negotiated_outbound) } ConnectionEvent::FullyNegotiatedInbound(fully_negotiated_inbound) => { self.on_fully_negotiated_inbound(fully_negotiated_inbound) } ConnectionEvent::DialUpgradeError(ev) => { if let Some(sender) = self.pending_streams.pop_front() { let _ = sender.send(Err(ev.error)); } } ConnectionEvent::RemoteProtocolsChange(change) => { let dirty = self.remote_supported_protocols.on_protocols_change(change); if dirty { let remote_supports_our_kademlia_protocols = self .remote_supported_protocols .iter() .any(|p| self.protocol_config.protocol_names().contains(p)); self.protocol_status = Some(compute_new_protocol_status( remote_supports_our_kademlia_protocols, self.protocol_status, )) } } _ => {} } } } fn compute_new_protocol_status( now_supported: bool, current_status: Option, ) -> ProtocolStatus { let current_status = match current_status { None => { return ProtocolStatus { supported: now_supported, reported: false, } } Some(current) => current, }; if now_supported == current_status.supported { return ProtocolStatus { supported: now_supported, reported: true, }; } if now_supported { tracing::debug!("Remote now supports our kademlia protocol"); } else { tracing::debug!("Remote no longer supports our kademlia protocol"); } ProtocolStatus { supported: now_supported, reported: false, } } impl Handler { fn answer_pending_request(&mut self, request_id: RequestId, mut msg: KadResponseMsg) { for state in self.inbound_substreams.iter_mut() { match state.try_answer_with(request_id, msg) { Ok(()) => return, Err(m) => { msg = m; } } } debug_assert!(false, "Cannot find inbound substream for {request_id:?}") } } impl futures::Stream for InboundSubstreamState { type Item = ConnectionHandlerEvent; fn poll_next(self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll> { let this = self.get_mut(); loop { match std::mem::replace( this, Self::Poisoned { phantom: PhantomData, }, ) { InboundSubstreamState::WaitingMessage { first, connection_id, mut substream, } => match substream.poll_next_unpin(cx) { Poll::Ready(Some(Ok(KadRequestMsg::Ping))) => { tracing::warn!("Kademlia PING messages are unsupported"); *this = InboundSubstreamState::Closing(substream); } Poll::Ready(Some(Ok(KadRequestMsg::FindNode { key }))) => { *this = InboundSubstreamState::WaitingBehaviour(connection_id, substream, None); return Poll::Ready(Some(ConnectionHandlerEvent::NotifyBehaviour( HandlerEvent::FindNodeReq { key, request_id: RequestId { connec_unique_id: connection_id, }, }, ))); } Poll::Ready(Some(Ok(KadRequestMsg::GetProviders { key }))) => { *this = InboundSubstreamState::WaitingBehaviour(connection_id, substream, None); return Poll::Ready(Some(ConnectionHandlerEvent::NotifyBehaviour( HandlerEvent::GetProvidersReq { key, request_id: RequestId { connec_unique_id: connection_id, }, }, ))); } Poll::Ready(Some(Ok(KadRequestMsg::AddProvider { key, provider }))) => { *this = InboundSubstreamState::WaitingMessage { first: false, connection_id, substream, }; return Poll::Ready(Some(ConnectionHandlerEvent::NotifyBehaviour( HandlerEvent::AddProvider { key, provider }, ))); } Poll::Ready(Some(Ok(KadRequestMsg::GetValue { key }))) => { *this = InboundSubstreamState::WaitingBehaviour(connection_id, substream, None); return Poll::Ready(Some(ConnectionHandlerEvent::NotifyBehaviour( HandlerEvent::GetRecord { key, request_id: RequestId { connec_unique_id: connection_id, }, }, ))); } Poll::Ready(Some(Ok(KadRequestMsg::PutValue { record }))) => { *this = InboundSubstreamState::WaitingBehaviour(connection_id, substream, None); return Poll::Ready(Some(ConnectionHandlerEvent::NotifyBehaviour( HandlerEvent::PutRecord { record, request_id: RequestId { connec_unique_id: connection_id, }, }, ))); } Poll::Pending => { *this = InboundSubstreamState::WaitingMessage { first, connection_id, substream, }; return Poll::Pending; } Poll::Ready(None) => { return Poll::Ready(None); } Poll::Ready(Some(Err(e))) => { tracing::trace!("Inbound substream error: {:?}", e); return Poll::Ready(None); } }, InboundSubstreamState::WaitingBehaviour(id, substream, _) => { *this = InboundSubstreamState::WaitingBehaviour( id, substream, Some(cx.waker().clone()), ); return Poll::Pending; } InboundSubstreamState::PendingSend(id, mut substream, msg) => { match substream.poll_ready_unpin(cx) { Poll::Ready(Ok(())) => match substream.start_send_unpin(msg) { Ok(()) => { *this = InboundSubstreamState::PendingFlush(id, substream); } Err(_) => return Poll::Ready(None), }, Poll::Pending => { *this = InboundSubstreamState::PendingSend(id, substream, msg); return Poll::Pending; } Poll::Ready(Err(_)) => return Poll::Ready(None), } } InboundSubstreamState::PendingFlush(id, mut substream) => { match substream.poll_flush_unpin(cx) { Poll::Ready(Ok(())) => { *this = InboundSubstreamState::WaitingMessage { first: false, connection_id: id, substream, }; } Poll::Pending => { *this = InboundSubstreamState::PendingFlush(id, substream); return Poll::Pending; } Poll::Ready(Err(_)) => return Poll::Ready(None), } } InboundSubstreamState::Closing(mut stream) => match stream.poll_close_unpin(cx) { Poll::Ready(Ok(())) | Poll::Ready(Err(_)) => return Poll::Ready(None), Poll::Pending => { *this = InboundSubstreamState::Closing(stream); return Poll::Pending; } }, InboundSubstreamState::Poisoned { .. } => unreachable!(), InboundSubstreamState::Cancelled => return Poll::Ready(None), } } } } /// Process a Kademlia message that's supposed to be a response to one of our requests. fn process_kad_response(event: KadResponseMsg, query_id: QueryId) -> HandlerEvent { // TODO: must check that the response corresponds to the request match event { KadResponseMsg::Pong => { // We never send out pings. HandlerEvent::QueryError { error: HandlerQueryErr::UnexpectedMessage, query_id, } } KadResponseMsg::FindNode { closer_peers } => HandlerEvent::FindNodeRes { closer_peers, query_id, }, KadResponseMsg::GetProviders { closer_peers, provider_peers, } => HandlerEvent::GetProvidersRes { closer_peers, provider_peers, query_id, }, KadResponseMsg::GetValue { record, closer_peers, } => HandlerEvent::GetRecordRes { record, closer_peers, query_id, }, KadResponseMsg::PutValue { key, value, .. } => HandlerEvent::PutRecordRes { key, value, query_id, }, } } #[cfg(test)] mod tests { use super::*; use quickcheck::{Arbitrary, Gen}; use tracing_subscriber::EnvFilter; impl Arbitrary for ProtocolStatus { fn arbitrary(g: &mut Gen) -> Self { Self { supported: bool::arbitrary(g), reported: bool::arbitrary(g), } } } #[test] fn compute_next_protocol_status_test() { let _ = tracing_subscriber::fmt() .with_env_filter(EnvFilter::from_default_env()) .try_init(); fn prop(now_supported: bool, current: Option) { let new = compute_new_protocol_status(now_supported, current); match current { None => { assert!(!new.reported); assert_eq!(new.supported, now_supported); } Some(current) => { if current.supported == now_supported { assert!(new.reported); } else { assert!(!new.reported); } assert_eq!(new.supported, now_supported); } } } quickcheck::quickcheck(prop as fn(_, _)) } }