mirror of
https://github.com/fluencelabs/rust-libp2p
synced 2025-04-28 20:22:16 +00:00
# Conflicts: # Cargo.toml # core/Cargo.toml # examples/distributed-key-value-store.rs # misc/multiaddr/Cargo.toml # misc/multiaddr/src/onion_addr.rs # misc/multistream-select/Cargo.toml # muxers/mplex/Cargo.toml # muxers/yamux/Cargo.toml # protocols/floodsub/Cargo.toml # protocols/gossipsub/Cargo.toml # protocols/identify/Cargo.toml # protocols/kad/Cargo.toml # protocols/kad/src/addresses.rs # protocols/kad/src/behaviour.rs # protocols/kad/src/behaviour/test.rs # protocols/kad/src/kbucket/bucket.rs # protocols/kad/src/kbucket/key.rs # protocols/mdns/Cargo.toml # protocols/ping/Cargo.toml # protocols/request-response/Cargo.toml # protocols/secio/Cargo.toml # swarm-derive/Cargo.toml # swarm/Cargo.toml # transports/deflate/Cargo.toml # transports/dns/Cargo.toml # transports/noise/Cargo.toml # transports/plaintext/Cargo.toml # transports/tcp/Cargo.toml # transports/uds/Cargo.toml # transports/wasm-ext/Cargo.toml # transports/websocket/Cargo.toml
613 lines
22 KiB
Rust
613 lines
22 KiB
Rust
// Copyright 2019 Parity Technologies (UK) Ltd.
|
|
//
|
|
// Permission is hereby granted, free of charge, to any person obtaining a
|
|
// copy of this software and associated documentation files (the "Software"),
|
|
// to deal in the Software without restriction, including without limitation
|
|
// the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
// and/or sell copies of the Software, and to permit persons to whom the
|
|
// Software is furnished to do so, subject to the following conditions:
|
|
//
|
|
// The above copyright notice and this permission notice shall be included in
|
|
// all copies or substantial portions of the Software.
|
|
//
|
|
// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
|
|
// OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
|
// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
// FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
// DEALINGS IN THE SOFTWARE.
|
|
|
|
mod peers;
|
|
|
|
use peers::PeersIterState;
|
|
use peers::closest::{ClosestPeersIterConfig, ClosestPeersIter, disjoint::ClosestDisjointPeersIter};
|
|
use peers::fixed::FixedPeersIter;
|
|
|
|
use crate::{ALPHA_VALUE, K_VALUE};
|
|
use crate::kbucket::{Key, KeyBytes};
|
|
use either::Either;
|
|
use fnv::FnvHashMap;
|
|
use libp2p_core::PeerId;
|
|
use std::{time::Duration, num::NonZeroUsize};
|
|
use wasm_timer::Instant;
|
|
|
|
/// Peer along with its weight
|
|
pub struct WeightedPeer {
|
|
/// Kademlia key & id of the peer
|
|
pub peer_id: Key<PeerId>,
|
|
/// Weight, calculated locally
|
|
pub weight: u32
|
|
}
|
|
|
|
/// A `QueryPool` provides an aggregate state machine for driving `Query`s to completion.
|
|
///
|
|
/// Internally, a `Query` is in turn driven by an underlying `QueryPeerIter`
|
|
/// that determines the peer selection strategy, i.e. the order in which the
|
|
/// peers involved in the query should be contacted.
|
|
pub struct QueryPool<TInner> {
|
|
next_id: usize,
|
|
config: QueryConfig,
|
|
queries: FnvHashMap<QueryId, Query<TInner>>,
|
|
}
|
|
|
|
/// The observable states emitted by [`QueryPool::poll`].
|
|
pub enum QueryPoolState<'a, TInner> {
|
|
/// The pool is idle, i.e. there are no queries to process.
|
|
Idle,
|
|
/// At least one query is waiting for results. `Some(request)` indicates
|
|
/// that a new request is now being waited on.
|
|
Waiting(Option<(&'a mut Query<TInner>, PeerId)>),
|
|
/// A query has finished.
|
|
Finished(Query<TInner>),
|
|
/// A query has timed out.
|
|
Timeout(Query<TInner>)
|
|
}
|
|
|
|
impl<TInner> QueryPool<TInner> {
|
|
/// Creates a new `QueryPool` with the given configuration.
|
|
pub fn new(config: QueryConfig) -> Self {
|
|
QueryPool {
|
|
next_id: 0,
|
|
config,
|
|
queries: Default::default()
|
|
}
|
|
}
|
|
|
|
/// Gets a reference to the `QueryConfig` used by the pool.
|
|
pub fn config(&self) -> &QueryConfig {
|
|
&self.config
|
|
}
|
|
|
|
/// Returns an iterator over the queries in the pool.
|
|
pub fn iter(&self) -> impl Iterator<Item = &Query<TInner>> {
|
|
self.queries.values()
|
|
}
|
|
|
|
/// Gets the current size of the pool, i.e. the number of running queries.
|
|
pub fn size(&self) -> usize {
|
|
self.queries.len()
|
|
}
|
|
|
|
/// Returns an iterator that allows modifying each query in the pool.
|
|
pub fn iter_mut(&mut self) -> impl Iterator<Item = &mut Query<TInner>> {
|
|
self.queries.values_mut()
|
|
}
|
|
|
|
/// Adds a query to the pool that contacts a fixed set of peers.
|
|
pub fn add_fixed<I>(&mut self, peers: I, inner: TInner) -> QueryId
|
|
where
|
|
I: IntoIterator<Item = WeightedPeer>
|
|
{
|
|
let id = self.next_query_id();
|
|
self.continue_fixed(id, peers, inner);
|
|
id
|
|
}
|
|
|
|
/// Continues an earlier query with a fixed set of peers, reusing
|
|
/// the given query ID, which must be from a query that finished
|
|
/// earlier.
|
|
pub fn continue_fixed<I>(&mut self, id: QueryId, peers: I, inner: TInner)
|
|
where
|
|
I: IntoIterator<Item = WeightedPeer>
|
|
{
|
|
assert!(!self.queries.contains_key(&id));
|
|
// TODO: why not alpha?
|
|
let parallelism = self.config.replication_factor;
|
|
|
|
let (swamp, weighted) = peers.into_iter().partition::<Vec<_>, _>(|p| p.weight == 0);
|
|
let swamp = swamp.into_iter().map(|p| p.peer_id.into_preimage());
|
|
let weighted = weighted.into_iter().map(|p| p.peer_id.into_preimage());
|
|
|
|
let weighted_iter = QueryPeerIter::Fixed(FixedPeersIter::new(weighted, parallelism));
|
|
let swamp_iter = QueryPeerIter::Fixed(FixedPeersIter::new(swamp, parallelism));
|
|
let query = Query::new(id, weighted_iter, swamp_iter, inner);
|
|
self.queries.insert(id, query);
|
|
}
|
|
|
|
/// Adds a query to the pool that iterates towards the closest peers to the target.
|
|
pub fn add_iter_closest<T, I>(&mut self, target: T, peers: I, inner: TInner) -> QueryId
|
|
where
|
|
T: Into<KeyBytes> + Clone,
|
|
I: IntoIterator<Item = WeightedPeer>
|
|
{
|
|
let id = self.next_query_id();
|
|
self.continue_iter_closest(id, target, peers, inner);
|
|
id
|
|
}
|
|
|
|
/// Adds a query to the pool that iterates towards the closest peers to the target.
|
|
pub fn continue_iter_closest<T, I>(&mut self, id: QueryId, target: T, peers: I, inner: TInner)
|
|
where
|
|
T: Into<KeyBytes> + Clone,
|
|
I: IntoIterator<Item = WeightedPeer>
|
|
{
|
|
let cfg = ClosestPeersIterConfig {
|
|
num_results: self.config.replication_factor,
|
|
parallelism: self.config.parallelism,
|
|
.. ClosestPeersIterConfig::default()
|
|
};
|
|
|
|
let (swamp, weighted) = peers.into_iter().partition::<Vec<_>, _>(|p| p.weight == 0);
|
|
let swamp = swamp.into_iter().map(|p| p.peer_id);
|
|
let weighted = weighted.into_iter().map(|p| p.peer_id);
|
|
|
|
let weighted_iter = if self.config.disjoint_query_paths {
|
|
QueryPeerIter::ClosestDisjoint(
|
|
ClosestDisjointPeersIter::with_config(cfg.clone(), target.clone(), weighted),
|
|
)
|
|
} else {
|
|
QueryPeerIter::Closest(
|
|
ClosestPeersIter::with_config(cfg.clone(), target.clone(), weighted)
|
|
)
|
|
};
|
|
|
|
let swamp_iter = if self.config.disjoint_query_paths {
|
|
QueryPeerIter::ClosestDisjoint(
|
|
ClosestDisjointPeersIter::with_config(cfg, target, swamp),
|
|
)
|
|
} else {
|
|
QueryPeerIter::Closest(
|
|
ClosestPeersIter::with_config(cfg, target, swamp)
|
|
)
|
|
};
|
|
|
|
let query = Query::new(id, weighted_iter, swamp_iter, inner);
|
|
self.queries.insert(id, query);
|
|
}
|
|
|
|
fn next_query_id(&mut self) -> QueryId {
|
|
let id = QueryId(self.next_id);
|
|
self.next_id = self.next_id.wrapping_add(1);
|
|
id
|
|
}
|
|
|
|
/// Returns a reference to a query with the given ID, if it is in the pool.
|
|
pub fn get(&self, id: &QueryId) -> Option<&Query<TInner>> {
|
|
self.queries.get(id)
|
|
}
|
|
|
|
/// Returns a mutablereference to a query with the given ID, if it is in the pool.
|
|
pub fn get_mut(&mut self, id: &QueryId) -> Option<&mut Query<TInner>> {
|
|
self.queries.get_mut(id)
|
|
}
|
|
|
|
/// Polls the pool to advance the queries.
|
|
pub fn poll(&mut self, now: Instant) -> QueryPoolState<'_, TInner> {
|
|
let mut finished = None;
|
|
let mut timeout = None;
|
|
let mut waiting = None;
|
|
|
|
for (&query_id, query) in self.queries.iter_mut() {
|
|
query.stats.start = query.stats.start.or(Some(now));
|
|
match query.next(now) {
|
|
PeersIterState::Finished => {
|
|
finished = Some(query_id);
|
|
break
|
|
}
|
|
PeersIterState::Waiting(Some(peer_id)) => {
|
|
let peer = peer_id.into_owned();
|
|
waiting = Some((query_id, peer));
|
|
break
|
|
}
|
|
PeersIterState::Waiting(None) | PeersIterState::WaitingAtCapacity => {
|
|
let elapsed = now - query.stats.start.unwrap_or(now);
|
|
if elapsed >= self.config.timeout {
|
|
timeout = Some(query_id);
|
|
break
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if let Some((query_id, peer_id)) = waiting {
|
|
let query = self.queries.get_mut(&query_id).expect("s.a.");
|
|
return QueryPoolState::Waiting(Some((query, peer_id)))
|
|
}
|
|
|
|
if let Some(query_id) = finished {
|
|
let mut query = self.queries.remove(&query_id).expect("s.a.");
|
|
query.stats.end = Some(now);
|
|
return QueryPoolState::Finished(query)
|
|
}
|
|
|
|
if let Some(query_id) = timeout {
|
|
let mut query = self.queries.remove(&query_id).expect("s.a.");
|
|
query.stats.end = Some(now);
|
|
return QueryPoolState::Timeout(query)
|
|
}
|
|
|
|
if self.queries.is_empty() {
|
|
QueryPoolState::Idle
|
|
} else {
|
|
QueryPoolState::Waiting(None)
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Unique identifier for an active query.
|
|
#[derive(Debug, Copy, Clone, Hash, PartialEq, Eq)]
|
|
pub struct QueryId(usize);
|
|
|
|
/// The configuration for queries in a `QueryPool`.
|
|
#[derive(Debug, Clone)]
|
|
pub struct QueryConfig {
|
|
/// Timeout of a single query.
|
|
///
|
|
/// See [`crate::behaviour::KademliaConfig::set_query_timeout`] for details.
|
|
pub timeout: Duration,
|
|
|
|
/// The replication factor to use.
|
|
///
|
|
/// See [`crate::behaviour::KademliaConfig::set_replication_factor`] for details.
|
|
pub replication_factor: NonZeroUsize,
|
|
|
|
/// Allowed level of parallelism for iterative queries.
|
|
///
|
|
/// See [`crate::behaviour::KademliaConfig::set_parallelism`] for details.
|
|
pub parallelism: NonZeroUsize,
|
|
|
|
/// Whether to use disjoint paths on iterative lookups.
|
|
///
|
|
/// See [`crate::behaviour::KademliaConfig::disjoint_query_paths`] for details.
|
|
pub disjoint_query_paths: bool,
|
|
}
|
|
|
|
impl Default for QueryConfig {
|
|
fn default() -> Self {
|
|
QueryConfig {
|
|
timeout: Duration::from_secs(60),
|
|
replication_factor: NonZeroUsize::new(K_VALUE.get()).expect("K_VALUE > 0"),
|
|
parallelism: ALPHA_VALUE,
|
|
disjoint_query_paths: false,
|
|
}
|
|
}
|
|
}
|
|
|
|
/// A query in a `QueryPool`.
|
|
pub struct Query<TInner> {
|
|
/// The unique ID of the query.
|
|
id: QueryId,
|
|
/// The peer iterator that drives the query state.
|
|
weighted_iter: QueryPeerIter,
|
|
/// The peer iterator that drives the query state.
|
|
swamp_iter: QueryPeerIter,
|
|
/// Execution statistics of the query.
|
|
stats: QueryStats,
|
|
/// The opaque inner query state.
|
|
pub inner: TInner,
|
|
}
|
|
|
|
/// The peer selection strategies that can be used by queries.
|
|
enum QueryPeerIter {
|
|
Closest(ClosestPeersIter),
|
|
ClosestDisjoint(ClosestDisjointPeersIter),
|
|
Fixed(FixedPeersIter)
|
|
}
|
|
|
|
impl<TInner> Query<TInner> {
|
|
/// Creates a new query without starting it.
|
|
fn new(id: QueryId, weighted_iter: QueryPeerIter, swamp_iter: QueryPeerIter, inner: TInner) -> Self {
|
|
Query { id, inner, weighted_iter, swamp_iter, stats: QueryStats::empty() }
|
|
}
|
|
|
|
/// Gets the unique ID of the query.
|
|
pub fn id(&self) -> QueryId {
|
|
self.id
|
|
}
|
|
|
|
/// Gets the current execution statistics of the query.
|
|
pub fn stats(&self) -> &QueryStats {
|
|
&self.stats
|
|
}
|
|
|
|
/// Informs the query that the attempt to contact `peer` failed.
|
|
pub fn on_failure(&mut self, peer: &PeerId) {
|
|
let updated_swamp = match &mut self.swamp_iter {
|
|
QueryPeerIter::Closest(iter) => iter.on_failure(peer),
|
|
QueryPeerIter::Fixed(iter) => iter.on_failure(peer),
|
|
QueryPeerIter::ClosestDisjoint(iter) => iter.on_failure(peer),
|
|
};
|
|
|
|
let updated_weighted = match &mut self.weighted_iter {
|
|
QueryPeerIter::Closest(iter) => iter.on_failure(peer),
|
|
QueryPeerIter::ClosestDisjoint(iter) => iter.on_failure(peer),
|
|
QueryPeerIter::Fixed(iter) => iter.on_failure(peer),
|
|
};
|
|
|
|
if updated_swamp || updated_weighted {
|
|
self.stats.failure += 1;
|
|
}
|
|
}
|
|
|
|
/// Informs the query that the attempt to contact `peer` succeeded,
|
|
/// possibly resulting in new peers that should be incorporated into
|
|
/// the query, if applicable.
|
|
pub fn on_success<I>(&mut self, peer: &PeerId, new_peers: I)
|
|
where
|
|
I: IntoIterator<Item = WeightedPeer>
|
|
{
|
|
let (swamp, weighted) = new_peers.into_iter().partition::<Vec<_>, _>(|p| p.weight == 0);
|
|
let swamp = swamp.into_iter().map(|p| p.peer_id);
|
|
let weighted = weighted.into_iter().map(|p| p.peer_id);
|
|
|
|
|
|
let updated_swamp = match &mut self.swamp_iter {
|
|
QueryPeerIter::Closest(iter) => iter.on_success(peer, swamp),
|
|
QueryPeerIter::ClosestDisjoint(iter) => iter.on_success(peer, swamp),
|
|
QueryPeerIter::Fixed(iter) => iter.on_success(peer),
|
|
};
|
|
|
|
let updated_weighted = match &mut self.weighted_iter {
|
|
QueryPeerIter::Closest(iter) => iter.on_success(peer, weighted),
|
|
QueryPeerIter::ClosestDisjoint(iter) => iter.on_success(peer, weighted),
|
|
QueryPeerIter::Fixed(iter) => iter.on_success(peer),
|
|
};
|
|
|
|
if updated_swamp || updated_weighted {
|
|
self.stats.success += 1;
|
|
}
|
|
}
|
|
|
|
/// Checks whether the query is currently waiting for a result from `peer`.
|
|
pub fn is_waiting(&self, peer: &PeerId) -> bool {
|
|
let weighted_waiting = match &self.weighted_iter {
|
|
QueryPeerIter::Closest(iter) => iter.is_waiting(peer),
|
|
QueryPeerIter::ClosestDisjoint(iter) => iter.is_waiting(peer),
|
|
QueryPeerIter::Fixed(iter) => iter.is_waiting(peer)
|
|
};
|
|
|
|
let swamp_waiting = match &self.swamp_iter {
|
|
QueryPeerIter::Closest(iter) => iter.is_waiting(peer),
|
|
QueryPeerIter::ClosestDisjoint(iter) => iter.is_waiting(peer),
|
|
QueryPeerIter::Fixed(iter) => iter.is_waiting(peer)
|
|
};
|
|
|
|
debug_assert_ne!(weighted_waiting, swamp_waiting);
|
|
|
|
weighted_waiting || swamp_waiting
|
|
}
|
|
|
|
/// Advances the state of the underlying peer iterator.
|
|
fn next(&mut self, now: Instant) -> PeersIterState<'_> {
|
|
use PeersIterState::*;
|
|
|
|
// First query weighted iter
|
|
let weighted_state = match &mut self.weighted_iter {
|
|
QueryPeerIter::Closest(iter) => iter.next(now),
|
|
QueryPeerIter::ClosestDisjoint(iter) => iter.next(now),
|
|
QueryPeerIter::Fixed(iter) => iter.next()
|
|
};
|
|
|
|
// If there's a new weighted peer to send rpc to, return it
|
|
if let Waiting(Some(_)) = weighted_state {
|
|
self.stats.requests += 1;
|
|
return weighted_state;
|
|
}
|
|
|
|
// If there was no new weighted peer, check swamp
|
|
let swamp_state = match &mut self.swamp_iter {
|
|
QueryPeerIter::Closest(iter) => iter.next(now),
|
|
QueryPeerIter::ClosestDisjoint(iter) => iter.next(now),
|
|
QueryPeerIter::Fixed(iter) => iter.next()
|
|
};
|
|
|
|
// If there's a new swamp peer to send rpc to, return it
|
|
if let Waiting(Some(_)) = swamp_state {
|
|
self.stats.requests += 1;
|
|
return swamp_state;
|
|
}
|
|
|
|
// Return remaining state: weighted has higher priority
|
|
match (weighted_state, swamp_state) {
|
|
// If weighted finished, return swamp state
|
|
(Finished, swamp) => swamp,
|
|
// Otherwise, return weighted state first
|
|
(weighted, _) => weighted
|
|
}
|
|
}
|
|
|
|
/// Tries to (gracefully) finish the query prematurely, providing the peers
|
|
/// that are no longer of interest for further progress of the query.
|
|
///
|
|
/// A query may require that in order to finish gracefully a certain subset
|
|
/// of peers must be contacted. E.g. in the case of disjoint query paths a
|
|
/// query may only finish gracefully if every path contacted a peer whose
|
|
/// response permits termination of the query. The given peers are those for
|
|
/// which this is considered to be the case, i.e. for which a termination
|
|
/// condition is satisfied.
|
|
///
|
|
/// Returns `true` if the query did indeed finish, `false` otherwise. In the
|
|
/// latter case, a new attempt at finishing the query may be made with new
|
|
/// `peers`.
|
|
///
|
|
/// A finished query immediately stops yielding new peers to contact and
|
|
/// will be reported by [`QueryPool::poll`] via
|
|
/// [`QueryPoolState::Finished`].
|
|
pub fn try_finish<'a, I>(&mut self, peers: I) -> bool
|
|
where
|
|
I: IntoIterator<Item = &'a PeerId> + Clone
|
|
{
|
|
let weighted = match &mut self.weighted_iter {
|
|
QueryPeerIter::Closest(iter) => { iter.finish(); true },
|
|
QueryPeerIter::ClosestDisjoint(iter) => iter.finish_paths(peers.clone()),
|
|
QueryPeerIter::Fixed(iter) => { iter.finish(); true },
|
|
};
|
|
|
|
let swamp = match &mut self.swamp_iter {
|
|
QueryPeerIter::Closest(iter) => { iter.finish(); true },
|
|
QueryPeerIter::ClosestDisjoint(iter) => iter.finish_paths(peers),
|
|
QueryPeerIter::Fixed(iter) => { iter.finish(); true },
|
|
};
|
|
|
|
weighted && swamp
|
|
}
|
|
|
|
/// Finishes the query prematurely.
|
|
///
|
|
/// A finished query immediately stops yielding new peers to contact and will be
|
|
/// reported by [`QueryPool::poll`] via [`QueryPoolState::Finished`].
|
|
pub fn finish(&mut self) {
|
|
match &mut self.weighted_iter {
|
|
QueryPeerIter::Closest(iter) => iter.finish(),
|
|
QueryPeerIter::ClosestDisjoint(iter) => iter.finish(),
|
|
QueryPeerIter::Fixed(iter) => iter.finish()
|
|
};
|
|
|
|
match &mut self.swamp_iter {
|
|
QueryPeerIter::Closest(iter) => iter.finish(),
|
|
QueryPeerIter::ClosestDisjoint(iter) => iter.finish(),
|
|
QueryPeerIter::Fixed(iter) => iter.finish()
|
|
};
|
|
}
|
|
|
|
/// Checks whether the query has finished.
|
|
///
|
|
/// A finished query is eventually reported by `QueryPool::next()` and
|
|
/// removed from the pool.
|
|
pub fn is_finished(&self) -> bool {
|
|
let weighted_finished = match &self.weighted_iter {
|
|
QueryPeerIter::Closest(iter) => iter.is_finished(),
|
|
QueryPeerIter::ClosestDisjoint(iter) => iter.is_finished(),
|
|
QueryPeerIter::Fixed(iter) => iter.is_finished()
|
|
};
|
|
|
|
let swamp_finished = match &self.swamp_iter {
|
|
QueryPeerIter::Closest(iter) => iter.is_finished(),
|
|
QueryPeerIter::ClosestDisjoint(iter) => iter.is_finished(),
|
|
QueryPeerIter::Fixed(iter) => iter.is_finished()
|
|
};
|
|
|
|
weighted_finished && swamp_finished
|
|
}
|
|
|
|
/// Consumes the query, producing the final `QueryResult`.
|
|
pub fn into_result(self) -> QueryResult<TInner, impl Iterator<Item = PeerId>> {
|
|
fn into_result_(iter: QueryPeerIter) -> impl Iterator<Item = PeerId> {
|
|
match iter {
|
|
QueryPeerIter::Closest(iter) => Either::Left(Either::Left(iter.into_result())),
|
|
QueryPeerIter::ClosestDisjoint(iter) => Either::Left(Either::Right(iter.into_result())),
|
|
QueryPeerIter::Fixed(iter) => Either::Right(iter.into_result())
|
|
}
|
|
}
|
|
|
|
let weighted = into_result_(self.weighted_iter);
|
|
let swamp = into_result_(self.swamp_iter);
|
|
let peers = weighted.chain(swamp);
|
|
QueryResult { peers, inner: self.inner, stats: self.stats }
|
|
}
|
|
}
|
|
|
|
/// The result of a `Query`.
|
|
pub struct QueryResult<TInner, TPeers> {
|
|
/// The opaque inner query state.
|
|
pub inner: TInner,
|
|
/// The successfully contacted peers.
|
|
pub peers: TPeers,
|
|
/// The collected query statistics.
|
|
pub stats: QueryStats
|
|
}
|
|
|
|
/// Execution statistics of a query.
|
|
#[derive(Clone, Debug, PartialEq, Eq)]
|
|
pub struct QueryStats {
|
|
requests: u32,
|
|
success: u32,
|
|
failure: u32,
|
|
start: Option<Instant>,
|
|
end: Option<Instant>
|
|
}
|
|
|
|
impl QueryStats {
|
|
pub fn empty() -> Self {
|
|
QueryStats {
|
|
requests: 0,
|
|
success: 0,
|
|
failure: 0,
|
|
start: None,
|
|
end: None,
|
|
}
|
|
}
|
|
|
|
/// Gets the total number of requests initiated by the query.
|
|
pub fn num_requests(&self) -> u32 {
|
|
self.requests
|
|
}
|
|
|
|
/// Gets the number of successful requests.
|
|
pub fn num_successes(&self) -> u32 {
|
|
self.success
|
|
}
|
|
|
|
/// Gets the number of failed requests.
|
|
pub fn num_failures(&self) -> u32 {
|
|
self.failure
|
|
}
|
|
|
|
/// Gets the number of pending requests.
|
|
///
|
|
/// > **Note**: A query can finish while still having pending
|
|
/// > requests, if the termination conditions are already met.
|
|
pub fn num_pending(&self) -> u32 {
|
|
self.requests - (self.success + self.failure)
|
|
}
|
|
|
|
/// Gets the duration of the query.
|
|
///
|
|
/// If the query has not yet finished, the duration is measured from the
|
|
/// start of the query to the current instant.
|
|
///
|
|
/// If the query did not yet start (i.e. yield the first peer to contact),
|
|
/// `None` is returned.
|
|
pub fn duration(&self) -> Option<Duration> {
|
|
if let Some(s) = self.start {
|
|
if let Some(e) = self.end {
|
|
Some(e - s)
|
|
} else {
|
|
Some(Instant::now() - s)
|
|
}
|
|
} else {
|
|
None
|
|
}
|
|
}
|
|
|
|
/// Merges these stats with the given stats of another query,
|
|
/// e.g. to accumulate statistics from a multi-phase query.
|
|
///
|
|
/// Counters are merged cumulatively while the instants for
|
|
/// start and end of the queries are taken as the minimum and
|
|
/// maximum, respectively.
|
|
pub fn merge(self, other: QueryStats) -> Self {
|
|
QueryStats {
|
|
requests: self.requests + other.requests,
|
|
success: self.success + other.success,
|
|
failure: self.failure + other.failure,
|
|
start: match (self.start, other.start) {
|
|
(Some(a), Some(b)) => Some(std::cmp::min(a, b)),
|
|
(a, b) => a.or(b)
|
|
},
|
|
end: std::cmp::max(self.end, other.end)
|
|
}
|
|
}
|
|
}
|