596 lines
15 KiB
Go
Raw Normal View History

2014-11-20 10:46:19 -08:00
// Package dht implements a distributed hash table that satisfies the ipfs routing
// interface. This DHT is modeled after kademlia with Coral and S/Kademlia modifications.
2014-07-23 04:48:30 -07:00
package dht
import (
2014-09-19 08:07:56 -07:00
"bytes"
"crypto/rand"
2014-09-16 00:56:40 -07:00
"errors"
"fmt"
"sync"
"time"
2014-09-14 04:52:08 -07:00
inet "github.com/jbenet/go-ipfs/net"
2014-09-16 00:56:40 -07:00
msg "github.com/jbenet/go-ipfs/net/message"
peer "github.com/jbenet/go-ipfs/peer"
routing "github.com/jbenet/go-ipfs/routing"
pb "github.com/jbenet/go-ipfs/routing/dht/pb"
2014-08-09 22:28:46 -07:00
kb "github.com/jbenet/go-ipfs/routing/kbucket"
u "github.com/jbenet/go-ipfs/util"
2014-10-25 07:12:01 -07:00
ctxc "github.com/jbenet/go-ipfs/util/ctxcloser"
"github.com/jbenet/go-ipfs/util/eventlog"
2014-09-16 00:56:40 -07:00
context "github.com/jbenet/go-ipfs/Godeps/_workspace/src/code.google.com/p/go.net/context"
2014-10-21 15:10:58 -07:00
ds "github.com/jbenet/go-ipfs/Godeps/_workspace/src/github.com/jbenet/go-datastore"
2014-07-30 20:16:34 -07:00
"github.com/jbenet/go-ipfs/Godeps/_workspace/src/code.google.com/p/goprotobuf/proto"
)
var log = eventlog.Logger("dht")
const doPinging = false
2014-07-23 04:48:30 -07:00
// TODO. SEE https://github.com/jbenet/node-ipfs/blob/master/submodules/ipfs-dht/index.js
// IpfsDHT is an implementation of Kademlia with Coral and S/Kademlia modifications.
// It is used to implement the base IpfsRouting module.
type IpfsDHT struct {
// Array of routing tables for differently distanced nodes
// NOTE: (currently, only a single table is used)
2014-08-16 23:03:36 -07:00
routingTables []*kb.RoutingTable
// the network services we need
dialer inet.Dialer
sender inet.Sender
2014-07-30 20:16:34 -07:00
// Local peer (yourself)
self peer.Peer
2014-07-30 20:16:34 -07:00
2014-09-16 06:18:26 -07:00
// Other peers
peerstore peer.Peerstore
2014-07-30 20:16:34 -07:00
// Local data
datastore ds.Datastore
dslock sync.Mutex
providers *ProviderManager
// When this peer started up
birth time.Time
//lock to make diagnostics work better
diaglock sync.Mutex
// record validator funcs
Validators map[string]ValidatorFunc
2014-10-25 07:12:01 -07:00
ctxc.ContextCloser
}
2014-08-09 22:28:46 -07:00
// NewDHT creates a new DHT object with the given peer as the 'local' host
func NewDHT(ctx context.Context, p peer.Peer, ps peer.Peerstore, dialer inet.Dialer, sender inet.Sender, dstore ds.Datastore) *IpfsDHT {
dht := new(IpfsDHT)
dht.dialer = dialer
2014-09-16 00:56:40 -07:00
dht.sender = sender
2014-08-26 14:24:51 -07:00
dht.datastore = dstore
dht.self = p
2014-09-16 06:18:26 -07:00
dht.peerstore = ps
2014-10-25 07:12:01 -07:00
dht.ContextCloser = ctxc.NewContextCloser(ctx, nil)
2014-09-16 00:56:40 -07:00
2014-10-25 07:12:01 -07:00
dht.providers = NewProviderManager(dht.Context(), p.ID())
dht.AddCloserChild(dht.providers)
2014-08-16 23:03:36 -07:00
dht.routingTables = make([]*kb.RoutingTable, 3)
dht.routingTables[0] = kb.NewRoutingTable(20, kb.ConvertPeerID(p.ID()), time.Millisecond*1000)
dht.routingTables[1] = kb.NewRoutingTable(20, kb.ConvertPeerID(p.ID()), time.Millisecond*1000)
dht.routingTables[2] = kb.NewRoutingTable(20, kb.ConvertPeerID(p.ID()), time.Hour)
dht.birth = time.Now()
dht.Validators = make(map[string]ValidatorFunc)
dht.Validators["pk"] = ValidatePublicKeyRecord
if doPinging {
2014-10-25 07:12:01 -07:00
dht.Children().Add(1)
go dht.PingRoutine(time.Second * 10)
}
2014-08-10 21:40:17 -07:00
return dht
}
// Connect to a new peer at the given address, ping and add to the routing table
func (dht *IpfsDHT) Connect(ctx context.Context, npeer peer.Peer) error {
2014-11-05 04:26:30 -08:00
err := dht.dialer.DialPeer(ctx, npeer)
if err != nil {
return err
}
2014-08-06 10:02:53 -07:00
// Ping new peer to register in their routing table
// NOTE: this should be done better...
err = dht.Ping(ctx, npeer)
2014-08-06 10:02:53 -07:00
if err != nil {
return fmt.Errorf("failed to ping newly connected peer: %s\n", err)
2014-08-06 10:02:53 -07:00
}
log.Event(ctx, "connect", dht.self, npeer)
2014-08-06 10:02:53 -07:00
dht.Update(ctx, npeer)
return nil
2014-07-30 20:16:34 -07:00
}
2014-09-16 00:56:40 -07:00
// HandleMessage implements the inet.Handler interface.
func (dht *IpfsDHT) HandleMessage(ctx context.Context, mes msg.NetMessage) msg.NetMessage {
2014-09-16 00:56:40 -07:00
mData := mes.Data()
if mData == nil {
log.Error("Message contained nil data.")
return nil
2014-09-16 00:56:40 -07:00
}
mPeer := mes.Peer()
if mPeer == nil {
log.Error("Message contained nil peer.")
return nil
2014-09-16 00:56:40 -07:00
}
// deserialize msg
pmes := new(pb.Message)
2014-09-16 00:56:40 -07:00
err := proto.Unmarshal(mData, pmes)
if err != nil {
log.Error("Error unmarshaling data")
return nil
2014-09-16 00:56:40 -07:00
}
// update the peer (on valid msgs only)
dht.Update(ctx, mPeer)
2014-09-16 00:56:40 -07:00
log.Event(ctx, "foo", dht.self, mPeer, pmes)
2014-09-16 00:56:40 -07:00
// get handler for this msg type.
handler := dht.handlerForMsgType(pmes.GetType())
if handler == nil {
log.Error("got back nil handler from handlerForMsgType")
return nil
2014-09-16 00:56:40 -07:00
}
// dispatch handler.
rpmes, err := handler(ctx, mPeer, pmes)
2014-09-16 00:56:40 -07:00
if err != nil {
log.Errorf("handle message error: %s", err)
return nil
2014-09-16 00:56:40 -07:00
}
2014-09-19 07:51:03 -07:00
// if nil response, return it before serializing
if rpmes == nil {
log.Warning("Got back nil response from request.")
return nil
2014-09-19 07:51:03 -07:00
}
2014-09-16 00:56:40 -07:00
// serialize response msg
rmes, err := msg.FromObject(mPeer, rpmes)
if err != nil {
log.Errorf("serialze response error: %s", err)
return nil
2014-09-16 00:56:40 -07:00
}
return rmes
2014-09-16 00:56:40 -07:00
}
2014-09-16 02:43:11 -07:00
// sendRequest sends out a request using dht.sender, but also makes sure to
// measure the RTT for latency measurements.
func (dht *IpfsDHT) sendRequest(ctx context.Context, p peer.Peer, pmes *pb.Message) (*pb.Message, error) {
2014-09-16 02:43:11 -07:00
mes, err := msg.FromObject(p, pmes)
if err != nil {
return nil, err
}
start := time.Now()
rmes, err := dht.sender.SendRequest(ctx, mes) // respect?
2014-09-16 02:43:11 -07:00
if err != nil {
return nil, err
}
2014-09-18 19:30:04 -07:00
if rmes == nil {
return nil, errors.New("no response to request")
}
log.Event(ctx, "sentMessage", dht.self, p, pmes)
2014-09-16 02:43:11 -07:00
rmes.Peer().SetLatency(time.Since(start))
2014-09-16 02:43:11 -07:00
rpmes := new(pb.Message)
2014-09-16 02:43:11 -07:00
if err := proto.Unmarshal(rmes.Data(), rpmes); err != nil {
return nil, err
}
return rpmes, nil
}
// putValueToNetwork stores the given key/value pair at the peer 'p'
func (dht *IpfsDHT) putValueToNetwork(ctx context.Context, p peer.Peer,
key string, rec *pb.Record) error {
2014-09-19 08:07:56 -07:00
pmes := pb.NewMessage(pb.Message_PUT_VALUE, string(key), 0)
pmes.Record = rec
2014-09-19 08:07:56 -07:00
rpmes, err := dht.sendRequest(ctx, p, pmes)
2014-09-17 07:19:40 -07:00
if err != nil {
return err
}
2014-09-19 08:07:56 -07:00
if !bytes.Equal(rpmes.GetRecord().Value, pmes.GetRecord().Value) {
2014-09-19 08:07:56 -07:00
return errors.New("value not put correctly")
}
return nil
2014-09-17 07:19:40 -07:00
}
2014-11-03 03:02:56 +00:00
// putProvider sends a message to peer 'p' saying that the local node
// can provide the value of 'key'
func (dht *IpfsDHT) putProvider(ctx context.Context, p peer.Peer, key string) error {
2014-09-17 07:19:40 -07:00
pmes := pb.NewMessage(pb.Message_ADD_PROVIDER, string(key), 0)
// add self as the provider
pmes.ProviderPeers = pb.PeersToPBPeers([]peer.Peer{dht.self})
2014-09-19 14:31:10 -07:00
rpmes, err := dht.sendRequest(ctx, p, pmes)
2014-09-17 07:19:40 -07:00
if err != nil {
return err
}
2014-09-19 14:31:10 -07:00
log.Debugf("%s putProvider: %s for %s", dht.self, p, u.Key(key))
if rpmes.GetKey() != pmes.GetKey() {
2014-09-19 14:31:10 -07:00
return errors.New("provider not added correctly")
}
return nil
2014-09-17 07:19:40 -07:00
}
func (dht *IpfsDHT) getValueOrPeers(ctx context.Context, p peer.Peer,
key u.Key, level int) ([]byte, []peer.Peer, error) {
2014-09-16 02:43:11 -07:00
pmes, err := dht.getValueSingle(ctx, p, key, level)
if err != nil {
return nil, nil, err
}
if record := pmes.GetRecord(); record != nil {
// Success! We were given the value
2014-09-28 00:13:07 -07:00
log.Debug("getValueOrPeers: got value")
// make sure record is still valid
err = dht.verifyRecord(record)
if err != nil {
2014-11-14 11:00:45 -08:00
log.Error("Received invalid record!")
return nil, nil, err
}
return record.GetValue(), nil, nil
2014-08-16 23:03:36 -07:00
}
// TODO decide on providers. This probably shouldn't be happening.
2014-09-18 19:30:04 -07:00
if prv := pmes.GetProviderPeers(); prv != nil && len(prv) > 0 {
val, err := dht.getFromPeerList(ctx, key, prv, level)
if err != nil {
return nil, nil, err
}
2014-09-28 00:13:07 -07:00
log.Debug("getValueOrPeers: get from providers")
2014-09-18 19:30:04 -07:00
return val, nil, nil
}
// Perhaps we were given closer peers
var peers []peer.Peer
for _, pb := range pmes.GetCloserPeers() {
2014-10-21 01:18:20 -07:00
pr, err := dht.peerFromInfo(pb)
2014-08-16 23:03:36 -07:00
if err != nil {
log.Error(err)
2014-08-16 23:03:36 -07:00
continue
}
2014-09-16 06:33:51 -07:00
peers = append(peers, pr)
}
if len(peers) > 0 {
log.Debug("getValueOrPeers: peers")
return nil, peers, nil
}
log.Warning("getValueOrPeers: routing.ErrNotFound")
return nil, nil, routing.ErrNotFound
}
// getValueSingle simply performs the get value RPC with the given parameters
func (dht *IpfsDHT) getValueSingle(ctx context.Context, p peer.Peer,
key u.Key, level int) (*pb.Message, error) {
2014-09-16 02:43:11 -07:00
pmes := pb.NewMessage(pb.Message_GET_VALUE, string(key), level)
2014-09-16 02:43:11 -07:00
return dht.sendRequest(ctx, p, pmes)
2014-08-09 22:28:46 -07:00
}
// TODO: Im not certain on this implementation, we get a list of peers/providers
// from someone what do we do with it? Connect to each of them? randomly pick
// one to get the value from? Or just connect to one at a time until we get a
// successful connection and request the value from it?
2014-09-16 06:33:51 -07:00
func (dht *IpfsDHT) getFromPeerList(ctx context.Context, key u.Key,
peerlist []*pb.Message_Peer, level int) ([]byte, error) {
2014-09-16 06:33:51 -07:00
for _, pinfo := range peerlist {
2014-11-05 04:26:30 -08:00
p, err := dht.ensureConnectedToPeer(ctx, pinfo)
2014-09-16 06:33:51 -07:00
if err != nil {
log.Errorf("getFromPeers error: %s", err)
2014-09-16 06:33:51 -07:00
continue
2014-08-09 22:28:46 -07:00
}
2014-09-16 06:33:51 -07:00
pmes, err := dht.getValueSingle(ctx, p, key, level)
2014-08-09 22:28:46 -07:00
if err != nil {
log.Errorf("getFromPeers error: %s\n", err)
2014-08-09 22:28:46 -07:00
continue
}
if record := pmes.GetRecord(); record != nil {
2014-09-16 06:33:51 -07:00
// Success! We were given the value
err := dht.verifyRecord(record)
if err != nil {
return nil, err
}
2014-09-16 06:33:51 -07:00
dht.providers.AddProvider(key, p)
return record.GetValue(), nil
}
2014-08-09 22:28:46 -07:00
}
return nil, routing.ErrNotFound
2014-08-09 22:28:46 -07:00
}
// getLocal attempts to retrieve the value from the datastore
2014-08-16 23:03:36 -07:00
func (dht *IpfsDHT) getLocal(key u.Key) ([]byte, error) {
dht.dslock.Lock()
defer dht.dslock.Unlock()
2014-11-11 16:28:20 -08:00
log.Debug("getLocal %s", key)
2014-10-03 15:34:08 -07:00
v, err := dht.datastore.Get(key.DsKey())
if err != nil {
return nil, err
}
2014-11-11 16:28:20 -08:00
log.Debug("found in db")
byt, ok := v.([]byte)
if !ok {
return nil, errors.New("value stored in datastore not []byte")
}
rec := new(pb.Record)
err = proto.Unmarshal(byt, rec)
if err != nil {
return nil, err
}
// TODO: 'if paranoid'
if u.Debug {
err = dht.verifyRecord(rec)
if err != nil {
2014-11-11 16:28:20 -08:00
log.Errorf("local record verify failed: %s", err)
return nil, err
}
}
return rec.GetValue(), nil
}
// putLocal stores the key value pair in the datastore
2014-08-16 23:03:36 -07:00
func (dht *IpfsDHT) putLocal(key u.Key, value []byte) error {
rec, err := dht.makePutRecord(key, value)
if err != nil {
return err
}
data, err := proto.Marshal(rec)
if err != nil {
return err
}
return dht.datastore.Put(key.DsKey(), data)
}
2014-09-16 06:40:17 -07:00
// Update signals to all routingTables to Update their last-seen status
// on the given peer.
func (dht *IpfsDHT) Update(ctx context.Context, p peer.Peer) {
log.Event(ctx, "updatePeer", p)
2014-09-16 06:40:17 -07:00
removedCount := 0
2014-08-16 23:03:36 -07:00
for _, route := range dht.routingTables {
removed := route.Update(p)
2014-08-30 00:00:52 -07:00
// Only close the connection if no tables refer to this peer
if removed != nil {
2014-09-16 06:40:17 -07:00
removedCount++
}
}
2014-09-16 06:40:17 -07:00
// Only close the connection if no tables refer to this peer
// if removedCount == len(dht.routingTables) {
// dht.network.ClosePeer(p)
// }
// ACTUALLY, no, let's not just close the connection. it may be connected
// due to other things. it seems that we just need connection timeouts
// after some deadline of inactivity.
}
2014-08-09 22:28:46 -07:00
2014-10-11 10:43:54 -07:00
// FindLocal looks for a peer with a given ID connected to this dht and returns the peer and the table it was found in.
func (dht *IpfsDHT) FindLocal(id peer.ID) (peer.Peer, *kb.RoutingTable) {
2014-08-16 23:03:36 -07:00
for _, table := range dht.routingTables {
2014-08-09 22:28:46 -07:00
p := table.Find(id)
if p != nil {
return p, table
}
}
return nil, nil
}
2014-11-03 03:02:56 +00:00
// findPeerSingle asks peer 'p' if they know where the peer with id 'id' is
func (dht *IpfsDHT) findPeerSingle(ctx context.Context, p peer.Peer, id peer.ID, level int) (*pb.Message, error) {
pmes := pb.NewMessage(pb.Message_FIND_NODE, string(id), level)
2014-09-16 07:17:55 -07:00
return dht.sendRequest(ctx, p, pmes)
}
func (dht *IpfsDHT) findProvidersSingle(ctx context.Context, p peer.Peer, key u.Key, level int) (*pb.Message, error) {
pmes := pb.NewMessage(pb.Message_GET_PROVIDERS, string(key), level)
2014-09-16 07:17:55 -07:00
return dht.sendRequest(ctx, p, pmes)
2014-08-14 08:32:17 -07:00
}
func (dht *IpfsDHT) addProviders(key u.Key, peers []*pb.Message_Peer) []peer.Peer {
var provArr []peer.Peer
2014-08-14 08:32:17 -07:00
for _, prov := range peers {
2014-09-16 07:17:55 -07:00
p, err := dht.peerFromInfo(prov)
if err != nil {
log.Errorf("error getting peer from info: %v", err)
2014-08-14 08:32:17 -07:00
continue
}
2014-09-16 07:17:55 -07:00
log.Debugf("%s adding provider: %s for %s", dht.self, p, key)
2014-09-19 18:11:05 -07:00
2014-09-16 07:17:55 -07:00
// Dont add outselves to the list
if p.ID().Equal(dht.self.ID()) {
2014-09-16 07:17:55 -07:00
continue
2014-08-14 08:32:17 -07:00
}
2014-09-16 07:17:55 -07:00
// TODO(jbenet) ensure providers is idempotent
dht.providers.AddProvider(key, p)
2014-08-16 23:03:36 -07:00
provArr = append(provArr, p)
2014-08-14 08:32:17 -07:00
}
2014-08-16 23:03:36 -07:00
return provArr
2014-08-14 08:32:17 -07:00
}
2014-08-26 14:24:51 -07:00
// nearestPeersToQuery returns the routing tables closest peers.
func (dht *IpfsDHT) nearestPeersToQuery(pmes *pb.Message, count int) []peer.Peer {
level := pmes.GetClusterLevel()
cluster := dht.routingTables[level]
key := u.Key(pmes.GetKey())
closer := cluster.NearestPeers(kb.ConvertKey(key), count)
return closer
}
// betterPeerToQuery returns nearestPeersToQuery, but iff closer than self.
func (dht *IpfsDHT) betterPeersToQuery(pmes *pb.Message, count int) []peer.Peer {
closer := dht.nearestPeersToQuery(pmes, count)
// no node? nil
if closer == nil {
return nil
}
// == to self? thats bad
for _, p := range closer {
if p.ID().Equal(dht.self.ID()) {
log.Error("Attempted to return self! this shouldnt happen...")
return nil
}
}
var filtered []peer.Peer
for _, p := range closer {
// must all be closer than self
key := u.Key(pmes.GetKey())
if !kb.Closer(dht.self.ID(), p.ID(), key) {
filtered = append(filtered, p)
}
}
// ok seems like closer nodes
return filtered
}
2014-11-03 03:02:56 +00:00
// getPeer searches the peerstore for a peer with the given peer ID
func (dht *IpfsDHT) getPeer(id peer.ID) (peer.Peer, error) {
p, err := dht.peerstore.FindOrCreate(id)
if err != nil {
err = fmt.Errorf("Failed to get peer from peerstore: %s", err)
log.Error(err)
return nil, err
}
return p, nil
}
2014-11-03 03:02:56 +00:00
// peerFromInfo returns a peer using info in the protobuf peer struct
// to lookup or create a peer
func (dht *IpfsDHT) peerFromInfo(pbp *pb.Message_Peer) (peer.Peer, error) {
2014-09-16 06:33:51 -07:00
id := peer.ID(pbp.GetId())
2014-09-17 07:19:40 -07:00
2014-10-21 01:18:20 -07:00
// bail out if it's ourselves
//TODO(jbenet) not sure this should be an error _here_
if id.Equal(dht.self.ID()) {
2014-09-17 07:19:40 -07:00
return nil, errors.New("found self")
}
p, err := dht.getPeer(id)
if err != nil {
return nil, err
2014-09-16 06:33:51 -07:00
}
// add addresses we've just discovered
maddrs, err := pbp.Addresses()
if err != nil {
return nil, err
2014-09-16 07:17:55 -07:00
}
for _, maddr := range maddrs {
p.AddAddress(maddr)
}
2014-09-16 07:17:55 -07:00
return p, nil
}
2014-11-05 04:26:30 -08:00
func (dht *IpfsDHT) ensureConnectedToPeer(ctx context.Context, pbp *pb.Message_Peer) (peer.Peer, error) {
2014-09-16 07:17:55 -07:00
p, err := dht.peerFromInfo(pbp)
if err != nil {
return nil, err
2014-08-26 14:24:51 -07:00
}
2014-09-16 06:33:51 -07:00
// dial connection
2014-11-05 04:26:30 -08:00
err = dht.dialer.DialPeer(ctx, p)
2014-09-16 06:33:51 -07:00
return p, err
2014-08-26 14:24:51 -07:00
}
2014-10-03 15:34:08 -07:00
//TODO: this should be smarter about which keys it selects.
func (dht *IpfsDHT) loadProvidableKeys() error {
kl, err := dht.datastore.KeyList()
if err != nil {
return err
}
2014-10-03 15:34:08 -07:00
for _, dsk := range kl {
k := u.KeyFromDsKey(dsk)
if len(k) == 0 {
log.Errorf("loadProvidableKeys error: %v", dsk)
2014-10-03 15:34:08 -07:00
}
dht.providers.AddProvider(k, dht.self)
}
return nil
}
// PingRoutine periodically pings nearest neighbors.
func (dht *IpfsDHT) PingRoutine(t time.Duration) {
2014-10-25 07:12:01 -07:00
defer dht.Children().Done()
tick := time.Tick(t)
for {
select {
case <-tick:
id := make([]byte, 16)
rand.Read(id)
peers := dht.routingTables[0].NearestPeers(kb.ConvertKey(u.Key(id)), 5)
for _, p := range peers {
2014-10-25 07:12:01 -07:00
ctx, _ := context.WithTimeout(dht.Context(), time.Second*5)
err := dht.Ping(ctx, p)
if err != nil {
log.Errorf("Ping error: %s", err)
}
}
2014-10-25 07:12:01 -07:00
case <-dht.Closing():
return
}
}
}
2014-09-16 07:17:55 -07:00
// Bootstrap builds up list of peers by requesting random peer IDs
func (dht *IpfsDHT) Bootstrap(ctx context.Context) {
id := make([]byte, 16)
rand.Read(id)
2014-10-25 03:36:00 -07:00
p, err := dht.FindPeer(ctx, peer.ID(id))
if err != nil {
2014-11-20 10:46:19 -08:00
log.Errorf("Bootstrap peer error: %s", err)
2014-10-25 03:36:00 -07:00
}
2014-11-05 04:26:30 -08:00
err = dht.dialer.DialPeer(ctx, p)
if err != nil {
log.Errorf("Bootstrap peer error: %s", err)
}
}