mirror of
https://github.com/fluencelabs/go-libp2p-kad-dht
synced 2025-04-24 14:22:13 +00:00
* test: augment TestProviderManager test, add notes of future tests * refactor(provider-manager): order funcs, update names for consistency, add code docs
336 lines
8.0 KiB
Go
336 lines
8.0 KiB
Go
package providers
|
|
|
|
import (
|
|
"context"
|
|
"encoding/binary"
|
|
"fmt"
|
|
"strings"
|
|
"time"
|
|
|
|
"github.com/libp2p/go-libp2p-core/peer"
|
|
|
|
lru "github.com/hashicorp/golang-lru/simplelru"
|
|
ds "github.com/ipfs/go-datastore"
|
|
autobatch "github.com/ipfs/go-datastore/autobatch"
|
|
dsq "github.com/ipfs/go-datastore/query"
|
|
logging "github.com/ipfs/go-log"
|
|
goprocess "github.com/jbenet/goprocess"
|
|
goprocessctx "github.com/jbenet/goprocess/context"
|
|
base32 "github.com/multiformats/go-base32"
|
|
)
|
|
|
|
// ProvidersKeyPrefix is the prefix/namespace for ALL provider record
|
|
// keys stored in the data store.
|
|
const ProvidersKeyPrefix = "/providers/"
|
|
|
|
// ProvideValidity is the default time that a provider record should last
|
|
var ProvideValidity = time.Hour * 24
|
|
var defaultCleanupInterval = time.Hour
|
|
var lruCacheSize = 256
|
|
var batchBufferSize = 256
|
|
var log = logging.Logger("providers")
|
|
|
|
// ProviderManager adds and pulls providers out of the datastore,
|
|
// caching them in between
|
|
type ProviderManager struct {
|
|
// all non channel fields are meant to be accessed only within
|
|
// the run method
|
|
cache *lru.LRU
|
|
dstore *autobatch.Datastore
|
|
|
|
newprovs chan *addProv
|
|
getprovs chan *getProv
|
|
proc goprocess.Process
|
|
|
|
cleanupInterval time.Duration
|
|
}
|
|
|
|
type addProv struct {
|
|
key []byte
|
|
val peer.ID
|
|
}
|
|
|
|
type getProv struct {
|
|
key []byte
|
|
resp chan []peer.ID
|
|
}
|
|
|
|
// NewProviderManager constructor
|
|
func NewProviderManager(ctx context.Context, local peer.ID, dstore ds.Batching) *ProviderManager {
|
|
pm := new(ProviderManager)
|
|
pm.getprovs = make(chan *getProv)
|
|
pm.newprovs = make(chan *addProv)
|
|
pm.dstore = autobatch.NewAutoBatching(dstore, batchBufferSize)
|
|
cache, err := lru.NewLRU(lruCacheSize, nil)
|
|
if err != nil {
|
|
panic(err) //only happens if negative value is passed to lru constructor
|
|
}
|
|
pm.cache = cache
|
|
|
|
pm.proc = goprocessctx.WithContext(ctx)
|
|
pm.cleanupInterval = defaultCleanupInterval
|
|
pm.proc.Go(pm.run)
|
|
|
|
return pm
|
|
}
|
|
|
|
// Process returns the ProviderManager process
|
|
func (pm *ProviderManager) Process() goprocess.Process {
|
|
return pm.proc
|
|
}
|
|
|
|
func (pm *ProviderManager) run(proc goprocess.Process) {
|
|
var (
|
|
gcQuery dsq.Results
|
|
gcQueryRes <-chan dsq.Result
|
|
gcSkip map[string]struct{}
|
|
gcTime time.Time
|
|
gcTimer = time.NewTimer(pm.cleanupInterval)
|
|
)
|
|
|
|
defer func() {
|
|
gcTimer.Stop()
|
|
if gcQuery != nil {
|
|
// don't really care if this fails.
|
|
_ = gcQuery.Close()
|
|
}
|
|
if err := pm.dstore.Flush(); err != nil {
|
|
log.Error("failed to flush datastore: ", err)
|
|
}
|
|
}()
|
|
|
|
for {
|
|
select {
|
|
case np := <-pm.newprovs:
|
|
err := pm.addProv(np.key, np.val)
|
|
if err != nil {
|
|
log.Error("error adding new providers: ", err)
|
|
continue
|
|
}
|
|
if gcSkip != nil {
|
|
// we have an gc, tell it to skip this provider
|
|
// as we've updated it since the GC started.
|
|
gcSkip[mkProvKeyFor(np.key, np.val)] = struct{}{}
|
|
}
|
|
case gp := <-pm.getprovs:
|
|
provs, err := pm.getProvidersForKey(gp.key)
|
|
if err != nil && err != ds.ErrNotFound {
|
|
log.Error("error reading providers: ", err)
|
|
}
|
|
|
|
// set the cap so the user can't append to this.
|
|
gp.resp <- provs[0:len(provs):len(provs)]
|
|
case res, ok := <-gcQueryRes:
|
|
if !ok {
|
|
if err := gcQuery.Close(); err != nil {
|
|
log.Error("failed to close provider GC query: ", err)
|
|
}
|
|
gcTimer.Reset(pm.cleanupInterval)
|
|
|
|
// cleanup GC round
|
|
gcQueryRes = nil
|
|
gcSkip = nil
|
|
gcQuery = nil
|
|
continue
|
|
}
|
|
if res.Error != nil {
|
|
log.Error("got error from GC query: ", res.Error)
|
|
continue
|
|
}
|
|
if _, ok := gcSkip[res.Key]; ok {
|
|
// We've updated this record since starting the
|
|
// GC round, skip it.
|
|
continue
|
|
}
|
|
|
|
// check expiration time
|
|
t, err := readTimeValue(res.Value)
|
|
switch {
|
|
case err != nil:
|
|
// couldn't parse the time
|
|
log.Error("parsing providers record from disk: ", err)
|
|
fallthrough
|
|
case gcTime.Sub(t) > ProvideValidity:
|
|
// or expired
|
|
err = pm.dstore.Delete(ds.RawKey(res.Key))
|
|
if err != nil && err != ds.ErrNotFound {
|
|
log.Error("failed to remove provider record from disk: ", err)
|
|
}
|
|
}
|
|
|
|
case gcTime = <-gcTimer.C:
|
|
// You know the wonderful thing about caches? You can
|
|
// drop them.
|
|
//
|
|
// Much faster than GCing.
|
|
pm.cache.Purge()
|
|
|
|
// Now, kick off a GC of the datastore.
|
|
q, err := pm.dstore.Query(dsq.Query{
|
|
Prefix: ProvidersKeyPrefix,
|
|
})
|
|
if err != nil {
|
|
log.Error("provider record GC query failed: ", err)
|
|
continue
|
|
}
|
|
gcQuery = q
|
|
gcQueryRes = q.Next()
|
|
gcSkip = make(map[string]struct{})
|
|
case <-proc.Closing():
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// AddProvider adds a provider
|
|
func (pm *ProviderManager) AddProvider(ctx context.Context, k []byte, val peer.ID) {
|
|
prov := &addProv{
|
|
key: k,
|
|
val: val,
|
|
}
|
|
select {
|
|
case pm.newprovs <- prov:
|
|
case <-ctx.Done():
|
|
}
|
|
}
|
|
|
|
// addProv updates the cache if needed
|
|
func (pm *ProviderManager) addProv(k []byte, p peer.ID) error {
|
|
now := time.Now()
|
|
if provs, ok := pm.cache.Get(string(k)); ok {
|
|
provs.(*providerSet).setVal(p, now)
|
|
} // else not cached, just write through
|
|
|
|
return writeProviderEntry(pm.dstore, k, p, now)
|
|
}
|
|
|
|
// writeProviderEntry writes the provider into the datastore
|
|
func writeProviderEntry(dstore ds.Datastore, k []byte, p peer.ID, t time.Time) error {
|
|
dsk := mkProvKeyFor(k, p)
|
|
|
|
buf := make([]byte, 16)
|
|
n := binary.PutVarint(buf, t.UnixNano())
|
|
|
|
return dstore.Put(ds.NewKey(dsk), buf[:n])
|
|
}
|
|
|
|
func mkProvKeyFor(k []byte, p peer.ID) string {
|
|
return mkProvKey(k) + "/" + base32.RawStdEncoding.EncodeToString([]byte(p))
|
|
}
|
|
|
|
func mkProvKey(k []byte) string {
|
|
return ProvidersKeyPrefix + base32.RawStdEncoding.EncodeToString(k)
|
|
}
|
|
|
|
// GetProviders returns the set of providers for the given key.
|
|
// This method _does not_ copy the set. Do not modify it.
|
|
func (pm *ProviderManager) GetProviders(ctx context.Context, k []byte) []peer.ID {
|
|
gp := &getProv{
|
|
key: k,
|
|
resp: make(chan []peer.ID, 1), // buffered to prevent sender from blocking
|
|
}
|
|
select {
|
|
case <-ctx.Done():
|
|
return nil
|
|
case pm.getprovs <- gp:
|
|
}
|
|
select {
|
|
case <-ctx.Done():
|
|
return nil
|
|
case peers := <-gp.resp:
|
|
return peers
|
|
}
|
|
}
|
|
|
|
func (pm *ProviderManager) getProvidersForKey(k []byte) ([]peer.ID, error) {
|
|
pset, err := pm.getProviderSetForKey(k)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return pset.providers, nil
|
|
}
|
|
|
|
// returns the ProviderSet if it already exists on cache, otherwise loads it from datasatore
|
|
func (pm *ProviderManager) getProviderSetForKey(k []byte) (*providerSet, error) {
|
|
cached, ok := pm.cache.Get(string(k))
|
|
if ok {
|
|
return cached.(*providerSet), nil
|
|
}
|
|
|
|
pset, err := loadProviderSet(pm.dstore, k)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if len(pset.providers) > 0 {
|
|
pm.cache.Add(string(k), pset)
|
|
}
|
|
|
|
return pset, nil
|
|
}
|
|
|
|
// loads the ProviderSet out of the datastore
|
|
func loadProviderSet(dstore ds.Datastore, k []byte) (*providerSet, error) {
|
|
res, err := dstore.Query(dsq.Query{Prefix: mkProvKey(k)})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
defer res.Close()
|
|
|
|
now := time.Now()
|
|
out := newProviderSet()
|
|
for {
|
|
e, ok := res.NextSync()
|
|
if !ok {
|
|
break
|
|
}
|
|
if e.Error != nil {
|
|
log.Error("got an error: ", e.Error)
|
|
continue
|
|
}
|
|
|
|
// check expiration time
|
|
t, err := readTimeValue(e.Value)
|
|
switch {
|
|
case err != nil:
|
|
// couldn't parse the time
|
|
log.Error("parsing providers record from disk: ", err)
|
|
fallthrough
|
|
case now.Sub(t) > ProvideValidity:
|
|
// or just expired
|
|
err = dstore.Delete(ds.RawKey(e.Key))
|
|
if err != nil && err != ds.ErrNotFound {
|
|
log.Error("failed to remove provider record from disk: ", err)
|
|
}
|
|
continue
|
|
}
|
|
|
|
lix := strings.LastIndex(e.Key, "/")
|
|
|
|
decstr, err := base32.RawStdEncoding.DecodeString(e.Key[lix+1:])
|
|
if err != nil {
|
|
log.Error("base32 decoding error: ", err)
|
|
err = dstore.Delete(ds.RawKey(e.Key))
|
|
if err != nil && err != ds.ErrNotFound {
|
|
log.Error("failed to remove provider record from disk: ", err)
|
|
}
|
|
continue
|
|
}
|
|
|
|
pid := peer.ID(decstr)
|
|
|
|
out.setVal(pid, t)
|
|
}
|
|
|
|
return out, nil
|
|
}
|
|
|
|
func readTimeValue(data []byte) (time.Time, error) {
|
|
nsec, n := binary.Varint(data)
|
|
if n <= 0 {
|
|
return time.Time{}, fmt.Errorf("failed to parse time")
|
|
}
|
|
|
|
return time.Unix(0, nsec), nil
|
|
}
|