Integrate BACKBEAT SDK and resolve KACHING license validation
Major integrations and fixes: - Added BACKBEAT SDK integration for P2P operation timing - Implemented beat-aware status tracking for distributed operations - Added Docker secrets support for secure license management - Resolved KACHING license validation via HTTPS/TLS - Updated docker-compose configuration for clean stack deployment - Disabled rollback policies to prevent deployment failures - Added license credential storage (CHORUS-DEV-MULTI-001) Technical improvements: - BACKBEAT P2P operation tracking with phase management - Enhanced configuration system with file-based secrets - Improved error handling for license validation - Clean separation of KACHING and CHORUS deployment stacks 🤖 Generated with [Claude Code](https://claude.ai/code) Co-Authored-By: Claude <noreply@anthropic.com>
This commit is contained in:
34
vendor/github.com/libp2p/go-libp2p-kad-dht/providers/provider_set.go
generated
vendored
Normal file
34
vendor/github.com/libp2p/go-libp2p-kad-dht/providers/provider_set.go
generated
vendored
Normal file
@@ -0,0 +1,34 @@
|
||||
package providers
|
||||
|
||||
import (
|
||||
"time"
|
||||
|
||||
"github.com/libp2p/go-libp2p/core/peer"
|
||||
)
|
||||
|
||||
// A providerSet has the list of providers and the time that they were added
|
||||
// It is used as an intermediary data struct between what is stored in the datastore
|
||||
// and the list of providers that get passed to the consumer of a .GetProviders call
|
||||
type providerSet struct {
|
||||
providers []peer.ID
|
||||
set map[peer.ID]time.Time
|
||||
}
|
||||
|
||||
func newProviderSet() *providerSet {
|
||||
return &providerSet{
|
||||
set: make(map[peer.ID]time.Time),
|
||||
}
|
||||
}
|
||||
|
||||
func (ps *providerSet) Add(p peer.ID) {
|
||||
ps.setVal(p, time.Now())
|
||||
}
|
||||
|
||||
func (ps *providerSet) setVal(p peer.ID, t time.Time) {
|
||||
_, found := ps.set[p]
|
||||
if !found {
|
||||
ps.providers = append(ps.providers, p)
|
||||
}
|
||||
|
||||
ps.set[p] = t
|
||||
}
|
||||
412
vendor/github.com/libp2p/go-libp2p-kad-dht/providers/providers_manager.go
generated
vendored
Normal file
412
vendor/github.com/libp2p/go-libp2p-kad-dht/providers/providers_manager.go
generated
vendored
Normal file
@@ -0,0 +1,412 @@
|
||||
package providers
|
||||
|
||||
import (
|
||||
"context"
|
||||
"encoding/binary"
|
||||
"fmt"
|
||||
"io"
|
||||
"strings"
|
||||
"sync"
|
||||
"time"
|
||||
|
||||
lru "github.com/hashicorp/golang-lru/simplelru"
|
||||
ds "github.com/ipfs/go-datastore"
|
||||
"github.com/ipfs/go-datastore/autobatch"
|
||||
dsq "github.com/ipfs/go-datastore/query"
|
||||
logging "github.com/ipfs/go-log"
|
||||
"github.com/libp2p/go-libp2p-kad-dht/internal"
|
||||
"github.com/libp2p/go-libp2p/core/peer"
|
||||
"github.com/libp2p/go-libp2p/core/peerstore"
|
||||
peerstoreImpl "github.com/libp2p/go-libp2p/p2p/host/peerstore"
|
||||
"github.com/multiformats/go-base32"
|
||||
)
|
||||
|
||||
const (
|
||||
// ProvidersKeyPrefix is the prefix/namespace for ALL provider record
|
||||
// keys stored in the data store.
|
||||
ProvidersKeyPrefix = "/providers/"
|
||||
|
||||
// ProviderAddrTTL is the TTL to keep the multi addresses of provider
|
||||
// peers around. Those addresses are returned alongside provider. After
|
||||
// it expires, the returned records will require an extra lookup, to
|
||||
// find the multiaddress associated with the returned peer id.
|
||||
ProviderAddrTTL = 24 * time.Hour
|
||||
)
|
||||
|
||||
// ProvideValidity is the default time that a Provider Record should last on DHT
|
||||
// This value is also known as Provider Record Expiration Interval.
|
||||
var ProvideValidity = time.Hour * 48
|
||||
var defaultCleanupInterval = time.Hour
|
||||
var lruCacheSize = 256
|
||||
var batchBufferSize = 256
|
||||
var log = logging.Logger("providers")
|
||||
|
||||
// ProviderStore represents a store that associates peers and their addresses to keys.
|
||||
type ProviderStore interface {
|
||||
AddProvider(ctx context.Context, key []byte, prov peer.AddrInfo) error
|
||||
GetProviders(ctx context.Context, key []byte) ([]peer.AddrInfo, error)
|
||||
io.Closer
|
||||
}
|
||||
|
||||
// ProviderManager adds and pulls providers out of the datastore,
|
||||
// caching them in between
|
||||
type ProviderManager struct {
|
||||
self peer.ID
|
||||
// all non channel fields are meant to be accessed only within
|
||||
// the run method
|
||||
cache lru.LRUCache
|
||||
pstore peerstore.Peerstore
|
||||
dstore *autobatch.Datastore
|
||||
|
||||
newprovs chan *addProv
|
||||
getprovs chan *getProv
|
||||
|
||||
cleanupInterval time.Duration
|
||||
|
||||
ctx context.Context
|
||||
cancel context.CancelFunc
|
||||
wg sync.WaitGroup
|
||||
}
|
||||
|
||||
var _ ProviderStore = (*ProviderManager)(nil)
|
||||
|
||||
// Option is a function that sets a provider manager option.
|
||||
type Option func(*ProviderManager) error
|
||||
|
||||
func (pm *ProviderManager) applyOptions(opts ...Option) error {
|
||||
for i, opt := range opts {
|
||||
if err := opt(pm); err != nil {
|
||||
return fmt.Errorf("provider manager option %d failed: %s", i, err)
|
||||
}
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
// CleanupInterval sets the time between GC runs.
|
||||
// Defaults to 1h.
|
||||
func CleanupInterval(d time.Duration) Option {
|
||||
return func(pm *ProviderManager) error {
|
||||
pm.cleanupInterval = d
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
// Cache sets the LRU cache implementation.
|
||||
// Defaults to a simple LRU cache.
|
||||
func Cache(c lru.LRUCache) Option {
|
||||
return func(pm *ProviderManager) error {
|
||||
pm.cache = c
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
type addProv struct {
|
||||
ctx context.Context
|
||||
key []byte
|
||||
val peer.ID
|
||||
}
|
||||
|
||||
type getProv struct {
|
||||
ctx context.Context
|
||||
key []byte
|
||||
resp chan []peer.ID
|
||||
}
|
||||
|
||||
// NewProviderManager constructor
|
||||
func NewProviderManager(local peer.ID, ps peerstore.Peerstore, dstore ds.Batching, opts ...Option) (*ProviderManager, error) {
|
||||
pm := new(ProviderManager)
|
||||
pm.self = local
|
||||
pm.getprovs = make(chan *getProv)
|
||||
pm.newprovs = make(chan *addProv)
|
||||
pm.pstore = ps
|
||||
pm.dstore = autobatch.NewAutoBatching(dstore, batchBufferSize)
|
||||
cache, err := lru.NewLRU(lruCacheSize, nil)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
pm.cache = cache
|
||||
pm.cleanupInterval = defaultCleanupInterval
|
||||
if err := pm.applyOptions(opts...); err != nil {
|
||||
return nil, err
|
||||
}
|
||||
pm.ctx, pm.cancel = context.WithCancel(context.Background())
|
||||
pm.run()
|
||||
return pm, nil
|
||||
}
|
||||
|
||||
func (pm *ProviderManager) run() {
|
||||
pm.wg.Add(1)
|
||||
go func() {
|
||||
defer pm.wg.Done()
|
||||
|
||||
var gcQuery dsq.Results
|
||||
gcTimer := time.NewTimer(pm.cleanupInterval)
|
||||
|
||||
defer func() {
|
||||
gcTimer.Stop()
|
||||
if gcQuery != nil {
|
||||
// don't really care if this fails.
|
||||
_ = gcQuery.Close()
|
||||
}
|
||||
if err := pm.dstore.Flush(context.Background()); err != nil {
|
||||
log.Error("failed to flush datastore: ", err)
|
||||
}
|
||||
}()
|
||||
|
||||
var gcQueryRes <-chan dsq.Result
|
||||
var gcSkip map[string]struct{}
|
||||
var gcTime time.Time
|
||||
for {
|
||||
select {
|
||||
case np := <-pm.newprovs:
|
||||
err := pm.addProv(np.ctx, np.key, np.val)
|
||||
if err != nil {
|
||||
log.Error("error adding new providers: ", err)
|
||||
continue
|
||||
}
|
||||
if gcSkip != nil {
|
||||
// we have an gc, tell it to skip this provider
|
||||
// as we've updated it since the GC started.
|
||||
gcSkip[mkProvKeyFor(np.key, np.val)] = struct{}{}
|
||||
}
|
||||
case gp := <-pm.getprovs:
|
||||
provs, err := pm.getProvidersForKey(gp.ctx, gp.key)
|
||||
if err != nil && err != ds.ErrNotFound {
|
||||
log.Error("error reading providers: ", err)
|
||||
}
|
||||
|
||||
// set the cap so the user can't append to this.
|
||||
gp.resp <- provs[0:len(provs):len(provs)]
|
||||
case res, ok := <-gcQueryRes:
|
||||
if !ok {
|
||||
if err := gcQuery.Close(); err != nil {
|
||||
log.Error("failed to close provider GC query: ", err)
|
||||
}
|
||||
gcTimer.Reset(pm.cleanupInterval)
|
||||
|
||||
// cleanup GC round
|
||||
gcQueryRes = nil
|
||||
gcSkip = nil
|
||||
gcQuery = nil
|
||||
continue
|
||||
}
|
||||
if res.Error != nil {
|
||||
log.Error("got error from GC query: ", res.Error)
|
||||
continue
|
||||
}
|
||||
if _, ok := gcSkip[res.Key]; ok {
|
||||
// We've updated this record since starting the
|
||||
// GC round, skip it.
|
||||
continue
|
||||
}
|
||||
|
||||
// check expiration time
|
||||
t, err := readTimeValue(res.Value)
|
||||
switch {
|
||||
case err != nil:
|
||||
// couldn't parse the time
|
||||
log.Error("parsing providers record from disk: ", err)
|
||||
fallthrough
|
||||
case gcTime.Sub(t) > ProvideValidity:
|
||||
// or expired
|
||||
err = pm.dstore.Delete(pm.ctx, ds.RawKey(res.Key))
|
||||
if err != nil && err != ds.ErrNotFound {
|
||||
log.Error("failed to remove provider record from disk: ", err)
|
||||
}
|
||||
}
|
||||
|
||||
case gcTime = <-gcTimer.C:
|
||||
// You know the wonderful thing about caches? You can
|
||||
// drop them.
|
||||
//
|
||||
// Much faster than GCing.
|
||||
pm.cache.Purge()
|
||||
|
||||
// Now, kick off a GC of the datastore.
|
||||
q, err := pm.dstore.Query(pm.ctx, dsq.Query{
|
||||
Prefix: ProvidersKeyPrefix,
|
||||
})
|
||||
if err != nil {
|
||||
log.Error("provider record GC query failed: ", err)
|
||||
continue
|
||||
}
|
||||
gcQuery = q
|
||||
gcQueryRes = q.Next()
|
||||
gcSkip = make(map[string]struct{})
|
||||
case <-pm.ctx.Done():
|
||||
return
|
||||
}
|
||||
}
|
||||
}()
|
||||
}
|
||||
|
||||
func (pm *ProviderManager) Close() error {
|
||||
pm.cancel()
|
||||
pm.wg.Wait()
|
||||
return nil
|
||||
}
|
||||
|
||||
// AddProvider adds a provider
|
||||
func (pm *ProviderManager) AddProvider(ctx context.Context, k []byte, provInfo peer.AddrInfo) error {
|
||||
ctx, span := internal.StartSpan(ctx, "ProviderManager.AddProvider")
|
||||
defer span.End()
|
||||
|
||||
if provInfo.ID != pm.self { // don't add own addrs.
|
||||
pm.pstore.AddAddrs(provInfo.ID, provInfo.Addrs, ProviderAddrTTL)
|
||||
}
|
||||
prov := &addProv{
|
||||
ctx: ctx,
|
||||
key: k,
|
||||
val: provInfo.ID,
|
||||
}
|
||||
select {
|
||||
case pm.newprovs <- prov:
|
||||
return nil
|
||||
case <-ctx.Done():
|
||||
return ctx.Err()
|
||||
}
|
||||
}
|
||||
|
||||
// addProv updates the cache if needed
|
||||
func (pm *ProviderManager) addProv(ctx context.Context, k []byte, p peer.ID) error {
|
||||
now := time.Now()
|
||||
if provs, ok := pm.cache.Get(string(k)); ok {
|
||||
provs.(*providerSet).setVal(p, now)
|
||||
} // else not cached, just write through
|
||||
|
||||
return writeProviderEntry(ctx, pm.dstore, k, p, now)
|
||||
}
|
||||
|
||||
// writeProviderEntry writes the provider into the datastore
|
||||
func writeProviderEntry(ctx context.Context, dstore ds.Datastore, k []byte, p peer.ID, t time.Time) error {
|
||||
dsk := mkProvKeyFor(k, p)
|
||||
|
||||
buf := make([]byte, 16)
|
||||
n := binary.PutVarint(buf, t.UnixNano())
|
||||
|
||||
return dstore.Put(ctx, ds.NewKey(dsk), buf[:n])
|
||||
}
|
||||
|
||||
func mkProvKeyFor(k []byte, p peer.ID) string {
|
||||
return mkProvKey(k) + "/" + base32.RawStdEncoding.EncodeToString([]byte(p))
|
||||
}
|
||||
|
||||
func mkProvKey(k []byte) string {
|
||||
return ProvidersKeyPrefix + base32.RawStdEncoding.EncodeToString(k)
|
||||
}
|
||||
|
||||
// GetProviders returns the set of providers for the given key.
|
||||
// This method _does not_ copy the set. Do not modify it.
|
||||
func (pm *ProviderManager) GetProviders(ctx context.Context, k []byte) ([]peer.AddrInfo, error) {
|
||||
ctx, span := internal.StartSpan(ctx, "ProviderManager.GetProviders")
|
||||
defer span.End()
|
||||
|
||||
gp := &getProv{
|
||||
ctx: ctx,
|
||||
key: k,
|
||||
resp: make(chan []peer.ID, 1), // buffered to prevent sender from blocking
|
||||
}
|
||||
select {
|
||||
case <-ctx.Done():
|
||||
return nil, ctx.Err()
|
||||
case pm.getprovs <- gp:
|
||||
}
|
||||
select {
|
||||
case <-ctx.Done():
|
||||
return nil, ctx.Err()
|
||||
case peers := <-gp.resp:
|
||||
return peerstoreImpl.PeerInfos(pm.pstore, peers), nil
|
||||
}
|
||||
}
|
||||
|
||||
func (pm *ProviderManager) getProvidersForKey(ctx context.Context, k []byte) ([]peer.ID, error) {
|
||||
pset, err := pm.getProviderSetForKey(ctx, k)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
return pset.providers, nil
|
||||
}
|
||||
|
||||
// returns the ProviderSet if it already exists on cache, otherwise loads it from datasatore
|
||||
func (pm *ProviderManager) getProviderSetForKey(ctx context.Context, k []byte) (*providerSet, error) {
|
||||
cached, ok := pm.cache.Get(string(k))
|
||||
if ok {
|
||||
return cached.(*providerSet), nil
|
||||
}
|
||||
|
||||
pset, err := loadProviderSet(ctx, pm.dstore, k)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
if len(pset.providers) > 0 {
|
||||
pm.cache.Add(string(k), pset)
|
||||
}
|
||||
|
||||
return pset, nil
|
||||
}
|
||||
|
||||
// loads the ProviderSet out of the datastore
|
||||
func loadProviderSet(ctx context.Context, dstore ds.Datastore, k []byte) (*providerSet, error) {
|
||||
res, err := dstore.Query(ctx, dsq.Query{Prefix: mkProvKey(k)})
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
defer res.Close()
|
||||
|
||||
now := time.Now()
|
||||
out := newProviderSet()
|
||||
for {
|
||||
e, ok := res.NextSync()
|
||||
if !ok {
|
||||
break
|
||||
}
|
||||
if e.Error != nil {
|
||||
log.Error("got an error: ", e.Error)
|
||||
continue
|
||||
}
|
||||
|
||||
// check expiration time
|
||||
t, err := readTimeValue(e.Value)
|
||||
switch {
|
||||
case err != nil:
|
||||
// couldn't parse the time
|
||||
log.Error("parsing providers record from disk: ", err)
|
||||
fallthrough
|
||||
case now.Sub(t) > ProvideValidity:
|
||||
// or just expired
|
||||
err = dstore.Delete(ctx, ds.RawKey(e.Key))
|
||||
if err != nil && err != ds.ErrNotFound {
|
||||
log.Error("failed to remove provider record from disk: ", err)
|
||||
}
|
||||
continue
|
||||
}
|
||||
|
||||
lix := strings.LastIndex(e.Key, "/")
|
||||
|
||||
decstr, err := base32.RawStdEncoding.DecodeString(e.Key[lix+1:])
|
||||
if err != nil {
|
||||
log.Error("base32 decoding error: ", err)
|
||||
err = dstore.Delete(ctx, ds.RawKey(e.Key))
|
||||
if err != nil && err != ds.ErrNotFound {
|
||||
log.Error("failed to remove provider record from disk: ", err)
|
||||
}
|
||||
continue
|
||||
}
|
||||
|
||||
pid := peer.ID(decstr)
|
||||
|
||||
out.setVal(pid, t)
|
||||
}
|
||||
|
||||
return out, nil
|
||||
}
|
||||
|
||||
func readTimeValue(data []byte) (time.Time, error) {
|
||||
nsec, n := binary.Varint(data)
|
||||
if n <= 0 {
|
||||
return time.Time{}, fmt.Errorf("failed to parse time")
|
||||
}
|
||||
|
||||
return time.Unix(0, nsec), nil
|
||||
}
|
||||
Reference in New Issue
Block a user