mirror of
https://github.com/LBRYFoundation/reflector.go.git
synced 2025-09-20 18:09:52 +00:00
353 lines
8.4 KiB
Go
353 lines
8.4 KiB
Go
package store
|
|
|
|
import (
|
|
"encoding/json"
|
|
"fmt"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/c2h5oh/datasize"
|
|
"github.com/lbryio/reflector.go/db"
|
|
"github.com/lbryio/reflector.go/shared"
|
|
|
|
"github.com/lbryio/lbry.go/v2/extras/errors"
|
|
"github.com/lbryio/lbry.go/v2/extras/stop"
|
|
"github.com/lbryio/lbry.go/v2/stream"
|
|
|
|
log "github.com/sirupsen/logrus"
|
|
"github.com/spf13/viper"
|
|
)
|
|
|
|
// DBBackedStore is a store that's backed by a DB. The DB contains data about what's in the store.
|
|
type DBBackedStore struct {
|
|
blobs BlobStore
|
|
db *db.SQL
|
|
blockedMu sync.RWMutex
|
|
blocked map[string]bool
|
|
deleteOnMiss bool
|
|
maxSize int
|
|
cleanerStop *stop.Group
|
|
name string
|
|
}
|
|
|
|
type DBBackedParams struct {
|
|
Name string `mapstructure:"name"`
|
|
Store BlobStore `mapstructure:"store"`
|
|
DB *db.SQL `mapstructure:"db"`
|
|
DeleteOnMiss bool `mapstructure:"delete_on_miss"`
|
|
MaxSize *int `mapstructure:"max_size"`
|
|
}
|
|
|
|
type DBBackedConfig struct {
|
|
Name string `mapstructure:"name"`
|
|
Store *viper.Viper
|
|
User string `mapstructure:"user"`
|
|
Password string `mapstructure:"password"`
|
|
Host string `mapstructure:"host"`
|
|
Port int `mapstructure:"port"`
|
|
Database string `mapstructure:"database"`
|
|
DeleteOnMiss bool `mapstructure:"delete_on_miss"`
|
|
AccessTracking int `mapstructure:"access_tracking"`
|
|
SoftDeletes bool `mapstructure:"soft_deletes"`
|
|
LogQueries bool `mapstructure:"log_queries"`
|
|
HasCap bool `mapstructure:"has_cap"`
|
|
MaxSize string `mapstructure:"max_size"`
|
|
}
|
|
|
|
// NewDBBackedStore returns an initialized store pointer.
|
|
func NewDBBackedStore(params DBBackedParams) *DBBackedStore {
|
|
store := &DBBackedStore{
|
|
blobs: params.Store,
|
|
db: params.DB,
|
|
deleteOnMiss: params.DeleteOnMiss,
|
|
cleanerStop: stop.New(),
|
|
name: params.Name,
|
|
}
|
|
if params.MaxSize != nil {
|
|
store.maxSize = *params.MaxSize
|
|
go store.cleanOldestBlobs()
|
|
}
|
|
return store
|
|
}
|
|
|
|
const nameDBBacked = "db_backed"
|
|
|
|
// Name is the cache type name
|
|
func (d *DBBackedStore) Name() string { return nameDBBacked + "-" + d.name }
|
|
|
|
// Has returns true if the blob is in the store
|
|
func (d *DBBackedStore) Has(hash string) (bool, error) {
|
|
return d.db.HasBlob(hash, false)
|
|
}
|
|
|
|
// Get gets the blob
|
|
func (d *DBBackedStore) Get(hash string) (stream.Blob, shared.BlobTrace, error) {
|
|
start := time.Now()
|
|
has, err := d.db.HasBlob(hash, true)
|
|
if err != nil {
|
|
return nil, shared.NewBlobTrace(time.Since(start), d.Name()), err
|
|
}
|
|
if !has {
|
|
return nil, shared.NewBlobTrace(time.Since(start), d.Name()), ErrBlobNotFound
|
|
}
|
|
|
|
b, stack, err := d.blobs.Get(hash)
|
|
if d.deleteOnMiss && errors.Is(err, ErrBlobNotFound) {
|
|
e2 := d.Delete(hash)
|
|
if e2 != nil {
|
|
log.Errorf("error while deleting blob from db: %s", errors.FullTrace(err))
|
|
}
|
|
}
|
|
|
|
return b, stack.Stack(time.Since(start), d.Name()), err
|
|
}
|
|
|
|
// Put stores the blob in the S3 store and stores the blob information in the DB.
|
|
func (d *DBBackedStore) Put(hash string, blob stream.Blob) error {
|
|
err := d.blobs.Put(hash, blob)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
return d.db.AddBlob(hash, len(blob), true)
|
|
}
|
|
|
|
// PutSD stores the SDBlob in the S3 store. It will return an error if the sd blob is missing the stream hash or if
|
|
// there is an error storing the blob information in the DB.
|
|
func (d *DBBackedStore) PutSD(hash string, blob stream.Blob) error {
|
|
var blobContents db.SdBlob
|
|
err := json.Unmarshal(blob, &blobContents)
|
|
if err != nil {
|
|
return errors.Err(err)
|
|
}
|
|
if blobContents.StreamHash == "" {
|
|
return errors.Err("sd blob is missing stream hash")
|
|
}
|
|
|
|
err = d.blobs.PutSD(hash, blob)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
return d.db.AddSDBlob(hash, len(blob), blobContents)
|
|
}
|
|
|
|
func (d *DBBackedStore) Delete(hash string) error {
|
|
err := d.blobs.Delete(hash)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
return d.db.Delete(hash)
|
|
}
|
|
|
|
// Block deletes the blob and prevents it from being uploaded in the future
|
|
func (d *DBBackedStore) Block(hash string) error {
|
|
blocked, err := d.isBlocked(hash)
|
|
if blocked || err != nil {
|
|
return err
|
|
}
|
|
|
|
log.Debugf("blocking %s", hash)
|
|
|
|
err = d.db.Block(hash)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
return d.markBlocked(hash)
|
|
}
|
|
|
|
// Wants returns false if the hash exists or is blocked, true otherwise
|
|
func (d *DBBackedStore) Wants(hash string) (bool, error) {
|
|
blocked, err := d.isBlocked(hash)
|
|
if blocked || err != nil {
|
|
return false, err
|
|
}
|
|
|
|
has, err := d.Has(hash)
|
|
return !has, err
|
|
}
|
|
|
|
// MissingBlobsForKnownStream returns missing blobs for an existing stream
|
|
// WARNING: if the stream does NOT exist, no blob hashes will be returned, which looks
|
|
// like no blobs are missing
|
|
func (d *DBBackedStore) MissingBlobsForKnownStream(sdHash string) ([]string, error) {
|
|
return d.db.MissingBlobsForKnownStream(sdHash)
|
|
}
|
|
|
|
func (d *DBBackedStore) markBlocked(hash string) error {
|
|
err := d.initBlocked()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
d.blockedMu.Lock()
|
|
defer d.blockedMu.Unlock()
|
|
|
|
d.blocked[hash] = true
|
|
return nil
|
|
}
|
|
|
|
func (d *DBBackedStore) isBlocked(hash string) (bool, error) {
|
|
err := d.initBlocked()
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
|
|
d.blockedMu.RLock()
|
|
defer d.blockedMu.RUnlock()
|
|
|
|
return d.blocked[hash], nil
|
|
}
|
|
|
|
func (d *DBBackedStore) initBlocked() error {
|
|
// first check without blocking since this is the most likely scenario
|
|
if d.blocked != nil {
|
|
return nil
|
|
}
|
|
|
|
d.blockedMu.Lock()
|
|
defer d.blockedMu.Unlock()
|
|
|
|
// check again in case of race condition
|
|
if d.blocked != nil {
|
|
return nil
|
|
}
|
|
|
|
var err error
|
|
d.blocked, err = d.db.GetBlocked()
|
|
|
|
return err
|
|
}
|
|
|
|
// cleanOldestBlobs periodically cleans up the oldest blobs if maxSize is set
|
|
func (d *DBBackedStore) cleanOldestBlobs() {
|
|
// Run on startup without waiting for 10 minutes
|
|
err := d.doClean()
|
|
if err != nil {
|
|
log.Error(errors.FullTrace(err))
|
|
}
|
|
const cleanupInterval = 10 * time.Minute
|
|
for {
|
|
select {
|
|
case <-d.cleanerStop.Ch():
|
|
log.Infoln("stopping self cleanup")
|
|
return
|
|
case <-time.After(cleanupInterval):
|
|
err := d.doClean()
|
|
if err != nil {
|
|
log.Error(errors.FullTrace(err))
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// doClean removes the least recently accessed blobs if the store exceeds maxItems
|
|
func (d *DBBackedStore) doClean() error {
|
|
blobsCount, err := d.db.Count()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
if blobsCount >= d.maxSize {
|
|
itemsToDelete := blobsCount / 10
|
|
blobs, err := d.db.LeastRecentlyAccessedHashes(itemsToDelete)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
blobsChan := make(chan string, len(blobs))
|
|
wg := stop.New()
|
|
go func() {
|
|
for _, hash := range blobs {
|
|
select {
|
|
case <-d.cleanerStop.Ch():
|
|
return
|
|
default:
|
|
}
|
|
blobsChan <- hash
|
|
}
|
|
close(blobsChan)
|
|
}()
|
|
for i := 0; i < 3; i++ {
|
|
wg.Add(1)
|
|
go func() {
|
|
defer wg.Done()
|
|
for h := range blobsChan {
|
|
select {
|
|
case <-d.cleanerStop.Ch():
|
|
return
|
|
default:
|
|
}
|
|
err = d.Delete(h)
|
|
if err != nil {
|
|
log.Errorf("error pruning %s: %s", h, errors.FullTrace(err))
|
|
continue
|
|
}
|
|
}
|
|
}()
|
|
}
|
|
wg.Wait()
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// Shutdown shuts down the store gracefully
|
|
func (d *DBBackedStore) Shutdown() {
|
|
d.cleanerStop.Stop()
|
|
d.blobs.Shutdown()
|
|
}
|
|
|
|
func DBBackedStoreFactory(config *viper.Viper) (BlobStore, error) {
|
|
var cfg DBBackedConfig
|
|
err := config.Unmarshal(&cfg)
|
|
if err != nil {
|
|
return nil, errors.Err(err)
|
|
}
|
|
|
|
cfg.Store = config.Sub("store")
|
|
|
|
storeType := strings.Split(cfg.Store.AllKeys()[0], ".")[0]
|
|
storeConfig := cfg.Store.Sub(storeType)
|
|
factory, ok := Factories[storeType]
|
|
if !ok {
|
|
return nil, errors.Err("unknown store type %s", storeType)
|
|
}
|
|
underlyingStore, err := factory(storeConfig)
|
|
if err != nil {
|
|
return nil, errors.Err(err)
|
|
}
|
|
|
|
parsedDb := &db.SQL{
|
|
TrackingLevel: db.AccessTrackingLevel(cfg.AccessTracking),
|
|
SoftDelete: cfg.SoftDeletes,
|
|
LogQueries: cfg.LogQueries || log.GetLevel() == log.DebugLevel,
|
|
}
|
|
|
|
err = parsedDb.Connect(fmt.Sprintf("%s:%s@tcp(%s:%d)/%s", cfg.User, cfg.Password, cfg.Host, cfg.Port, cfg.Database))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
params := DBBackedParams{
|
|
Name: cfg.Name,
|
|
Store: underlyingStore,
|
|
DB: parsedDb,
|
|
DeleteOnMiss: cfg.DeleteOnMiss,
|
|
}
|
|
if cfg.HasCap {
|
|
var parsedSize datasize.ByteSize
|
|
err = parsedSize.UnmarshalText([]byte(cfg.MaxSize))
|
|
if err != nil {
|
|
return nil, errors.Err(err)
|
|
}
|
|
maxSize := int(float64(parsedSize) / float64(stream.MaxBlobSize))
|
|
params.MaxSize = &maxSize
|
|
}
|
|
|
|
return NewDBBackedStore(params), nil
|
|
}
|
|
|
|
func init() {
|
|
RegisterStore(nameDBBacked, DBBackedStoreFactory)
|
|
}
|