forked from TrueCloudLab/frostfs-node
[#384] shard: Cancel GC if change mode requested
Signed-off-by: Dmitrii Stepanov <d.stepanov@yadro.com>
This commit is contained in:
parent
1501f11e4d
commit
1dd84eef77
8 changed files with 98 additions and 63 deletions
|
@ -128,7 +128,7 @@ func (e *StorageEngine) removeObjects(ctx context.Context, ch <-chan oid.Address
|
||||||
|
|
||||||
var deletePrm shard.DeletePrm
|
var deletePrm shard.DeletePrm
|
||||||
deletePrm.SetAddresses(addr)
|
deletePrm.SetAddresses(addr)
|
||||||
_, err = shards[i].Delete(deletePrm)
|
_, err = shards[i].Delete(ctx, deletePrm)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
|
@ -296,8 +296,8 @@ func (s *Shard) Reload(opts ...Option) error {
|
||||||
opts[i](&c)
|
opts[i](&c)
|
||||||
}
|
}
|
||||||
|
|
||||||
s.m.Lock()
|
unlock := s.lockExclusive()
|
||||||
defer s.m.Unlock()
|
defer unlock()
|
||||||
|
|
||||||
ok, err := s.metaBase.Reload(c.metaOpts...)
|
ok, err := s.metaBase.Reload(c.metaOpts...)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -327,3 +327,15 @@ func (s *Shard) Reload(opts ...Option) error {
|
||||||
s.log.Info("trying to restore read-write mode")
|
s.log.Info("trying to restore read-write mode")
|
||||||
return s.setMode(mode.ReadWrite)
|
return s.setMode(mode.ReadWrite)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (s *Shard) lockExclusive() func() {
|
||||||
|
s.setModeRequested.Store(true)
|
||||||
|
val := s.gcCancel.Load()
|
||||||
|
if val != nil {
|
||||||
|
cancelGC := val.(context.CancelFunc)
|
||||||
|
cancelGC()
|
||||||
|
}
|
||||||
|
s.m.Lock()
|
||||||
|
s.setModeRequested.Store(false)
|
||||||
|
return s.m.Unlock
|
||||||
|
}
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
package shard
|
package shard
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"errors"
|
"errors"
|
||||||
|
|
||||||
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/local_object_storage/blobstor/common"
|
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/local_object_storage/blobstor/common"
|
||||||
|
@ -27,84 +28,88 @@ func (p *DeletePrm) SetAddresses(addr ...oid.Address) {
|
||||||
|
|
||||||
// Delete removes data from the shard's writeCache, metaBase and
|
// Delete removes data from the shard's writeCache, metaBase and
|
||||||
// blobStor.
|
// blobStor.
|
||||||
func (s *Shard) Delete(prm DeletePrm) (DeleteRes, error) {
|
func (s *Shard) Delete(ctx context.Context, prm DeletePrm) (DeleteRes, error) {
|
||||||
s.m.RLock()
|
s.m.RLock()
|
||||||
defer s.m.RUnlock()
|
defer s.m.RUnlock()
|
||||||
|
|
||||||
return s.delete(prm)
|
return s.delete(ctx, prm)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Shard) delete(prm DeletePrm) (DeleteRes, error) {
|
func (s *Shard) delete(ctx context.Context, prm DeletePrm) (DeleteRes, error) {
|
||||||
if s.info.Mode.ReadOnly() {
|
if s.info.Mode.ReadOnly() {
|
||||||
return DeleteRes{}, ErrReadOnlyMode
|
return DeleteRes{}, ErrReadOnlyMode
|
||||||
} else if s.info.Mode.NoMetabase() {
|
} else if s.info.Mode.NoMetabase() {
|
||||||
return DeleteRes{}, ErrDegradedMode
|
return DeleteRes{}, ErrDegradedMode
|
||||||
}
|
}
|
||||||
|
|
||||||
ln := len(prm.addr)
|
for _, addr := range prm.addr {
|
||||||
|
select {
|
||||||
smalls := make(map[oid.Address][]byte, ln)
|
case <-ctx.Done():
|
||||||
|
return DeleteRes{}, ctx.Err()
|
||||||
for i := range prm.addr {
|
default:
|
||||||
if s.hasWriteCache() {
|
|
||||||
err := s.writeCache.Delete(prm.addr[i])
|
|
||||||
if err != nil && !IsErrNotFound(err) && !errors.Is(err, writecache.ErrReadOnly) {
|
|
||||||
s.log.Warn("can't delete object from write cache", zap.String("error", err.Error()))
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
var sPrm meta.StorageIDPrm
|
s.deleteObjectFromWriteCacheSafe(addr)
|
||||||
sPrm.SetAddress(prm.addr[i])
|
|
||||||
|
|
||||||
res, err := s.metaBase.StorageID(sPrm)
|
s.deleteFromBlobstorSafe(addr)
|
||||||
if err != nil {
|
|
||||||
s.log.Debug("can't get storage ID from metabase",
|
|
||||||
zap.Stringer("object", prm.addr[i]),
|
|
||||||
zap.String("error", err.Error()))
|
|
||||||
|
|
||||||
continue
|
if err := s.deleteFromMetabase(addr); err != nil {
|
||||||
}
|
return DeleteRes{}, err // stop on metabase error ?
|
||||||
|
|
||||||
if res.StorageID() != nil {
|
|
||||||
smalls[prm.addr[i]] = res.StorageID()
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
return DeleteRes{}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Shard) deleteObjectFromWriteCacheSafe(addr oid.Address) {
|
||||||
|
if s.hasWriteCache() {
|
||||||
|
err := s.writeCache.Delete(addr)
|
||||||
|
if err != nil && !IsErrNotFound(err) && !errors.Is(err, writecache.ErrReadOnly) {
|
||||||
|
s.log.Warn("can't delete object from write cache", zap.String("error", err.Error()))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Shard) deleteFromMetabase(addr oid.Address) error {
|
||||||
var delPrm meta.DeletePrm
|
var delPrm meta.DeletePrm
|
||||||
delPrm.SetAddresses(prm.addr...)
|
delPrm.SetAddresses(addr)
|
||||||
|
|
||||||
res, err := s.metaBase.Delete(delPrm)
|
res, err := s.metaBase.Delete(delPrm)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return DeleteRes{}, err // stop on metabase error ?
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
var totalRemovedPayload uint64
|
|
||||||
|
|
||||||
s.decObjectCounterBy(physical, res.RawObjectsRemoved())
|
s.decObjectCounterBy(physical, res.RawObjectsRemoved())
|
||||||
s.decObjectCounterBy(logical, res.AvailableObjectsRemoved())
|
s.decObjectCounterBy(logical, res.AvailableObjectsRemoved())
|
||||||
for i := range prm.addr {
|
removedPayload := res.RemovedPhysicalObjectSizes()[0]
|
||||||
removedPayload := res.RemovedPhysicalObjectSizes()[i]
|
logicalRemovedPayload := res.RemovedLogicalObjectSizes()[0]
|
||||||
totalRemovedPayload += removedPayload
|
if logicalRemovedPayload > 0 {
|
||||||
logicalRemovedPayload := res.RemovedLogicalObjectSizes()[i]
|
s.addToContainerSize(addr.Container().EncodeToString(), -int64(logicalRemovedPayload))
|
||||||
if logicalRemovedPayload > 0 {
|
|
||||||
s.addToContainerSize(prm.addr[i].Container().EncodeToString(), -int64(logicalRemovedPayload))
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
s.addToPayloadSize(-int64(totalRemovedPayload))
|
s.addToPayloadSize(-int64(removedPayload))
|
||||||
|
|
||||||
for i := range prm.addr {
|
return nil
|
||||||
var delPrm common.DeletePrm
|
}
|
||||||
delPrm.Address = prm.addr[i]
|
|
||||||
id := smalls[prm.addr[i]]
|
func (s *Shard) deleteFromBlobstorSafe(addr oid.Address) {
|
||||||
delPrm.StorageID = id
|
var sPrm meta.StorageIDPrm
|
||||||
|
sPrm.SetAddress(addr)
|
||||||
_, err = s.blobStor.Delete(delPrm)
|
|
||||||
if err != nil {
|
res, err := s.metaBase.StorageID(sPrm)
|
||||||
s.log.Debug("can't remove object from blobStor",
|
if err != nil {
|
||||||
zap.Stringer("object_address", prm.addr[i]),
|
s.log.Debug("can't get storage ID from metabase",
|
||||||
zap.String("error", err.Error()))
|
zap.Stringer("object", addr),
|
||||||
}
|
zap.String("error", err.Error()))
|
||||||
}
|
}
|
||||||
|
storageID := res.StorageID()
|
||||||
return DeleteRes{}, nil
|
|
||||||
|
var delPrm common.DeletePrm
|
||||||
|
delPrm.Address = addr
|
||||||
|
delPrm.StorageID = storageID
|
||||||
|
|
||||||
|
_, err = s.blobStor.Delete(delPrm)
|
||||||
|
if err != nil {
|
||||||
|
s.log.Debug("can't remove object from blobStor",
|
||||||
|
zap.Stringer("object_address", addr),
|
||||||
|
zap.String("error", err.Error()))
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -49,7 +49,7 @@ func testShardDelete(t *testing.T, hasWriteCache bool) {
|
||||||
_, err = testGet(t, sh, getPrm, hasWriteCache)
|
_, err = testGet(t, sh, getPrm, hasWriteCache)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
_, err = sh.Delete(delPrm)
|
_, err = sh.Delete(context.Background(), delPrm)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
_, err = sh.Get(context.Background(), getPrm)
|
_, err = sh.Get(context.Background(), getPrm)
|
||||||
|
@ -73,7 +73,7 @@ func testShardDelete(t *testing.T, hasWriteCache bool) {
|
||||||
_, err = sh.Get(context.Background(), getPrm)
|
_, err = sh.Get(context.Background(), getPrm)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
_, err = sh.Delete(delPrm)
|
_, err = sh.Delete(context.Background(), delPrm)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
_, err = sh.Get(context.Background(), getPrm)
|
_, err = sh.Get(context.Background(), getPrm)
|
||||||
|
|
|
@ -196,6 +196,14 @@ func (gc *gc) stop() {
|
||||||
// with GC-marked graves.
|
// with GC-marked graves.
|
||||||
// Does nothing if shard is in "read-only" mode.
|
// Does nothing if shard is in "read-only" mode.
|
||||||
func (s *Shard) removeGarbage() {
|
func (s *Shard) removeGarbage() {
|
||||||
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
|
defer cancel()
|
||||||
|
|
||||||
|
s.gcCancel.Store(cancel)
|
||||||
|
if s.setModeRequested.Load() {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
s.m.RLock()
|
s.m.RLock()
|
||||||
defer s.m.RUnlock()
|
defer s.m.RUnlock()
|
||||||
|
|
||||||
|
@ -207,6 +215,12 @@ func (s *Shard) removeGarbage() {
|
||||||
|
|
||||||
var iterPrm meta.GarbageIterationPrm
|
var iterPrm meta.GarbageIterationPrm
|
||||||
iterPrm.SetHandler(func(g meta.GarbageObject) error {
|
iterPrm.SetHandler(func(g meta.GarbageObject) error {
|
||||||
|
select {
|
||||||
|
case <-ctx.Done():
|
||||||
|
return ctx.Err()
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
|
||||||
buf = append(buf, g.Address())
|
buf = append(buf, g.Address())
|
||||||
|
|
||||||
if len(buf) == s.rmBatchSize {
|
if len(buf) == s.rmBatchSize {
|
||||||
|
@ -233,7 +247,7 @@ func (s *Shard) removeGarbage() {
|
||||||
deletePrm.SetAddresses(buf...)
|
deletePrm.SetAddresses(buf...)
|
||||||
|
|
||||||
// delete accumulated objects
|
// delete accumulated objects
|
||||||
_, err = s.delete(deletePrm)
|
_, err = s.delete(ctx, deletePrm)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
s.log.Warn("could not delete the objects",
|
s.log.Warn("could not delete the objects",
|
||||||
zap.String("error", err.Error()),
|
zap.String("error", err.Error()),
|
||||||
|
|
|
@ -168,7 +168,7 @@ func TestCounters(t *testing.T) {
|
||||||
deletedNumber := int(phy / 4)
|
deletedNumber := int(phy / 4)
|
||||||
prm.SetAddresses(addrFromObjs(oo[:deletedNumber])...)
|
prm.SetAddresses(addrFromObjs(oo[:deletedNumber])...)
|
||||||
|
|
||||||
_, err := sh.Delete(prm)
|
_, err := sh.Delete(context.Background(), prm)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
require.Equal(t, phy-uint64(deletedNumber), mm.objCounters[physical])
|
require.Equal(t, phy-uint64(deletedNumber), mm.objCounters[physical])
|
||||||
|
|
|
@ -18,8 +18,8 @@ var ErrDegradedMode = logicerr.New("shard is in degraded mode")
|
||||||
// Returns any error encountered that did not allow
|
// Returns any error encountered that did not allow
|
||||||
// setting shard mode.
|
// setting shard mode.
|
||||||
func (s *Shard) SetMode(m mode.Mode) error {
|
func (s *Shard) SetMode(m mode.Mode) error {
|
||||||
s.m.Lock()
|
unlock := s.lockExclusive()
|
||||||
defer s.m.Unlock()
|
defer unlock()
|
||||||
|
|
||||||
return s.setMode(m)
|
return s.setMode(m)
|
||||||
}
|
}
|
||||||
|
|
|
@ -3,6 +3,7 @@ package shard
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"sync"
|
"sync"
|
||||||
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/local_object_storage/blobstor"
|
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/local_object_storage/blobstor"
|
||||||
|
@ -31,6 +32,9 @@ type Shard struct {
|
||||||
metaBase *meta.DB
|
metaBase *meta.DB
|
||||||
|
|
||||||
tsSource TombstoneSource
|
tsSource TombstoneSource
|
||||||
|
|
||||||
|
gcCancel atomic.Value
|
||||||
|
setModeRequested atomic.Bool
|
||||||
}
|
}
|
||||||
|
|
||||||
// Option represents Shard's constructor option.
|
// Option represents Shard's constructor option.
|
||||||
|
@ -209,12 +213,12 @@ func WithWriteCache(use bool) Option {
|
||||||
}
|
}
|
||||||
|
|
||||||
// hasWriteCache returns bool if write cache exists on shards.
|
// hasWriteCache returns bool if write cache exists on shards.
|
||||||
func (s Shard) hasWriteCache() bool {
|
func (s *Shard) hasWriteCache() bool {
|
||||||
return s.cfg.useWriteCache
|
return s.cfg.useWriteCache
|
||||||
}
|
}
|
||||||
|
|
||||||
// needRefillMetabase returns true if metabase is needed to be refilled.
|
// needRefillMetabase returns true if metabase is needed to be refilled.
|
||||||
func (s Shard) needRefillMetabase() bool {
|
func (s *Shard) needRefillMetabase() bool {
|
||||||
return s.cfg.refillMetabase
|
return s.cfg.refillMetabase
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue