forked from TrueCloudLab/frostfs-node
[#1636] storage: Refactor shard rebuild
Drop redundant interfaces. Rename fields. Signed-off-by: Dmitrii Stepanov <d.stepanov@yadro.com>
This commit is contained in:
parent
e0dc3c3d0c
commit
ceff5e1f6a
7 changed files with 72 additions and 79 deletions
|
@ -8,6 +8,7 @@ import (
|
|||
"git.frostfs.info/TrueCloudLab/frostfs-node/internal/logs"
|
||||
"git.frostfs.info/TrueCloudLab/frostfs-node/internal/qos"
|
||||
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/local_object_storage/blobstor"
|
||||
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/local_object_storage/blobstor/common"
|
||||
meta "git.frostfs.info/TrueCloudLab/frostfs-node/pkg/local_object_storage/metabase"
|
||||
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/util/logger"
|
||||
"git.frostfs.info/TrueCloudLab/frostfs-observability/tracing"
|
||||
|
@ -20,37 +21,9 @@ import (
|
|||
|
||||
var ErrRebuildInProgress = errors.New("shard rebuild in progress")
|
||||
|
||||
type RebuildWorkerLimiter interface {
|
||||
AcquireWorkSlot(ctx context.Context) error
|
||||
ReleaseWorkSlot()
|
||||
}
|
||||
|
||||
type rebuildLimiter struct {
|
||||
semaphore chan struct{}
|
||||
}
|
||||
|
||||
func NewRebuildLimiter(workersCount uint32) RebuildWorkerLimiter {
|
||||
return &rebuildLimiter{
|
||||
semaphore: make(chan struct{}, workersCount),
|
||||
}
|
||||
}
|
||||
|
||||
func (l *rebuildLimiter) AcquireWorkSlot(ctx context.Context) error {
|
||||
select {
|
||||
case l.semaphore <- struct{}{}:
|
||||
return nil
|
||||
case <-ctx.Done():
|
||||
return ctx.Err()
|
||||
}
|
||||
}
|
||||
|
||||
func (l *rebuildLimiter) ReleaseWorkSlot() {
|
||||
<-l.semaphore
|
||||
}
|
||||
|
||||
type rebuildTask struct {
|
||||
limiter RebuildWorkerLimiter
|
||||
fillPercent int
|
||||
concurrencyLimiter common.RebuildLimiter
|
||||
fillPercent int
|
||||
}
|
||||
|
||||
type rebuilder struct {
|
||||
|
@ -90,14 +63,14 @@ func (r *rebuilder) Start(ctx context.Context, bs *blobstor.BlobStor, mb *meta.D
|
|||
if !ok {
|
||||
continue
|
||||
}
|
||||
runRebuild(ctx, bs, mb, log, t.fillPercent, t.limiter)
|
||||
runRebuild(ctx, bs, mb, log, t.fillPercent, t.concurrencyLimiter)
|
||||
}
|
||||
}
|
||||
}()
|
||||
}
|
||||
|
||||
func runRebuild(ctx context.Context, bs *blobstor.BlobStor, mb *meta.DB, log *logger.Logger,
|
||||
fillPercent int, limiter RebuildWorkerLimiter,
|
||||
fillPercent int, concLimiter common.RebuildLimiter,
|
||||
) {
|
||||
select {
|
||||
case <-ctx.Done():
|
||||
|
@ -107,21 +80,21 @@ func runRebuild(ctx context.Context, bs *blobstor.BlobStor, mb *meta.DB, log *lo
|
|||
log.Info(ctx, logs.BlobstoreRebuildStarted)
|
||||
ctx = tagging.ContextWithIOTag(ctx, qos.IOTagBackground.String())
|
||||
// TODO use shard limiter
|
||||
if err := bs.Rebuild(ctx, &mbStorageIDUpdate{mb: mb}, limiter, fillPercent); err != nil {
|
||||
if err := bs.Rebuild(ctx, &mbStorageIDUpdate{mb: mb}, concLimiter, fillPercent); err != nil {
|
||||
log.Warn(ctx, logs.FailedToRebuildBlobstore, zap.Error(err))
|
||||
} else {
|
||||
log.Info(ctx, logs.BlobstoreRebuildCompletedSuccessfully)
|
||||
}
|
||||
}
|
||||
|
||||
func (r *rebuilder) ScheduleRebuild(ctx context.Context, limiter RebuildWorkerLimiter, fillPercent int,
|
||||
func (r *rebuilder) ScheduleRebuild(ctx context.Context, limiter common.RebuildLimiter, fillPercent int,
|
||||
) error {
|
||||
select {
|
||||
case <-ctx.Done():
|
||||
return ctx.Err()
|
||||
case r.tasks <- rebuildTask{
|
||||
limiter: limiter,
|
||||
fillPercent: fillPercent,
|
||||
concurrencyLimiter: limiter,
|
||||
fillPercent: fillPercent,
|
||||
}:
|
||||
return nil
|
||||
default:
|
||||
|
@ -170,7 +143,7 @@ func (u *mbStorageIDUpdate) UpdateStorageID(ctx context.Context, addr oid.Addres
|
|||
}
|
||||
|
||||
type RebuildPrm struct {
|
||||
ConcurrencyLimiter RebuildWorkerLimiter
|
||||
ConcurrencyLimiter common.ConcurrencyLimiter
|
||||
TargetFillPercent uint32
|
||||
}
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue