forked from TrueCloudLab/frostfs-node
[#1226] blobovniczatree: Drop leaf width limitation
Signed-off-by: Dmitrii Stepanov <d.stepanov@yadro.com>
This commit is contained in:
parent
40c9ddb6ba
commit
78b1d9b18d
10 changed files with 50 additions and 55 deletions
|
@ -188,7 +188,6 @@ type subStorageCfg struct {
|
||||||
// blobovnicza-specific
|
// blobovnicza-specific
|
||||||
size uint64
|
size uint64
|
||||||
width uint64
|
width uint64
|
||||||
leafWidth uint64
|
|
||||||
openedCacheSize int
|
openedCacheSize int
|
||||||
initWorkerCount int
|
initWorkerCount int
|
||||||
rebuildDropTimeout time.Duration
|
rebuildDropTimeout time.Duration
|
||||||
|
@ -313,7 +312,6 @@ func (a *applicationConfiguration) setShardStorageConfig(newConfig *shardCfg, ol
|
||||||
sCfg.size = sub.Size()
|
sCfg.size = sub.Size()
|
||||||
sCfg.depth = sub.ShallowDepth()
|
sCfg.depth = sub.ShallowDepth()
|
||||||
sCfg.width = sub.ShallowWidth()
|
sCfg.width = sub.ShallowWidth()
|
||||||
sCfg.leafWidth = sub.LeafWidth()
|
|
||||||
sCfg.openedCacheSize = sub.OpenedCacheSize()
|
sCfg.openedCacheSize = sub.OpenedCacheSize()
|
||||||
sCfg.openedCacheTTL = sub.OpenedCacheTTL()
|
sCfg.openedCacheTTL = sub.OpenedCacheTTL()
|
||||||
sCfg.openedCacheExpInterval = sub.OpenedCacheExpInterval()
|
sCfg.openedCacheExpInterval = sub.OpenedCacheExpInterval()
|
||||||
|
@ -904,7 +902,6 @@ func (c *cfg) getSubstorageOpts(ctx context.Context, shCfg shardCfg) []blobstor.
|
||||||
blobovniczatree.WithBlobovniczaSize(sRead.size),
|
blobovniczatree.WithBlobovniczaSize(sRead.size),
|
||||||
blobovniczatree.WithBlobovniczaShallowDepth(sRead.depth),
|
blobovniczatree.WithBlobovniczaShallowDepth(sRead.depth),
|
||||||
blobovniczatree.WithBlobovniczaShallowWidth(sRead.width),
|
blobovniczatree.WithBlobovniczaShallowWidth(sRead.width),
|
||||||
blobovniczatree.WithBlobovniczaLeafWidth(sRead.leafWidth),
|
|
||||||
blobovniczatree.WithOpenedCacheSize(sRead.openedCacheSize),
|
blobovniczatree.WithOpenedCacheSize(sRead.openedCacheSize),
|
||||||
blobovniczatree.WithOpenedCacheTTL(sRead.openedCacheTTL),
|
blobovniczatree.WithOpenedCacheTTL(sRead.openedCacheTTL),
|
||||||
blobovniczatree.WithOpenedCacheExpInterval(sRead.openedCacheExpInterval),
|
blobovniczatree.WithOpenedCacheExpInterval(sRead.openedCacheExpInterval),
|
||||||
|
|
|
@ -100,7 +100,6 @@ func TestEngineSection(t *testing.T) {
|
||||||
require.EqualValues(t, 50, blz.OpenedCacheSize())
|
require.EqualValues(t, 50, blz.OpenedCacheSize())
|
||||||
require.EqualValues(t, time.Minute, blz.OpenedCacheTTL())
|
require.EqualValues(t, time.Minute, blz.OpenedCacheTTL())
|
||||||
require.EqualValues(t, 30*time.Second, blz.OpenedCacheExpInterval())
|
require.EqualValues(t, 30*time.Second, blz.OpenedCacheExpInterval())
|
||||||
require.EqualValues(t, 10, blz.LeafWidth())
|
|
||||||
require.EqualValues(t, 10, blz.InitWorkerCount())
|
require.EqualValues(t, 10, blz.InitWorkerCount())
|
||||||
require.EqualValues(t, 30*time.Second, blz.RebuildDropTimeout())
|
require.EqualValues(t, 30*time.Second, blz.RebuildDropTimeout())
|
||||||
|
|
||||||
|
@ -154,7 +153,6 @@ func TestEngineSection(t *testing.T) {
|
||||||
require.EqualValues(t, 50, blz.OpenedCacheSize())
|
require.EqualValues(t, 50, blz.OpenedCacheSize())
|
||||||
require.EqualValues(t, 5*time.Minute, blz.OpenedCacheTTL())
|
require.EqualValues(t, 5*time.Minute, blz.OpenedCacheTTL())
|
||||||
require.EqualValues(t, 15*time.Second, blz.OpenedCacheExpInterval())
|
require.EqualValues(t, 15*time.Second, blz.OpenedCacheExpInterval())
|
||||||
require.EqualValues(t, 10, blz.LeafWidth())
|
|
||||||
require.EqualValues(t, blobovniczaconfig.InitWorkerCountDefault, blz.InitWorkerCount())
|
require.EqualValues(t, blobovniczaconfig.InitWorkerCountDefault, blz.InitWorkerCount())
|
||||||
require.EqualValues(t, blobovniczaconfig.RebuildDropTimeoutDefault, blz.RebuildDropTimeout())
|
require.EqualValues(t, blobovniczaconfig.RebuildDropTimeoutDefault, blz.RebuildDropTimeout())
|
||||||
|
|
||||||
|
|
|
@ -149,16 +149,6 @@ func (x *Config) BoltDB() *boltdbconfig.Config {
|
||||||
return (*boltdbconfig.Config)(x)
|
return (*boltdbconfig.Config)(x)
|
||||||
}
|
}
|
||||||
|
|
||||||
// LeafWidth returns the value of "leaf_width" config parameter.
|
|
||||||
//
|
|
||||||
// Returns 0 if the value is not a positive number.
|
|
||||||
func (x *Config) LeafWidth() uint64 {
|
|
||||||
return config.UintSafe(
|
|
||||||
(*config.Config)(x),
|
|
||||||
"leaf_width",
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
// InitWorkerCount returns the value of "init_worker_count" config parameter.
|
// InitWorkerCount returns the value of "init_worker_count" config parameter.
|
||||||
//
|
//
|
||||||
// Returns InitWorkerCountDefault if the value is not a positive number.
|
// Returns InitWorkerCountDefault if the value is not a positive number.
|
||||||
|
|
|
@ -126,7 +126,6 @@ FROSTFS_STORAGE_SHARD_0_BLOBSTOR_0_WIDTH=4
|
||||||
FROSTFS_STORAGE_SHARD_0_BLOBSTOR_0_OPENED_CACHE_CAPACITY=50
|
FROSTFS_STORAGE_SHARD_0_BLOBSTOR_0_OPENED_CACHE_CAPACITY=50
|
||||||
FROSTFS_STORAGE_SHARD_0_BLOBSTOR_0_OPENED_CACHE_TTL=1m
|
FROSTFS_STORAGE_SHARD_0_BLOBSTOR_0_OPENED_CACHE_TTL=1m
|
||||||
FROSTFS_STORAGE_SHARD_0_BLOBSTOR_0_OPENED_CACHE_EXP_INTERVAL=30s
|
FROSTFS_STORAGE_SHARD_0_BLOBSTOR_0_OPENED_CACHE_EXP_INTERVAL=30s
|
||||||
FROSTFS_STORAGE_SHARD_0_BLOBSTOR_0_LEAF_WIDTH=10
|
|
||||||
FROSTFS_STORAGE_SHARD_0_BLOBSTOR_0_INIT_WORKER_COUNT=10
|
FROSTFS_STORAGE_SHARD_0_BLOBSTOR_0_INIT_WORKER_COUNT=10
|
||||||
FROSTFS_STORAGE_SHARD_0_BLOBSTOR_0_REBUILD_DROP_TIMEOUT=30s
|
FROSTFS_STORAGE_SHARD_0_BLOBSTOR_0_REBUILD_DROP_TIMEOUT=30s
|
||||||
### FSTree config
|
### FSTree config
|
||||||
|
@ -177,7 +176,6 @@ FROSTFS_STORAGE_SHARD_1_BLOBSTOR_0_WIDTH=4
|
||||||
FROSTFS_STORAGE_SHARD_1_BLOBSTOR_0_OPENED_CACHE_CAPACITY=50
|
FROSTFS_STORAGE_SHARD_1_BLOBSTOR_0_OPENED_CACHE_CAPACITY=50
|
||||||
FROSTFS_STORAGE_SHARD_1_BLOBSTOR_0_OPENED_CACHE_TTL=5m
|
FROSTFS_STORAGE_SHARD_1_BLOBSTOR_0_OPENED_CACHE_TTL=5m
|
||||||
FROSTFS_STORAGE_SHARD_1_BLOBSTOR_0_OPENED_CACHE_EXP_INTERVAL=15s
|
FROSTFS_STORAGE_SHARD_1_BLOBSTOR_0_OPENED_CACHE_EXP_INTERVAL=15s
|
||||||
FROSTFS_STORAGE_SHARD_1_BLOBSTOR_0_LEAF_WIDTH=10
|
|
||||||
### FSTree config
|
### FSTree config
|
||||||
FROSTFS_STORAGE_SHARD_1_BLOBSTOR_1_TYPE=fstree
|
FROSTFS_STORAGE_SHARD_1_BLOBSTOR_1_TYPE=fstree
|
||||||
FROSTFS_STORAGE_SHARD_1_BLOBSTOR_1_PATH=tmp/1/blob
|
FROSTFS_STORAGE_SHARD_1_BLOBSTOR_1_PATH=tmp/1/blob
|
||||||
|
|
|
@ -174,7 +174,6 @@
|
||||||
"opened_cache_capacity": 50,
|
"opened_cache_capacity": 50,
|
||||||
"opened_cache_ttl": "1m",
|
"opened_cache_ttl": "1m",
|
||||||
"opened_cache_exp_interval": "30s",
|
"opened_cache_exp_interval": "30s",
|
||||||
"leaf_width": 10,
|
|
||||||
"init_worker_count": 10,
|
"init_worker_count": 10,
|
||||||
"rebuild_drop_timeout": "30s"
|
"rebuild_drop_timeout": "30s"
|
||||||
},
|
},
|
||||||
|
@ -227,8 +226,7 @@
|
||||||
"width": 4,
|
"width": 4,
|
||||||
"opened_cache_capacity": 50,
|
"opened_cache_capacity": 50,
|
||||||
"opened_cache_ttl": "5m",
|
"opened_cache_ttl": "5m",
|
||||||
"opened_cache_exp_interval": "15s",
|
"opened_cache_exp_interval": "15s"
|
||||||
"leaf_width": 10
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"type": "fstree",
|
"type": "fstree",
|
||||||
|
|
|
@ -149,7 +149,6 @@ storage:
|
||||||
opened_cache_capacity: 50 # maximum number of opened database files
|
opened_cache_capacity: 50 # maximum number of opened database files
|
||||||
opened_cache_ttl: 5m # ttl for opened database file
|
opened_cache_ttl: 5m # ttl for opened database file
|
||||||
opened_cache_exp_interval: 15s # cache cleanup interval for expired blobovnicza's
|
opened_cache_exp_interval: 15s # cache cleanup interval for expired blobovnicza's
|
||||||
leaf_width: 10 # max count of key-value DB on leafs of object tree storage
|
|
||||||
- perm: 0644 # permissions for blobstor files(directories: +x for current user and group)
|
- perm: 0644 # permissions for blobstor files(directories: +x for current user and group)
|
||||||
depth: 5 # max depth of object tree storage in FS
|
depth: 5 # max depth of object tree storage in FS
|
||||||
|
|
||||||
|
|
|
@ -37,17 +37,17 @@ type activeDBManager struct {
|
||||||
closed bool
|
closed bool
|
||||||
|
|
||||||
dbManager *dbManager
|
dbManager *dbManager
|
||||||
leafWidth uint64
|
rootPath string
|
||||||
}
|
}
|
||||||
|
|
||||||
func newActiveDBManager(dbManager *dbManager, leafWidth uint64) *activeDBManager {
|
func newActiveDBManager(dbManager *dbManager, rootPath string) *activeDBManager {
|
||||||
return &activeDBManager{
|
return &activeDBManager{
|
||||||
levelToActiveDBGuard: &sync.RWMutex{},
|
levelToActiveDBGuard: &sync.RWMutex{},
|
||||||
levelToActiveDB: make(map[string]*sharedDB),
|
levelToActiveDB: make(map[string]*sharedDB),
|
||||||
levelLock: utilSync.NewKeyLocker[string](),
|
levelLock: utilSync.NewKeyLocker[string](),
|
||||||
|
|
||||||
dbManager: dbManager,
|
dbManager: dbManager,
|
||||||
leafWidth: leafWidth,
|
rootPath: rootPath,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -144,30 +144,25 @@ func (m *activeDBManager) updateAndGetActive(lvlPath string) (*activeDB, error)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *activeDBManager) getNextSharedDB(lvlPath string) (*sharedDB, error) {
|
func (m *activeDBManager) getNextSharedDB(lvlPath string) (*sharedDB, error) {
|
||||||
var idx uint64
|
var nextActiveDBIdx uint64
|
||||||
var iterCount uint64
|
|
||||||
hasActive, currentIdx := m.hasActiveDB(lvlPath)
|
hasActive, currentIdx := m.hasActiveDB(lvlPath)
|
||||||
if hasActive {
|
if hasActive {
|
||||||
idx = (currentIdx + 1) % m.leafWidth
|
nextActiveDBIdx = currentIdx + 1
|
||||||
}
|
} else {
|
||||||
|
hasDBs, maxIdx, err := getBlobovniczaMaxIndex(filepath.Join(m.rootPath, lvlPath))
|
||||||
var next *sharedDB
|
|
||||||
|
|
||||||
for iterCount < m.leafWidth {
|
|
||||||
path := filepath.Join(lvlPath, u64ToHexStringExt(idx))
|
|
||||||
shDB := m.dbManager.GetByPath(path)
|
|
||||||
db, err := shDB.Open() // open db to hold active DB open, will be closed if db is full, after m.replace or by activeDBManager.Close()
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
if db.IsFull() {
|
if hasDBs {
|
||||||
shDB.Close()
|
nextActiveDBIdx = maxIdx
|
||||||
} else {
|
|
||||||
next = shDB
|
|
||||||
break
|
|
||||||
}
|
}
|
||||||
idx = (idx + 1) % m.leafWidth
|
}
|
||||||
iterCount++
|
|
||||||
|
path := filepath.Join(lvlPath, u64ToHexStringExt(nextActiveDBIdx))
|
||||||
|
next := m.dbManager.GetByPath(path)
|
||||||
|
_, err := next.Open() // open db to hold active DB open, will be closed if db is full, after m.replace or by activeDBManager.Close()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
previous, updated := m.replace(lvlPath, next)
|
previous, updated := m.replace(lvlPath, next)
|
||||||
|
|
|
@ -3,6 +3,7 @@ package blobovniczatree
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"errors"
|
"errors"
|
||||||
|
"os"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
|
@ -81,12 +82,8 @@ func NewBlobovniczaTree(ctx context.Context, opts ...Option) (blz *Blobovniczas)
|
||||||
opts[i](&blz.cfg)
|
opts[i](&blz.cfg)
|
||||||
}
|
}
|
||||||
|
|
||||||
if blz.blzLeafWidth == 0 {
|
|
||||||
blz.blzLeafWidth = blz.blzShallowWidth
|
|
||||||
}
|
|
||||||
|
|
||||||
blz.commondbManager = newDBManager(blz.rootPath, blz.blzOpts, blz.readOnly, blz.metrics.Blobovnicza(), blz.log)
|
blz.commondbManager = newDBManager(blz.rootPath, blz.blzOpts, blz.readOnly, blz.metrics.Blobovnicza(), blz.log)
|
||||||
blz.activeDBManager = newActiveDBManager(blz.commondbManager, blz.blzLeafWidth)
|
blz.activeDBManager = newActiveDBManager(blz.commondbManager, blz.rootPath)
|
||||||
blz.dbCache = newDBCache(ctx, blz.openedCacheSize,
|
blz.dbCache = newDBCache(ctx, blz.openedCacheSize,
|
||||||
blz.openedCacheTTL, blz.openedCacheExpInterval, blz.commondbManager)
|
blz.openedCacheTTL, blz.openedCacheExpInterval, blz.commondbManager)
|
||||||
blz.deleteProtectedObjects = newAddressMap()
|
blz.deleteProtectedObjects = newAddressMap()
|
||||||
|
@ -124,6 +121,29 @@ func u64FromHexString(str string) uint64 {
|
||||||
return v
|
return v
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func getBlobovniczaMaxIndex(directory string) (bool, uint64, error) {
|
||||||
|
entries, err := os.ReadDir(directory)
|
||||||
|
if os.IsNotExist(err) { // non initialized tree
|
||||||
|
return false, 0, nil
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return false, 0, err
|
||||||
|
}
|
||||||
|
if len(entries) == 0 {
|
||||||
|
return false, 0, nil
|
||||||
|
}
|
||||||
|
var hasDBs bool
|
||||||
|
var maxIdx uint64
|
||||||
|
for _, e := range entries {
|
||||||
|
if e.IsDir() {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
hasDBs = true
|
||||||
|
maxIdx = max(u64FromHexString(e.Name()), maxIdx)
|
||||||
|
}
|
||||||
|
return hasDBs, maxIdx, nil
|
||||||
|
}
|
||||||
|
|
||||||
// Type is blobovniczatree storage type used in logs and configuration.
|
// Type is blobovniczatree storage type used in logs and configuration.
|
||||||
const Type = "blobovnicza"
|
const Type = "blobovnicza"
|
||||||
|
|
||||||
|
|
|
@ -130,7 +130,14 @@ func (b *Blobovniczas) iterateSorted(ctx context.Context, addr *oid.Address, cur
|
||||||
isLeafLevel := uint64(len(curPath)) == b.blzShallowDepth
|
isLeafLevel := uint64(len(curPath)) == b.blzShallowDepth
|
||||||
levelWidth := b.blzShallowWidth
|
levelWidth := b.blzShallowWidth
|
||||||
if isLeafLevel {
|
if isLeafLevel {
|
||||||
levelWidth = b.blzLeafWidth
|
hasDBs, maxIdx, err := getBlobovniczaMaxIndex(filepath.Join(append([]string{b.rootPath}, curPath...)...))
|
||||||
|
if err != nil {
|
||||||
|
return false, err
|
||||||
|
}
|
||||||
|
levelWidth = 0
|
||||||
|
if hasDBs {
|
||||||
|
levelWidth = maxIdx + 1
|
||||||
|
}
|
||||||
}
|
}
|
||||||
indices := indexSlice(levelWidth)
|
indices := indexSlice(levelWidth)
|
||||||
|
|
||||||
|
|
|
@ -18,7 +18,6 @@ type cfg struct {
|
||||||
openedCacheSize int
|
openedCacheSize int
|
||||||
blzShallowDepth uint64
|
blzShallowDepth uint64
|
||||||
blzShallowWidth uint64
|
blzShallowWidth uint64
|
||||||
blzLeafWidth uint64
|
|
||||||
compression *compression.Config
|
compression *compression.Config
|
||||||
blzOpts []blobovnicza.Option
|
blzOpts []blobovnicza.Option
|
||||||
reportError func(string, error) // reportError is the function called when encountering disk errors.
|
reportError func(string, error) // reportError is the function called when encountering disk errors.
|
||||||
|
@ -82,12 +81,6 @@ func WithBlobovniczaShallowWidth(width uint64) Option {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func WithBlobovniczaLeafWidth(w uint64) Option {
|
|
||||||
return func(c *cfg) {
|
|
||||||
c.blzLeafWidth = w
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func WithBlobovniczaShallowDepth(depth uint64) Option {
|
func WithBlobovniczaShallowDepth(depth uint64) Option {
|
||||||
return func(c *cfg) {
|
return func(c *cfg) {
|
||||||
c.blzShallowDepth = depth
|
c.blzShallowDepth = depth
|
||||||
|
|
Loading…
Reference in a new issue