Aleksey Savchuk
7fc6101bec
All checks were successful
Tests and linters / Run gofumpt (pull_request) Successful in 2m24s
DCO action / DCO (pull_request) Successful in 2m37s
Pre-commit hooks / Pre-commit (pull_request) Successful in 3m46s
Vulncheck / Vulncheck (pull_request) Successful in 3m46s
Build / Build Components (pull_request) Successful in 4m4s
Tests and linters / Staticcheck (pull_request) Successful in 4m1s
Tests and linters / gopls check (pull_request) Successful in 4m14s
Tests and linters / Lint (pull_request) Successful in 4m49s
Tests and linters / Tests with -race (pull_request) Successful in 5m25s
Tests and linters / Tests (pull_request) Successful in 5m46s
- Remove `testNewShard` and `setInitializedShards` because they violated the default engine workflow. The correct workflow is: first use `New()`, followed by `Open()`, and then `Init()`. As a result, adding new logic to `(*StorageEngine).Init` caused several tests to fail with a panic when attempting to access uninitialized resources. Now, all engines created with the test utils must be initialized manually. The new helper method `prepare` can be used for that purpose. - Additionally, `setInitializedShards` hardcoded the shard worker pool size, which prevented it from being configured in tests and benchmarks. This has been fixed as well. - Ensure engine initialization is done wherever it was missing. - Refactor `setShardsNumOpts`, `setShardsNumAdditionalOpts`, and `setShardsNum`. Make them all depend on `setShardsNumOpts`. Signed-off-by: Aleksey Savchuk <a.savchuk@yadro.com>
84 lines
2.2 KiB
Go
84 lines
2.2 KiB
Go
package engine
|
|
|
|
import (
|
|
"context"
|
|
"testing"
|
|
|
|
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/local_object_storage/internal/testutil"
|
|
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/local_object_storage/shard"
|
|
cidtest "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/container/id/test"
|
|
objectSDK "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/object"
|
|
oid "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/object/id"
|
|
"github.com/stretchr/testify/require"
|
|
)
|
|
|
|
func TestHeadRaw(t *testing.T) {
|
|
cnr := cidtest.ID()
|
|
splitID := objectSDK.NewSplitID()
|
|
|
|
parent := testutil.GenerateObjectWithCID(cnr)
|
|
testutil.AddAttribute(parent, "foo", "bar")
|
|
|
|
var parentAddr oid.Address
|
|
parentAddr.SetContainer(cnr)
|
|
|
|
idParent, _ := parent.ID()
|
|
parentAddr.SetObject(idParent)
|
|
|
|
child := testutil.GenerateObjectWithCID(cnr)
|
|
child.SetParent(parent)
|
|
child.SetParentID(idParent)
|
|
child.SetSplitID(splitID)
|
|
|
|
link := testutil.GenerateObjectWithCID(cnr)
|
|
link.SetParent(parent)
|
|
link.SetParentID(idParent)
|
|
|
|
idChild, _ := child.ID()
|
|
link.SetChildren(idChild)
|
|
link.SetSplitID(splitID)
|
|
|
|
t.Run("virtual object split in different shards", func(t *testing.T) {
|
|
te := testNewEngine(t).setShardsNum(t, 2).prepare(t)
|
|
e := te.engine
|
|
defer func() { require.NoError(t, e.Close(context.Background())) }()
|
|
|
|
s1, s2 := te.shards[0], te.shards[1]
|
|
|
|
var putPrmLeft shard.PutPrm
|
|
putPrmLeft.SetObject(child)
|
|
|
|
var putPrmLink shard.PutPrm
|
|
putPrmLink.SetObject(link)
|
|
|
|
// put most left object in one shard
|
|
_, err := s1.Put(context.Background(), putPrmLeft)
|
|
require.NoError(t, err)
|
|
|
|
// put link object in another shard
|
|
_, err = s2.Put(context.Background(), putPrmLink)
|
|
require.NoError(t, err)
|
|
|
|
// head with raw flag should return SplitInfoError
|
|
var headPrm HeadPrm
|
|
headPrm.WithAddress(parentAddr)
|
|
headPrm.WithRaw(true)
|
|
|
|
_, err = e.Head(context.Background(), headPrm)
|
|
require.Error(t, err)
|
|
|
|
var si *objectSDK.SplitInfoError
|
|
require.ErrorAs(t, err, &si)
|
|
|
|
// SplitInfoError should contain info from both shards
|
|
require.Equal(t, splitID, si.SplitInfo().SplitID())
|
|
|
|
id1, _ := child.ID()
|
|
id2, _ := si.SplitInfo().LastPart()
|
|
require.Equal(t, id1, id2)
|
|
|
|
id1, _ = link.ID()
|
|
id2, _ = si.SplitInfo().Link()
|
|
require.Equal(t, id1, id2)
|
|
})
|
|
}
|