Compare commits
No commits in common. "support/v0.42" and "v0.42.9" have entirely different histories.
support/v0
...
v0.42.9
36 changed files with 96 additions and 540 deletions
|
@ -8,7 +8,7 @@ jobs:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
strategy:
|
strategy:
|
||||||
matrix:
|
matrix:
|
||||||
go_versions: [ '1.22', '1.23' ]
|
go_versions: [ '1.21', '1.22' ]
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v3
|
||||||
|
|
|
@ -16,7 +16,7 @@ jobs:
|
||||||
- name: Set up Go
|
- name: Set up Go
|
||||||
uses: actions/setup-go@v3
|
uses: actions/setup-go@v3
|
||||||
with:
|
with:
|
||||||
go-version: 1.23
|
go-version: 1.22
|
||||||
- name: Set up Python
|
- name: Set up Python
|
||||||
run: |
|
run: |
|
||||||
apt update
|
apt update
|
||||||
|
|
|
@ -11,7 +11,7 @@ jobs:
|
||||||
- name: Set up Go
|
- name: Set up Go
|
||||||
uses: actions/setup-go@v3
|
uses: actions/setup-go@v3
|
||||||
with:
|
with:
|
||||||
go-version: '1.23'
|
go-version: '1.22'
|
||||||
cache: true
|
cache: true
|
||||||
|
|
||||||
- name: Install linters
|
- name: Install linters
|
||||||
|
@ -25,7 +25,7 @@ jobs:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
strategy:
|
strategy:
|
||||||
matrix:
|
matrix:
|
||||||
go_versions: [ '1.22', '1.23' ]
|
go_versions: [ '1.21', '1.22' ]
|
||||||
fail-fast: false
|
fail-fast: false
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v3
|
||||||
|
@ -48,7 +48,7 @@ jobs:
|
||||||
- name: Set up Go
|
- name: Set up Go
|
||||||
uses: actions/setup-go@v3
|
uses: actions/setup-go@v3
|
||||||
with:
|
with:
|
||||||
go-version: '1.22'
|
go-version: '1.21'
|
||||||
cache: true
|
cache: true
|
||||||
|
|
||||||
- name: Run tests
|
- name: Run tests
|
||||||
|
@ -63,7 +63,7 @@ jobs:
|
||||||
- name: Set up Go
|
- name: Set up Go
|
||||||
uses: actions/setup-go@v3
|
uses: actions/setup-go@v3
|
||||||
with:
|
with:
|
||||||
go-version: '1.23'
|
go-version: '1.22'
|
||||||
cache: true
|
cache: true
|
||||||
|
|
||||||
- name: Install staticcheck
|
- name: Install staticcheck
|
||||||
|
@ -81,7 +81,7 @@ jobs:
|
||||||
- name: Set up Go
|
- name: Set up Go
|
||||||
uses: actions/setup-go@v3
|
uses: actions/setup-go@v3
|
||||||
with:
|
with:
|
||||||
go-version: '1.22'
|
go-version: '1.21'
|
||||||
cache: true
|
cache: true
|
||||||
|
|
||||||
- name: Install gopls
|
- name: Install gopls
|
||||||
|
|
|
@ -13,7 +13,7 @@ jobs:
|
||||||
- name: Setup Go
|
- name: Setup Go
|
||||||
uses: actions/setup-go@v3
|
uses: actions/setup-go@v3
|
||||||
with:
|
with:
|
||||||
go-version: '1.23'
|
go-version: '1.22'
|
||||||
|
|
||||||
- name: Install govulncheck
|
- name: Install govulncheck
|
||||||
run: go install golang.org/x/vuln/cmd/govulncheck@latest
|
run: go install golang.org/x/vuln/cmd/govulncheck@latest
|
||||||
|
|
|
@ -12,8 +12,7 @@ run:
|
||||||
# output configuration options
|
# output configuration options
|
||||||
output:
|
output:
|
||||||
# colored-line-number|line-number|json|tab|checkstyle|code-climate, default is "colored-line-number"
|
# colored-line-number|line-number|json|tab|checkstyle|code-climate, default is "colored-line-number"
|
||||||
formats:
|
format: tab
|
||||||
- format: tab
|
|
||||||
|
|
||||||
# all available settings of specific linters
|
# all available settings of specific linters
|
||||||
linters-settings:
|
linters-settings:
|
||||||
|
|
10
Makefile
10
Makefile
|
@ -8,8 +8,8 @@ HUB_IMAGE ?= truecloudlab/frostfs
|
||||||
HUB_TAG ?= "$(shell echo ${VERSION} | sed 's/^v//')"
|
HUB_TAG ?= "$(shell echo ${VERSION} | sed 's/^v//')"
|
||||||
|
|
||||||
GO_VERSION ?= 1.22
|
GO_VERSION ?= 1.22
|
||||||
LINT_VERSION ?= 1.60.1
|
LINT_VERSION ?= 1.56.1
|
||||||
TRUECLOUDLAB_LINT_VERSION ?= 0.0.7
|
TRUECLOUDLAB_LINT_VERSION ?= 0.0.5
|
||||||
PROTOC_VERSION ?= 25.0
|
PROTOC_VERSION ?= 25.0
|
||||||
PROTOC_GEN_GO_VERSION ?= $(shell go list -f '{{.Version}}' -m google.golang.org/protobuf)
|
PROTOC_GEN_GO_VERSION ?= $(shell go list -f '{{.Version}}' -m google.golang.org/protobuf)
|
||||||
PROTOGEN_FROSTFS_VERSION ?= $(shell go list -f '{{.Version}}' -m git.frostfs.info/TrueCloudLab/frostfs-api-go/v2)
|
PROTOGEN_FROSTFS_VERSION ?= $(shell go list -f '{{.Version}}' -m git.frostfs.info/TrueCloudLab/frostfs-api-go/v2)
|
||||||
|
@ -17,7 +17,7 @@ PROTOC_OS_VERSION=osx-x86_64
|
||||||
ifeq ($(shell uname), Linux)
|
ifeq ($(shell uname), Linux)
|
||||||
PROTOC_OS_VERSION=linux-x86_64
|
PROTOC_OS_VERSION=linux-x86_64
|
||||||
endif
|
endif
|
||||||
STATICCHECK_VERSION ?= 2024.1.1
|
STATICCHECK_VERSION ?= 2023.1.6
|
||||||
ARCH = amd64
|
ARCH = amd64
|
||||||
|
|
||||||
BIN = bin
|
BIN = bin
|
||||||
|
@ -190,7 +190,7 @@ lint-install:
|
||||||
@@make -C $(TMP_DIR)/linters lib CGO_ENABLED=1 OUT_DIR=$(OUTPUT_LINT_DIR)
|
@@make -C $(TMP_DIR)/linters lib CGO_ENABLED=1 OUT_DIR=$(OUTPUT_LINT_DIR)
|
||||||
@rm -rf $(TMP_DIR)/linters
|
@rm -rf $(TMP_DIR)/linters
|
||||||
@rmdir $(TMP_DIR) 2>/dev/null || true
|
@rmdir $(TMP_DIR) 2>/dev/null || true
|
||||||
@CGO_ENABLED=1 GOBIN=$(LINT_DIR) go install -trimpath github.com/golangci/golangci-lint/cmd/golangci-lint@v$(LINT_VERSION)
|
@CGO_ENABLED=1 GOBIN=$(LINT_DIR) go install github.com/golangci/golangci-lint/cmd/golangci-lint@v$(LINT_VERSION)
|
||||||
|
|
||||||
# Run linters
|
# Run linters
|
||||||
lint:
|
lint:
|
||||||
|
@ -210,7 +210,7 @@ staticcheck-run:
|
||||||
@if [ ! -d "$(STATICCHECK_VERSION_DIR)" ]; then \
|
@if [ ! -d "$(STATICCHECK_VERSION_DIR)" ]; then \
|
||||||
make staticcheck-install; \
|
make staticcheck-install; \
|
||||||
fi
|
fi
|
||||||
@$(STATICCHECK_VERSION_DIR)/staticcheck -checks inherit,-SA1019 ./...
|
@$(STATICCHECK_VERSION_DIR)/staticcheck ./...
|
||||||
|
|
||||||
# Install gopls
|
# Install gopls
|
||||||
gopls-install:
|
gopls-install:
|
||||||
|
|
|
@ -49,7 +49,7 @@ The latest version of frostfs-node works with frostfs-contract
|
||||||
|
|
||||||
# Building
|
# Building
|
||||||
|
|
||||||
To make all binaries you need Go 1.22+ and `make`:
|
To make all binaries you need Go 1.21+ and `make`:
|
||||||
```
|
```
|
||||||
make all
|
make all
|
||||||
```
|
```
|
||||||
|
|
|
@ -1070,7 +1070,7 @@ func initLocalStorage(ctx context.Context, c *cfg) {
|
||||||
c.onShutdown(func() {
|
c.onShutdown(func() {
|
||||||
c.log.Info(logs.FrostFSNodeClosingComponentsOfTheStorageEngine)
|
c.log.Info(logs.FrostFSNodeClosingComponentsOfTheStorageEngine)
|
||||||
|
|
||||||
err := ls.Close(context.WithoutCancel(ctx))
|
err := ls.Close(context.Background())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
c.log.Info(logs.FrostFSNodeStorageEngineClosingFailure,
|
c.log.Info(logs.FrostFSNodeStorageEngineClosingFailure,
|
||||||
zap.String("error", err.Error()),
|
zap.String("error", err.Error()),
|
||||||
|
|
|
@ -21,7 +21,7 @@ func initTracing(ctx context.Context, c *cfg) {
|
||||||
c.closers = append(c.closers, closer{
|
c.closers = append(c.closers, closer{
|
||||||
name: "tracing",
|
name: "tracing",
|
||||||
fn: func() {
|
fn: func() {
|
||||||
ctx, cancel := context.WithTimeout(context.WithoutCancel(ctx), time.Second*5)
|
ctx, cancel := context.WithTimeout(context.Background(), time.Second*5)
|
||||||
defer cancel()
|
defer cancel()
|
||||||
err := tracing.Shutdown(ctx) // cfg context cancels before close
|
err := tracing.Shutdown(ctx) // cfg context cancels before close
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
34
go.mod
34
go.mod
|
@ -1,14 +1,14 @@
|
||||||
module git.frostfs.info/TrueCloudLab/frostfs-node
|
module git.frostfs.info/TrueCloudLab/frostfs-node
|
||||||
|
|
||||||
go 1.22
|
go 1.21
|
||||||
|
|
||||||
require (
|
require (
|
||||||
code.gitea.io/sdk/gitea v0.17.1
|
code.gitea.io/sdk/gitea v0.17.1
|
||||||
git.frostfs.info/TrueCloudLab/frostfs-api-go/v2 v2.16.1-0.20240906121927-2c79f770e449
|
git.frostfs.info/TrueCloudLab/frostfs-api-go/v2 v2.16.1-0.20240726072425-3dfa2f4fd65e
|
||||||
git.frostfs.info/TrueCloudLab/frostfs-contract v0.19.3-0.20240621131249-49e5270f673e
|
git.frostfs.info/TrueCloudLab/frostfs-contract v0.19.3-0.20240621131249-49e5270f673e
|
||||||
git.frostfs.info/TrueCloudLab/frostfs-crypto v0.6.0
|
git.frostfs.info/TrueCloudLab/frostfs-crypto v0.6.0
|
||||||
git.frostfs.info/TrueCloudLab/frostfs-observability v0.0.0-20231101111734-b3ad3335ff65
|
git.frostfs.info/TrueCloudLab/frostfs-observability v0.0.0-20231101111734-b3ad3335ff65
|
||||||
git.frostfs.info/TrueCloudLab/frostfs-sdk-go v0.0.0-20240906135451-a7aabe53491c
|
git.frostfs.info/TrueCloudLab/frostfs-sdk-go v0.0.0-20240726111349-9da46f566fec
|
||||||
git.frostfs.info/TrueCloudLab/hrw v1.2.1
|
git.frostfs.info/TrueCloudLab/hrw v1.2.1
|
||||||
git.frostfs.info/TrueCloudLab/policy-engine v0.0.0-20240712081403-2628f6184984
|
git.frostfs.info/TrueCloudLab/policy-engine v0.0.0-20240712081403-2628f6184984
|
||||||
git.frostfs.info/TrueCloudLab/tzhash v1.8.0
|
git.frostfs.info/TrueCloudLab/tzhash v1.8.0
|
||||||
|
@ -23,7 +23,7 @@ require (
|
||||||
github.com/mitchellh/go-homedir v1.1.0
|
github.com/mitchellh/go-homedir v1.1.0
|
||||||
github.com/mr-tron/base58 v1.2.0
|
github.com/mr-tron/base58 v1.2.0
|
||||||
github.com/multiformats/go-multiaddr v0.12.1
|
github.com/multiformats/go-multiaddr v0.12.1
|
||||||
github.com/nspcc-dev/neo-go v0.106.3
|
github.com/nspcc-dev/neo-go v0.106.2
|
||||||
github.com/olekukonko/tablewriter v0.0.5
|
github.com/olekukonko/tablewriter v0.0.5
|
||||||
github.com/panjf2000/ants/v2 v2.9.0
|
github.com/panjf2000/ants/v2 v2.9.0
|
||||||
github.com/paulmach/orb v0.11.0
|
github.com/paulmach/orb v0.11.0
|
||||||
|
@ -39,11 +39,11 @@ require (
|
||||||
go.opentelemetry.io/otel/trace v1.22.0
|
go.opentelemetry.io/otel/trace v1.22.0
|
||||||
go.uber.org/zap v1.27.0
|
go.uber.org/zap v1.27.0
|
||||||
golang.org/x/exp v0.0.0-20240222234643-814bf88cf225
|
golang.org/x/exp v0.0.0-20240222234643-814bf88cf225
|
||||||
golang.org/x/sync v0.8.0
|
golang.org/x/sync v0.6.0
|
||||||
golang.org/x/sys v0.25.0
|
golang.org/x/sys v0.18.0
|
||||||
golang.org/x/term v0.24.0
|
golang.org/x/term v0.18.0
|
||||||
google.golang.org/grpc v1.66.0
|
google.golang.org/grpc v1.63.2
|
||||||
google.golang.org/protobuf v1.34.2
|
google.golang.org/protobuf v1.33.0
|
||||||
gopkg.in/yaml.v3 v3.0.1
|
gopkg.in/yaml.v3 v3.0.1
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -60,10 +60,10 @@ require (
|
||||||
github.com/beorn7/perks v1.0.1 // indirect
|
github.com/beorn7/perks v1.0.1 // indirect
|
||||||
github.com/bits-and-blooms/bitset v1.13.0 // indirect
|
github.com/bits-and-blooms/bitset v1.13.0 // indirect
|
||||||
github.com/cenkalti/backoff/v4 v4.2.1 // indirect
|
github.com/cenkalti/backoff/v4 v4.2.1 // indirect
|
||||||
github.com/cespare/xxhash/v2 v2.3.0 // indirect
|
github.com/cespare/xxhash/v2 v2.2.0 // indirect
|
||||||
github.com/consensys/bavard v0.1.13 // indirect
|
github.com/consensys/bavard v0.1.13 // indirect
|
||||||
github.com/consensys/gnark-crypto v0.12.2-0.20231222162921-eb75782795d2 // indirect
|
github.com/consensys/gnark-crypto v0.12.2-0.20231222162921-eb75782795d2 // indirect
|
||||||
github.com/cpuguy83/go-md2man/v2 v2.0.4 // indirect
|
github.com/cpuguy83/go-md2man/v2 v2.0.3 // indirect
|
||||||
github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc // indirect
|
github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc // indirect
|
||||||
github.com/davidmz/go-pageant v1.0.2 // indirect
|
github.com/davidmz/go-pageant v1.0.2 // indirect
|
||||||
github.com/decred/dcrd/dcrec/secp256k1/v4 v4.2.0 // indirect
|
github.com/decred/dcrd/dcrec/secp256k1/v4 v4.2.0 // indirect
|
||||||
|
@ -96,7 +96,7 @@ require (
|
||||||
github.com/multiformats/go-multihash v0.2.3 // indirect
|
github.com/multiformats/go-multihash v0.2.3 // indirect
|
||||||
github.com/multiformats/go-varint v0.0.7 // indirect
|
github.com/multiformats/go-varint v0.0.7 // indirect
|
||||||
github.com/nspcc-dev/go-ordered-json v0.0.0-20240301084351-0246b013f8b2 // indirect
|
github.com/nspcc-dev/go-ordered-json v0.0.0-20240301084351-0246b013f8b2 // indirect
|
||||||
github.com/nspcc-dev/neo-go/pkg/interop v0.0.0-20240727093519-1a48f1ce43ec // indirect
|
github.com/nspcc-dev/neo-go/pkg/interop v0.0.0-20240521091047-78685785716d // indirect
|
||||||
github.com/nspcc-dev/rfc6979 v0.2.1 // indirect
|
github.com/nspcc-dev/rfc6979 v0.2.1 // indirect
|
||||||
github.com/pelletier/go-toml/v2 v2.1.1 // indirect
|
github.com/pelletier/go-toml/v2 v2.1.1 // indirect
|
||||||
github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 // indirect
|
github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 // indirect
|
||||||
|
@ -118,11 +118,11 @@ require (
|
||||||
go.opentelemetry.io/otel/sdk v1.22.0 // indirect
|
go.opentelemetry.io/otel/sdk v1.22.0 // indirect
|
||||||
go.opentelemetry.io/proto/otlp v1.1.0 // indirect
|
go.opentelemetry.io/proto/otlp v1.1.0 // indirect
|
||||||
go.uber.org/multierr v1.11.0 // indirect
|
go.uber.org/multierr v1.11.0 // indirect
|
||||||
golang.org/x/crypto v0.27.0 // indirect
|
golang.org/x/crypto v0.21.0 // indirect
|
||||||
golang.org/x/net v0.29.0 // indirect
|
golang.org/x/net v0.23.0 // indirect
|
||||||
golang.org/x/text v0.18.0 // indirect
|
golang.org/x/text v0.14.0 // indirect
|
||||||
google.golang.org/genproto/googleapis/api v0.0.0-20240604185151-ef581f913117 // indirect
|
google.golang.org/genproto/googleapis/api v0.0.0-20240227224415-6ceb2ff114de // indirect
|
||||||
google.golang.org/genproto/googleapis/rpc v0.0.0-20240903143218-8af14fe29dc1 // indirect
|
google.golang.org/genproto/googleapis/rpc v0.0.0-20240227224415-6ceb2ff114de // indirect
|
||||||
gopkg.in/ini.v1 v1.67.0 // indirect
|
gopkg.in/ini.v1 v1.67.0 // indirect
|
||||||
lukechampine.com/blake3 v1.2.1 // indirect
|
lukechampine.com/blake3 v1.2.1 // indirect
|
||||||
rsc.io/tmplfunc v0.0.3 // indirect
|
rsc.io/tmplfunc v0.0.3 // indirect
|
||||||
|
|
BIN
go.sum
BIN
go.sum
Binary file not shown.
|
@ -27,21 +27,21 @@ type objectDesc struct {
|
||||||
storageID []byte
|
storageID []byte
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestAll(t *testing.T, cons Constructor, minSize, maxSize uint64) {
|
func TestAll(t *testing.T, cons Constructor, min, max uint64) {
|
||||||
t.Run("get", func(t *testing.T) {
|
t.Run("get", func(t *testing.T) {
|
||||||
TestGet(t, cons, minSize, maxSize)
|
TestGet(t, cons, min, max)
|
||||||
})
|
})
|
||||||
t.Run("get range", func(t *testing.T) {
|
t.Run("get range", func(t *testing.T) {
|
||||||
TestGetRange(t, cons, minSize, maxSize)
|
TestGetRange(t, cons, min, max)
|
||||||
})
|
})
|
||||||
t.Run("delete", func(t *testing.T) {
|
t.Run("delete", func(t *testing.T) {
|
||||||
TestDelete(t, cons, minSize, maxSize)
|
TestDelete(t, cons, min, max)
|
||||||
})
|
})
|
||||||
t.Run("exists", func(t *testing.T) {
|
t.Run("exists", func(t *testing.T) {
|
||||||
TestExists(t, cons, minSize, maxSize)
|
TestExists(t, cons, min, max)
|
||||||
})
|
})
|
||||||
t.Run("iterate", func(t *testing.T) {
|
t.Run("iterate", func(t *testing.T) {
|
||||||
TestIterate(t, cons, minSize, maxSize)
|
TestIterate(t, cons, min, max)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -51,12 +51,12 @@ func TestInfo(t *testing.T, cons Constructor, expectedType string, expectedPath
|
||||||
require.Equal(t, expectedPath, s.Path())
|
require.Equal(t, expectedPath, s.Path())
|
||||||
}
|
}
|
||||||
|
|
||||||
func prepare(t *testing.T, count int, s common.Storage, minSize, maxSize uint64) []objectDesc {
|
func prepare(t *testing.T, count int, s common.Storage, min, max uint64) []objectDesc {
|
||||||
objects := make([]objectDesc, count)
|
objects := make([]objectDesc, count)
|
||||||
|
|
||||||
r := mrand.New(mrand.NewSource(0))
|
r := mrand.New(mrand.NewSource(0))
|
||||||
for i := range objects {
|
for i := range objects {
|
||||||
objects[i].obj = NewObject(minSize + uint64(r.Intn(int(maxSize-minSize+1)))) // not too large
|
objects[i].obj = NewObject(min + uint64(r.Intn(int(max-min+1)))) // not too large
|
||||||
objects[i].addr = objectCore.AddressOf(objects[i].obj)
|
objects[i].addr = objectCore.AddressOf(objects[i].obj)
|
||||||
|
|
||||||
raw, err := objects[i].obj.Marshal()
|
raw, err := objects[i].obj.Marshal()
|
||||||
|
|
|
@ -13,12 +13,12 @@ import (
|
||||||
|
|
||||||
// TestControl checks correctness of a read-only mode.
|
// TestControl checks correctness of a read-only mode.
|
||||||
// cons must return a storage which is NOT opened.
|
// cons must return a storage which is NOT opened.
|
||||||
func TestControl(t *testing.T, cons Constructor, minSize, maxSize uint64) {
|
func TestControl(t *testing.T, cons Constructor, min, max uint64) {
|
||||||
s := cons(t)
|
s := cons(t)
|
||||||
require.NoError(t, s.Open(mode.ComponentReadWrite))
|
require.NoError(t, s.Open(mode.ComponentReadWrite))
|
||||||
require.NoError(t, s.Init())
|
require.NoError(t, s.Init())
|
||||||
|
|
||||||
objects := prepare(t, 10, s, minSize, maxSize)
|
objects := prepare(t, 10, s, min, max)
|
||||||
require.NoError(t, s.Close())
|
require.NoError(t, s.Close())
|
||||||
|
|
||||||
require.NoError(t, s.Open(mode.ComponentReadOnly))
|
require.NoError(t, s.Open(mode.ComponentReadOnly))
|
||||||
|
@ -34,7 +34,7 @@ func TestControl(t *testing.T, cons Constructor, minSize, maxSize uint64) {
|
||||||
|
|
||||||
t.Run("put fails", func(t *testing.T) {
|
t.Run("put fails", func(t *testing.T) {
|
||||||
var prm common.PutPrm
|
var prm common.PutPrm
|
||||||
prm.Object = NewObject(minSize + uint64(rand.Intn(int(maxSize-minSize+1))))
|
prm.Object = NewObject(min + uint64(rand.Intn(int(max-min+1))))
|
||||||
prm.Address = objectCore.AddressOf(prm.Object)
|
prm.Address = objectCore.AddressOf(prm.Object)
|
||||||
|
|
||||||
_, err := s.Put(context.Background(), prm)
|
_, err := s.Put(context.Background(), prm)
|
||||||
|
|
|
@ -11,13 +11,13 @@ import (
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestDelete(t *testing.T, cons Constructor, minSize, maxSize uint64) {
|
func TestDelete(t *testing.T, cons Constructor, min, max uint64) {
|
||||||
s := cons(t)
|
s := cons(t)
|
||||||
require.NoError(t, s.Open(mode.ComponentReadWrite))
|
require.NoError(t, s.Open(mode.ComponentReadWrite))
|
||||||
require.NoError(t, s.Init())
|
require.NoError(t, s.Init())
|
||||||
defer func() { require.NoError(t, s.Close()) }()
|
defer func() { require.NoError(t, s.Close()) }()
|
||||||
|
|
||||||
objects := prepare(t, 4, s, minSize, maxSize)
|
objects := prepare(t, 4, s, min, max)
|
||||||
|
|
||||||
t.Run("delete non-existent", func(t *testing.T) {
|
t.Run("delete non-existent", func(t *testing.T) {
|
||||||
var prm common.DeletePrm
|
var prm common.DeletePrm
|
||||||
|
|
|
@ -10,13 +10,13 @@ import (
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestExists(t *testing.T, cons Constructor, minSize, maxSize uint64) {
|
func TestExists(t *testing.T, cons Constructor, min, max uint64) {
|
||||||
s := cons(t)
|
s := cons(t)
|
||||||
require.NoError(t, s.Open(mode.ComponentReadWrite))
|
require.NoError(t, s.Open(mode.ComponentReadWrite))
|
||||||
require.NoError(t, s.Init())
|
require.NoError(t, s.Init())
|
||||||
defer func() { require.NoError(t, s.Close()) }()
|
defer func() { require.NoError(t, s.Close()) }()
|
||||||
|
|
||||||
objects := prepare(t, 1, s, minSize, maxSize)
|
objects := prepare(t, 1, s, min, max)
|
||||||
|
|
||||||
t.Run("missing object", func(t *testing.T) {
|
t.Run("missing object", func(t *testing.T) {
|
||||||
prm := common.ExistsPrm{Address: oidtest.Address()}
|
prm := common.ExistsPrm{Address: oidtest.Address()}
|
||||||
|
|
|
@ -11,13 +11,13 @@ import (
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestGet(t *testing.T, cons Constructor, minSize, maxSize uint64) {
|
func TestGet(t *testing.T, cons Constructor, min, max uint64) {
|
||||||
s := cons(t)
|
s := cons(t)
|
||||||
require.NoError(t, s.Open(mode.ComponentReadWrite))
|
require.NoError(t, s.Open(mode.ComponentReadWrite))
|
||||||
require.NoError(t, s.Init())
|
require.NoError(t, s.Init())
|
||||||
defer func() { require.NoError(t, s.Close()) }()
|
defer func() { require.NoError(t, s.Close()) }()
|
||||||
|
|
||||||
objects := prepare(t, 2, s, minSize, maxSize)
|
objects := prepare(t, 2, s, min, max)
|
||||||
|
|
||||||
t.Run("missing object", func(t *testing.T) {
|
t.Run("missing object", func(t *testing.T) {
|
||||||
gPrm := common.GetPrm{Address: oidtest.Address()}
|
gPrm := common.GetPrm{Address: oidtest.Address()}
|
||||||
|
|
|
@ -13,13 +13,13 @@ import (
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestGetRange(t *testing.T, cons Constructor, minSize, maxSize uint64) {
|
func TestGetRange(t *testing.T, cons Constructor, min, max uint64) {
|
||||||
s := cons(t)
|
s := cons(t)
|
||||||
require.NoError(t, s.Open(mode.ComponentReadWrite))
|
require.NoError(t, s.Open(mode.ComponentReadWrite))
|
||||||
require.NoError(t, s.Init())
|
require.NoError(t, s.Init())
|
||||||
defer func() { require.NoError(t, s.Close()) }()
|
defer func() { require.NoError(t, s.Close()) }()
|
||||||
|
|
||||||
objects := prepare(t, 1, s, minSize, maxSize)
|
objects := prepare(t, 1, s, min, max)
|
||||||
|
|
||||||
t.Run("missing object", func(t *testing.T) {
|
t.Run("missing object", func(t *testing.T) {
|
||||||
gPrm := common.GetRangePrm{Address: oidtest.Address()}
|
gPrm := common.GetRangePrm{Address: oidtest.Address()}
|
||||||
|
|
|
@ -10,13 +10,13 @@ import (
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestIterate(t *testing.T, cons Constructor, minSize, maxSize uint64) {
|
func TestIterate(t *testing.T, cons Constructor, min, max uint64) {
|
||||||
s := cons(t)
|
s := cons(t)
|
||||||
require.NoError(t, s.Open(mode.ComponentReadWrite))
|
require.NoError(t, s.Open(mode.ComponentReadWrite))
|
||||||
require.NoError(t, s.Init())
|
require.NoError(t, s.Init())
|
||||||
defer func() { require.NoError(t, s.Close()) }()
|
defer func() { require.NoError(t, s.Close()) }()
|
||||||
|
|
||||||
objects := prepare(t, 10, s, minSize, maxSize)
|
objects := prepare(t, 10, s, min, max)
|
||||||
|
|
||||||
// Delete random object to ensure it is not iterated over.
|
// Delete random object to ensure it is not iterated over.
|
||||||
const delID = 2
|
const delID = 2
|
||||||
|
|
|
@ -236,7 +236,7 @@ func (db *DB) inhumeTx(tx *bbolt.Tx, epoch uint64, prm InhumePrm, res *InhumeRes
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
} else if errors.As(err, &ecErr) {
|
} else if errors.As(err, &ecErr) {
|
||||||
err = db.inhumeECInfo(tx, epoch, prm.tomb, res, garbageBKT, graveyardBKT, ecErr.ECInfo(), cnr, bkt, value)
|
err = db.inhumeECInfo(tx, epoch, prm.tomb, res, garbageBKT, graveyardBKT, ecErr.ECInfo(), cnr, bkt, value, targetKey)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
@ -280,7 +280,7 @@ func (db *DB) inhumeTx(tx *bbolt.Tx, epoch uint64, prm InhumePrm, res *InhumeRes
|
||||||
|
|
||||||
func (db *DB) inhumeECInfo(tx *bbolt.Tx, epoch uint64, tomb *oid.Address, res *InhumeRes,
|
func (db *DB) inhumeECInfo(tx *bbolt.Tx, epoch uint64, tomb *oid.Address, res *InhumeRes,
|
||||||
garbageBKT *bbolt.Bucket, graveyardBKT *bbolt.Bucket,
|
garbageBKT *bbolt.Bucket, graveyardBKT *bbolt.Bucket,
|
||||||
ecInfo *objectSDK.ECInfo, cnr cid.ID, targetBucket *bbolt.Bucket, value []byte,
|
ecInfo *objectSDK.ECInfo, cnr cid.ID, targetBucket *bbolt.Bucket, value []byte, targetKey []byte,
|
||||||
) error {
|
) error {
|
||||||
for _, chunk := range ecInfo.Chunks {
|
for _, chunk := range ecInfo.Chunks {
|
||||||
chunkBuf := make([]byte, addressKeySize)
|
chunkBuf := make([]byte, addressKeySize)
|
||||||
|
@ -296,11 +296,11 @@ func (db *DB) inhumeECInfo(tx *bbolt.Tx, epoch uint64, tomb *oid.Address, res *I
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
chunkKey := addressKey(chunkAddr, chunkBuf)
|
err = db.updateDeleteInfo(tx, garbageBKT, graveyardBKT, targetKey, cnr, chunkObj, res)
|
||||||
err = db.updateDeleteInfo(tx, garbageBKT, graveyardBKT, chunkKey, cnr, chunkObj, res)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
chunkKey := addressKey(chunkAddr, chunkBuf)
|
||||||
if tomb != nil {
|
if tomb != nil {
|
||||||
_, err = db.markAsGC(graveyardBKT, garbageBKT, chunkKey)
|
_, err = db.markAsGC(graveyardBKT, garbageBKT, chunkKey)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
|
@ -1,116 +0,0 @@
|
||||||
package meta
|
|
||||||
|
|
||||||
import (
|
|
||||||
"context"
|
|
||||||
"path/filepath"
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/local_object_storage/internal/testutil"
|
|
||||||
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/local_object_storage/shard/mode"
|
|
||||||
apistatus "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/client/status"
|
|
||||||
cidtest "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/container/id/test"
|
|
||||||
objectSDK "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/object"
|
|
||||||
oid "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/object/id"
|
|
||||||
oidtest "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/object/id/test"
|
|
||||||
"github.com/stretchr/testify/require"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestInhumeECObject(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
db := New(
|
|
||||||
WithPath(filepath.Join(t.TempDir(), "metabase")),
|
|
||||||
WithPermissions(0o600),
|
|
||||||
WithEpochState(epochState{uint64(12)}),
|
|
||||||
)
|
|
||||||
|
|
||||||
require.NoError(t, db.Open(context.Background(), mode.ReadWrite))
|
|
||||||
require.NoError(t, db.Init())
|
|
||||||
defer func() { require.NoError(t, db.Close()) }()
|
|
||||||
|
|
||||||
cnr := cidtest.ID()
|
|
||||||
ecChunk := oidtest.ID()
|
|
||||||
ecChunk2 := oidtest.ID()
|
|
||||||
ecParent := oidtest.ID()
|
|
||||||
tombstoneID := oidtest.ID()
|
|
||||||
|
|
||||||
chunkObj := testutil.GenerateObjectWithCID(cnr)
|
|
||||||
chunkObj.SetContainerID(cnr)
|
|
||||||
chunkObj.SetID(ecChunk)
|
|
||||||
chunkObj.SetPayload([]byte{0, 1, 2, 3, 4})
|
|
||||||
chunkObj.SetPayloadSize(uint64(5))
|
|
||||||
chunkObj.SetECHeader(objectSDK.NewECHeader(objectSDK.ECParentInfo{ID: ecParent}, 0, 3, []byte{}, 0))
|
|
||||||
|
|
||||||
chunkObj2 := testutil.GenerateObjectWithCID(cnr)
|
|
||||||
chunkObj2.SetContainerID(cnr)
|
|
||||||
chunkObj2.SetID(ecChunk2)
|
|
||||||
chunkObj2.SetPayload([]byte{0, 1, 2, 3, 4, 5, 6, 7, 8, 9})
|
|
||||||
chunkObj2.SetPayloadSize(uint64(10))
|
|
||||||
chunkObj2.SetECHeader(objectSDK.NewECHeader(objectSDK.ECParentInfo{ID: ecParent}, 1, 3, []byte{}, 0))
|
|
||||||
|
|
||||||
// put object with EC
|
|
||||||
|
|
||||||
var prm PutPrm
|
|
||||||
prm.SetObject(chunkObj)
|
|
||||||
prm.SetStorageID([]byte("0/0"))
|
|
||||||
_, err := db.Put(context.Background(), prm)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
prm.SetObject(chunkObj2)
|
|
||||||
_, err = db.Put(context.Background(), prm)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
var ecChunkAddress oid.Address
|
|
||||||
ecChunkAddress.SetContainer(cnr)
|
|
||||||
ecChunkAddress.SetObject(ecChunk)
|
|
||||||
|
|
||||||
var ecParentAddress oid.Address
|
|
||||||
ecParentAddress.SetContainer(cnr)
|
|
||||||
ecParentAddress.SetObject(ecParent)
|
|
||||||
|
|
||||||
var chunkObjectAddress oid.Address
|
|
||||||
chunkObjectAddress.SetContainer(cnr)
|
|
||||||
chunkObjectAddress.SetObject(ecChunk)
|
|
||||||
|
|
||||||
var getPrm GetPrm
|
|
||||||
|
|
||||||
getPrm.SetAddress(ecChunkAddress)
|
|
||||||
_, err = db.Get(context.Background(), getPrm)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
var ecInfoError *objectSDK.ECInfoError
|
|
||||||
getPrm.SetAddress(ecParentAddress)
|
|
||||||
_, err = db.Get(context.Background(), getPrm)
|
|
||||||
require.ErrorAs(t, err, &ecInfoError)
|
|
||||||
require.True(t, len(ecInfoError.ECInfo().Chunks) == 2 &&
|
|
||||||
ecInfoError.ECInfo().Chunks[0].Index == 0 &&
|
|
||||||
ecInfoError.ECInfo().Chunks[0].Total == 3)
|
|
||||||
|
|
||||||
// inhume Chunk
|
|
||||||
var inhumePrm InhumePrm
|
|
||||||
var tombAddress oid.Address
|
|
||||||
inhumePrm.SetAddresses(chunkObjectAddress)
|
|
||||||
res, err := db.Inhume(context.Background(), inhumePrm)
|
|
||||||
require.NoError(t, err)
|
|
||||||
require.True(t, len(res.deletionDetails) == 1)
|
|
||||||
require.True(t, res.deletionDetails[0].Size == 5)
|
|
||||||
|
|
||||||
// inhume EC parent (like Delete does)
|
|
||||||
tombAddress.SetContainer(cnr)
|
|
||||||
tombAddress.SetObject(tombstoneID)
|
|
||||||
inhumePrm.SetAddresses(ecParentAddress)
|
|
||||||
inhumePrm.SetTombstoneAddress(tombAddress)
|
|
||||||
res, err = db.Inhume(context.Background(), inhumePrm)
|
|
||||||
require.NoError(t, err)
|
|
||||||
// Previously deleted chunk shouldn't be in the details, because it is marked as garbage
|
|
||||||
require.True(t, len(res.deletionDetails) == 1)
|
|
||||||
require.True(t, res.deletionDetails[0].Size == 10)
|
|
||||||
|
|
||||||
getPrm.SetAddress(ecParentAddress)
|
|
||||||
_, err = db.Get(context.Background(), getPrm)
|
|
||||||
require.ErrorAs(t, err, new(*apistatus.ObjectAlreadyRemoved))
|
|
||||||
|
|
||||||
getPrm.SetAddress(ecChunkAddress)
|
|
||||||
_, err = db.Get(context.Background(), getPrm)
|
|
||||||
require.ErrorAs(t, err, new(*apistatus.ObjectAlreadyRemoved))
|
|
||||||
}
|
|
|
@ -1161,7 +1161,6 @@ func (t *boltForest) fillSortedChildren(b *bbolt.Bucket, nodeIDs MultiNode, h *f
|
||||||
lastFilename = nil
|
lastFilename = nil
|
||||||
nodes = nil
|
nodes = nil
|
||||||
length = actualLength + 1
|
length = actualLength + 1
|
||||||
count = 0
|
|
||||||
c.Seek(append(prefix, byte(length), byte(length>>8)))
|
c.Seek(append(prefix, byte(length), byte(length>>8)))
|
||||||
c.Prev() // c.Next() will be performed by for loop
|
c.Prev() // c.Next() will be performed by for loop
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,13 +1,11 @@
|
||||||
package pilorama
|
package pilorama
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
|
||||||
"context"
|
"context"
|
||||||
"crypto/rand"
|
"crypto/rand"
|
||||||
"fmt"
|
"fmt"
|
||||||
mrand "math/rand"
|
mrand "math/rand"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
"slices"
|
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
|
@ -234,65 +232,6 @@ func BenchmarkForestSortedIteration(b *testing.B) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// The issue which we call "BugWithSkip" is easiest to understand when filenames are
|
|
||||||
// monotonically increasing numbers. We want the list of sorted filenames to have different length interleaved.
|
|
||||||
// The bug happens when we switch between length during listing.
|
|
||||||
// Thus this test contains numbers from 1 to 2000 and batch size of size 10.
|
|
||||||
func TestForest_TreeSortedIterationBugWithSkip(t *testing.T) {
|
|
||||||
for i := range providers {
|
|
||||||
t.Run(providers[i].name, func(t *testing.T) {
|
|
||||||
testForestTreeSortedIterationBugWithSkip(t, providers[i].construct(t))
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func testForestTreeSortedIterationBugWithSkip(t *testing.T, s ForestStorage) {
|
|
||||||
defer func() { require.NoError(t, s.Close()) }()
|
|
||||||
|
|
||||||
cid := cidtest.ID()
|
|
||||||
d := CIDDescriptor{cid, 0, 1}
|
|
||||||
treeID := "version"
|
|
||||||
treeAdd := func(t *testing.T, ts int, filename string) {
|
|
||||||
_, err := s.TreeMove(context.Background(), d, treeID, &Move{
|
|
||||||
Child: RootID + uint64(ts),
|
|
||||||
Parent: RootID,
|
|
||||||
Meta: Meta{
|
|
||||||
Time: Timestamp(ts),
|
|
||||||
Items: []KeyValue{
|
|
||||||
{Key: AttributeFilename, Value: []byte(filename)},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
})
|
|
||||||
require.NoError(t, err)
|
|
||||||
}
|
|
||||||
|
|
||||||
const count = 2000
|
|
||||||
treeAdd(t, 1, "")
|
|
||||||
for i := 1; i < count; i++ {
|
|
||||||
treeAdd(t, i+1, strconv.Itoa(i+1))
|
|
||||||
}
|
|
||||||
|
|
||||||
var result []MultiNodeInfo
|
|
||||||
treeAppend := func(t *testing.T, last *string, count int) *string {
|
|
||||||
res, cursor, err := s.TreeSortedByFilename(context.Background(), d.CID, treeID, MultiNode{RootID}, last, count)
|
|
||||||
require.NoError(t, err)
|
|
||||||
result = append(result, res...)
|
|
||||||
return cursor
|
|
||||||
}
|
|
||||||
|
|
||||||
const batchSize = 10
|
|
||||||
last := treeAppend(t, nil, batchSize)
|
|
||||||
for i := 1; i < count/batchSize; i++ {
|
|
||||||
last = treeAppend(t, last, batchSize)
|
|
||||||
}
|
|
||||||
require.Len(t, result, count)
|
|
||||||
require.True(t, slices.IsSortedFunc(result, func(a, b MultiNodeInfo) int {
|
|
||||||
filenameA := findAttr(a.Meta, AttributeFilename)
|
|
||||||
filenameB := findAttr(b.Meta, AttributeFilename)
|
|
||||||
return bytes.Compare(filenameA, filenameB)
|
|
||||||
}))
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestForest_TreeSortedIteration(t *testing.T) {
|
func TestForest_TreeSortedIteration(t *testing.T) {
|
||||||
for i := range providers {
|
for i := range providers {
|
||||||
t.Run(providers[i].name, func(t *testing.T) {
|
t.Run(providers[i].name, func(t *testing.T) {
|
||||||
|
|
|
@ -2,8 +2,6 @@ package pilorama
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"container/heap"
|
"container/heap"
|
||||||
"slices"
|
|
||||||
"strings"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
type heapInfo struct {
|
type heapInfo struct {
|
||||||
|
@ -30,10 +28,9 @@ func (h *filenameHeap) Pop() any {
|
||||||
|
|
||||||
// fixedHeap maintains a fixed number of smallest elements started at some point.
|
// fixedHeap maintains a fixed number of smallest elements started at some point.
|
||||||
type fixedHeap struct {
|
type fixedHeap struct {
|
||||||
start *string
|
start *string
|
||||||
sorted bool
|
count int
|
||||||
count int
|
h *filenameHeap
|
||||||
h *filenameHeap
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func newHeap(start *string, count int) *fixedHeap {
|
func newHeap(start *string, count int) *fixedHeap {
|
||||||
|
@ -47,39 +44,20 @@ func newHeap(start *string, count int) *fixedHeap {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
const amortizationMultiplier = 5
|
|
||||||
|
|
||||||
func (h *fixedHeap) push(id MultiNode, filename string) bool {
|
func (h *fixedHeap) push(id MultiNode, filename string) bool {
|
||||||
if h.start != nil && filename <= *h.start {
|
if h.start != nil && filename <= *h.start {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
heap.Push(h.h, heapInfo{id: id, filename: filename})
|
||||||
*h.h = append(*h.h, heapInfo{id: id, filename: filename})
|
if h.h.Len() > h.count {
|
||||||
h.sorted = false
|
heap.Remove(h.h, h.h.Len()-1)
|
||||||
|
|
||||||
if h.h.Len() > h.count*amortizationMultiplier {
|
|
||||||
slices.SortFunc(*h.h, func(a, b heapInfo) int {
|
|
||||||
return strings.Compare(a.filename, b.filename)
|
|
||||||
})
|
|
||||||
*h.h = (*h.h)[:h.count]
|
|
||||||
}
|
}
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
func (h *fixedHeap) pop() (heapInfo, bool) {
|
func (h *fixedHeap) pop() (heapInfo, bool) {
|
||||||
if !h.sorted {
|
if h.h.Len() != 0 {
|
||||||
slices.SortFunc(*h.h, func(a, b heapInfo) int {
|
return heap.Pop(h.h).(heapInfo), true
|
||||||
return strings.Compare(a.filename, b.filename)
|
|
||||||
})
|
|
||||||
if len(*h.h) > h.count {
|
|
||||||
*h.h = (*h.h)[:h.count]
|
|
||||||
}
|
|
||||||
h.sorted = true
|
|
||||||
}
|
|
||||||
if len(*h.h) != 0 {
|
|
||||||
info := (*h.h)[0]
|
|
||||||
*h.h = (*h.h)[1:]
|
|
||||||
return info, true
|
|
||||||
}
|
}
|
||||||
return heapInfo{}, false
|
return heapInfo{}, false
|
||||||
}
|
}
|
||||||
|
|
|
@ -641,8 +641,8 @@ func (c *Client) notaryTxValidationLimit() (uint32, error) {
|
||||||
return 0, fmt.Errorf("can't get current blockchain height: %w", err)
|
return 0, fmt.Errorf("can't get current blockchain height: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
minTime := bc + c.notary.txValidTime
|
min := bc + c.notary.txValidTime
|
||||||
rounded := (minTime/c.notary.roundTime + 1) * c.notary.roundTime
|
rounded := (min/c.notary.roundTime + 1) * c.notary.roundTime
|
||||||
|
|
||||||
return rounded, nil
|
return rounded, nil
|
||||||
}
|
}
|
||||||
|
|
|
@ -33,15 +33,10 @@ func (a *auditService) AddChain(ctx context.Context, req *apemanager.AddChainReq
|
||||||
return res, err
|
return res, err
|
||||||
}
|
}
|
||||||
|
|
||||||
var respChainID []byte
|
|
||||||
if respBody := res.GetBody(); respBody != nil {
|
|
||||||
respChainID = respBody.GetChainID()
|
|
||||||
}
|
|
||||||
|
|
||||||
audit.LogRequest(a.log, ape_grpc.APEManagerService_AddChain_FullMethodName, req,
|
audit.LogRequest(a.log, ape_grpc.APEManagerService_AddChain_FullMethodName, req,
|
||||||
audit.TargetFromChainID(req.GetBody().GetTarget().GetTargetType().String(),
|
audit.TargetFromChainID(req.GetBody().GetTarget().GetTargetType().String(),
|
||||||
req.GetBody().GetTarget().GetName(),
|
req.GetBody().GetTarget().GetName(),
|
||||||
respChainID),
|
res.GetBody().GetChainID()),
|
||||||
err == nil)
|
err == nil)
|
||||||
|
|
||||||
return res, err
|
return res, err
|
||||||
|
|
|
@ -2,11 +2,9 @@ package getsvc
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"errors"
|
|
||||||
|
|
||||||
"git.frostfs.info/TrueCloudLab/frostfs-node/internal/logs"
|
"git.frostfs.info/TrueCloudLab/frostfs-node/internal/logs"
|
||||||
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/util"
|
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/util"
|
||||||
apistatus "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/client/status"
|
|
||||||
objectSDK "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/object"
|
objectSDK "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/object"
|
||||||
"go.uber.org/zap"
|
"go.uber.org/zap"
|
||||||
)
|
)
|
||||||
|
@ -122,12 +120,6 @@ func (exec *request) analyzeStatus(ctx context.Context, execCnr bool) {
|
||||||
exec.log.Debug(logs.OperationFinishedWithError,
|
exec.log.Debug(logs.OperationFinishedWithError,
|
||||||
zap.Error(exec.err),
|
zap.Error(exec.err),
|
||||||
)
|
)
|
||||||
var errAccessDenied *apistatus.ObjectAccessDenied
|
|
||||||
if execCnr && errors.As(exec.err, &errAccessDenied) {
|
|
||||||
// Local get can't return access denied error, so this error was returned by
|
|
||||||
// write to the output stream. So there is no need to try to find object on other nodes.
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
if execCnr {
|
if execCnr {
|
||||||
exec.executeOnContainer(ctx)
|
exec.executeOnContainer(ctx)
|
||||||
|
|
|
@ -31,7 +31,6 @@ func (r *request) processNode(ctx context.Context, info client.NodeInfo) bool {
|
||||||
var errECInfo *objectSDK.ECInfoError
|
var errECInfo *objectSDK.ECInfoError
|
||||||
var errRemoved *apistatus.ObjectAlreadyRemoved
|
var errRemoved *apistatus.ObjectAlreadyRemoved
|
||||||
var errOutOfRange *apistatus.ObjectOutOfRange
|
var errOutOfRange *apistatus.ObjectOutOfRange
|
||||||
var errAccessDenied *apistatus.ObjectAccessDenied
|
|
||||||
|
|
||||||
switch {
|
switch {
|
||||||
default:
|
default:
|
||||||
|
@ -39,11 +38,7 @@ func (r *request) processNode(ctx context.Context, info client.NodeInfo) bool {
|
||||||
if r.status != statusEC {
|
if r.status != statusEC {
|
||||||
// for raw requests, continue to collect other parts
|
// for raw requests, continue to collect other parts
|
||||||
r.status = statusUndefined
|
r.status = statusUndefined
|
||||||
if errors.As(err, &errAccessDenied) {
|
r.err = new(apistatus.ObjectNotFound)
|
||||||
r.err = err
|
|
||||||
} else {
|
|
||||||
r.err = new(apistatus.ObjectNotFound)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
return false
|
return false
|
||||||
case err == nil:
|
case err == nil:
|
||||||
|
|
|
@ -23,14 +23,12 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
type getRequestForwarder struct {
|
type getRequestForwarder struct {
|
||||||
OnceResign sync.Once
|
OnceResign sync.Once
|
||||||
GlobalProgress int
|
OnceHeaderSending sync.Once
|
||||||
Key *ecdsa.PrivateKey
|
GlobalProgress int
|
||||||
Request *objectV2.GetRequest
|
Key *ecdsa.PrivateKey
|
||||||
Stream *streamObjectWriter
|
Request *objectV2.GetRequest
|
||||||
|
Stream *streamObjectWriter
|
||||||
headerSent bool
|
|
||||||
headerSentGuard sync.Mutex
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f *getRequestForwarder) forwardRequestToNode(ctx context.Context, addr network.Address, c client.MultiAddressClient, pubkey []byte) (*objectSDK.Object, error) {
|
func (f *getRequestForwarder) forwardRequestToNode(ctx context.Context, addr network.Address, c client.MultiAddressClient, pubkey []byte) (*objectSDK.Object, error) {
|
||||||
|
@ -85,15 +83,13 @@ func (f *getRequestForwarder) writeHeader(ctx context.Context, v *objectV2.GetOb
|
||||||
obj.SetSignature(v.GetSignature())
|
obj.SetSignature(v.GetSignature())
|
||||||
obj.SetHeader(v.GetHeader())
|
obj.SetHeader(v.GetHeader())
|
||||||
|
|
||||||
f.headerSentGuard.Lock()
|
var err error
|
||||||
defer f.headerSentGuard.Unlock()
|
f.OnceHeaderSending.Do(func() {
|
||||||
if f.headerSent {
|
err = f.Stream.WriteHeader(ctx, objectSDK.NewFromV2(obj))
|
||||||
return nil
|
})
|
||||||
}
|
if err != nil {
|
||||||
if err := f.Stream.WriteHeader(ctx, objectSDK.NewFromV2(obj)); err != nil {
|
|
||||||
return errCouldNotWriteObjHeader(err)
|
return errCouldNotWriteObjHeader(err)
|
||||||
}
|
}
|
||||||
f.headerSent = true
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -24,8 +24,6 @@ type distributedTarget struct {
|
||||||
nodeTargetInitializer func(nodeDesc) preparedObjectTarget
|
nodeTargetInitializer func(nodeDesc) preparedObjectTarget
|
||||||
|
|
||||||
relay func(context.Context, nodeDesc) error
|
relay func(context.Context, nodeDesc) error
|
||||||
|
|
||||||
resetSuccessAfterOnBroadcast bool
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// parameters and state of container traversal.
|
// parameters and state of container traversal.
|
||||||
|
@ -37,8 +35,6 @@ type traversal struct {
|
||||||
|
|
||||||
// container nodes which was processed during the primary object placement
|
// container nodes which was processed during the primary object placement
|
||||||
mExclude map[string]*bool
|
mExclude map[string]*bool
|
||||||
|
|
||||||
resetSuccessAfterOnBroadcast bool
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// updates traversal parameters after the primary placement finish and
|
// updates traversal parameters after the primary placement finish and
|
||||||
|
@ -48,10 +44,6 @@ func (x *traversal) submitPrimaryPlacementFinish() bool {
|
||||||
// do not track success during container broadcast (best-effort)
|
// do not track success during container broadcast (best-effort)
|
||||||
x.opts = append(x.opts, placement.WithoutSuccessTracking())
|
x.opts = append(x.opts, placement.WithoutSuccessTracking())
|
||||||
|
|
||||||
if x.resetSuccessAfterOnBroadcast {
|
|
||||||
x.opts = append(x.opts, placement.ResetSuccessAfter())
|
|
||||||
}
|
|
||||||
|
|
||||||
// avoid 2nd broadcast
|
// avoid 2nd broadcast
|
||||||
x.extraBroadcastEnabled = false
|
x.extraBroadcastEnabled = false
|
||||||
|
|
||||||
|
@ -126,6 +118,5 @@ func (t *distributedTarget) iteratePlacement(ctx context.Context) error {
|
||||||
|
|
||||||
iter := t.cfg.newNodeIterator(append(t.placementOpts, placement.ForObject(id)))
|
iter := t.cfg.newNodeIterator(append(t.placementOpts, placement.ForObject(id)))
|
||||||
iter.extraBroadcastEnabled = needAdditionalBroadcast(t.obj, false /* Distributed target is for cluster-wide PUT */)
|
iter.extraBroadcastEnabled = needAdditionalBroadcast(t.obj, false /* Distributed target is for cluster-wide PUT */)
|
||||||
iter.resetSuccessAfterOnBroadcast = t.resetSuccessAfterOnBroadcast
|
|
||||||
return iter.forEachNode(ctx, t.sendObject)
|
return iter.forEachNode(ctx, t.sendObject)
|
||||||
}
|
}
|
||||||
|
|
|
@ -197,15 +197,14 @@ func (e *ecWriter) writeRawObject(ctx context.Context, obj *objectSDK.Object) er
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
partsProcessed := make([]atomic.Bool, len(parts))
|
|
||||||
objID, _ := obj.ID()
|
objID, _ := obj.ID()
|
||||||
t, err := placement.NewTraverser(append(e.placementOpts, placement.ForObject(objID))...)
|
t, err := placement.NewTraverser(append(e.placementOpts, placement.ForObject(objID))...)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
eg, egCtx := errgroup.WithContext(ctx)
|
||||||
for {
|
for {
|
||||||
eg, egCtx := errgroup.WithContext(ctx)
|
|
||||||
nodes := t.Next()
|
nodes := t.Next()
|
||||||
if len(nodes) == 0 {
|
if len(nodes) == 0 {
|
||||||
break
|
break
|
||||||
|
@ -217,20 +216,14 @@ func (e *ecWriter) writeRawObject(ctx context.Context, obj *objectSDK.Object) er
|
||||||
}
|
}
|
||||||
|
|
||||||
for idx := range parts {
|
for idx := range parts {
|
||||||
if !partsProcessed[idx].Load() {
|
idx := idx
|
||||||
eg.Go(func() error {
|
eg.Go(func() error {
|
||||||
err := e.writePart(egCtx, parts[idx], idx, nodes, visited)
|
return e.writePart(egCtx, parts[idx], idx, nodes, visited)
|
||||||
if err == nil {
|
})
|
||||||
partsProcessed[idx].Store(true)
|
t.SubmitSuccess()
|
||||||
t.SubmitSuccess()
|
|
||||||
}
|
|
||||||
return err
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
err = eg.Wait()
|
|
||||||
}
|
}
|
||||||
if err != nil {
|
if err := eg.Wait(); err != nil {
|
||||||
return errIncompletePut{
|
return errIncompletePut{
|
||||||
singleErr: err,
|
singleErr: err,
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,191 +0,0 @@
|
||||||
package putsvc
|
|
||||||
|
|
||||||
import (
|
|
||||||
"bytes"
|
|
||||||
"context"
|
|
||||||
"crypto/rand"
|
|
||||||
"crypto/sha256"
|
|
||||||
"errors"
|
|
||||||
"fmt"
|
|
||||||
"strconv"
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
rawclient "git.frostfs.info/TrueCloudLab/frostfs-api-go/v2/rpc/client"
|
|
||||||
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/core/client"
|
|
||||||
netmapcore "git.frostfs.info/TrueCloudLab/frostfs-node/pkg/core/netmap"
|
|
||||||
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/network"
|
|
||||||
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/services/object_manager/placement"
|
|
||||||
"git.frostfs.info/TrueCloudLab/frostfs-node/pkg/util/logger"
|
|
||||||
"git.frostfs.info/TrueCloudLab/frostfs-sdk-go/checksum"
|
|
||||||
apiclient "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/client"
|
|
||||||
"git.frostfs.info/TrueCloudLab/frostfs-sdk-go/container"
|
|
||||||
cid "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/container/id"
|
|
||||||
cidtest "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/container/id/test"
|
|
||||||
"git.frostfs.info/TrueCloudLab/frostfs-sdk-go/netmap"
|
|
||||||
objectSDK "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/object"
|
|
||||||
oid "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/object/id"
|
|
||||||
oidtest "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/object/id/test"
|
|
||||||
usertest "git.frostfs.info/TrueCloudLab/frostfs-sdk-go/user/test"
|
|
||||||
"git.frostfs.info/TrueCloudLab/frostfs-sdk-go/version"
|
|
||||||
"git.frostfs.info/TrueCloudLab/tzhash/tz"
|
|
||||||
"github.com/nspcc-dev/neo-go/pkg/crypto/keys"
|
|
||||||
"github.com/panjf2000/ants/v2"
|
|
||||||
"github.com/stretchr/testify/require"
|
|
||||||
)
|
|
||||||
|
|
||||||
type testPlacementBuilder struct {
|
|
||||||
vectors [][]netmap.NodeInfo
|
|
||||||
}
|
|
||||||
|
|
||||||
func (p *testPlacementBuilder) BuildPlacement(_ cid.ID, _ *oid.ID, _ netmap.PlacementPolicy) (
|
|
||||||
[][]netmap.NodeInfo, error,
|
|
||||||
) {
|
|
||||||
arr := make([]netmap.NodeInfo, len(p.vectors[0]))
|
|
||||||
copy(arr, p.vectors[0])
|
|
||||||
return [][]netmap.NodeInfo{arr}, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
type nmKeys struct{}
|
|
||||||
|
|
||||||
func (nmKeys) IsLocalKey(_ []byte) bool {
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
|
|
||||||
type clientConstructor struct {
|
|
||||||
vectors [][]netmap.NodeInfo
|
|
||||||
}
|
|
||||||
|
|
||||||
func (c clientConstructor) Get(info client.NodeInfo) (client.MultiAddressClient, error) {
|
|
||||||
if bytes.Equal(info.PublicKey(), c.vectors[0][0].PublicKey()) ||
|
|
||||||
bytes.Equal(info.PublicKey(), c.vectors[0][1].PublicKey()) {
|
|
||||||
return multiAddressClient{err: errors.New("node unavailable")}, nil
|
|
||||||
}
|
|
||||||
return multiAddressClient{}, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
type multiAddressClient struct {
|
|
||||||
client.MultiAddressClient
|
|
||||||
err error
|
|
||||||
}
|
|
||||||
|
|
||||||
func (c multiAddressClient) ObjectPutSingle(_ context.Context, _ apiclient.PrmObjectPutSingle) (*apiclient.ResObjectPutSingle, error) {
|
|
||||||
if c.err != nil {
|
|
||||||
return nil, c.err
|
|
||||||
}
|
|
||||||
return &apiclient.ResObjectPutSingle{}, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (c multiAddressClient) ReportError(error) {
|
|
||||||
}
|
|
||||||
|
|
||||||
func (multiAddressClient) RawForAddress(context.Context, network.Address, func(cli *rawclient.Client) error) error {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestECWriter(t *testing.T) {
|
|
||||||
// Create container with policy EC 1.1
|
|
||||||
cnr := container.Container{}
|
|
||||||
p1 := netmap.PlacementPolicy{}
|
|
||||||
p1.SetContainerBackupFactor(1)
|
|
||||||
x1 := netmap.ReplicaDescriptor{}
|
|
||||||
x1.SetECDataCount(1)
|
|
||||||
x1.SetECParityCount(1)
|
|
||||||
p1.AddReplicas(x1)
|
|
||||||
cnr.SetPlacementPolicy(p1)
|
|
||||||
cnr.SetAttribute("cnr", "cnr1")
|
|
||||||
|
|
||||||
cid := cidtest.ID()
|
|
||||||
|
|
||||||
// Create 4 nodes, 2 nodes for chunks,
|
|
||||||
// 2 nodes for the case when the first two will fail.
|
|
||||||
ns, _ := testNodeMatrix(t, []int{4})
|
|
||||||
|
|
||||||
data := make([]byte, 100)
|
|
||||||
_, _ = rand.Read(data)
|
|
||||||
ver := version.Current()
|
|
||||||
|
|
||||||
var csum checksum.Checksum
|
|
||||||
csum.SetSHA256(sha256.Sum256(data))
|
|
||||||
|
|
||||||
var csumTZ checksum.Checksum
|
|
||||||
csumTZ.SetTillichZemor(tz.Sum(csum.Value()))
|
|
||||||
|
|
||||||
obj := objectSDK.New()
|
|
||||||
obj.SetID(oidtest.ID())
|
|
||||||
obj.SetOwnerID(usertest.ID())
|
|
||||||
obj.SetContainerID(cid)
|
|
||||||
obj.SetVersion(&ver)
|
|
||||||
obj.SetPayload(data)
|
|
||||||
obj.SetPayloadSize(uint64(len(data)))
|
|
||||||
obj.SetPayloadChecksum(csum)
|
|
||||||
obj.SetPayloadHomomorphicHash(csumTZ)
|
|
||||||
|
|
||||||
// Builder return nodes without sort by hrw
|
|
||||||
builder := &testPlacementBuilder{
|
|
||||||
vectors: ns,
|
|
||||||
}
|
|
||||||
|
|
||||||
ownerKey, err := keys.NewPrivateKey()
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
pool, err := ants.NewPool(4, ants.WithNonblocking(true))
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
log, err := logger.NewLogger(nil)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
var n nmKeys
|
|
||||||
ecw := ecWriter{
|
|
||||||
cfg: &cfg{
|
|
||||||
netmapKeys: n,
|
|
||||||
remotePool: pool,
|
|
||||||
log: log,
|
|
||||||
clientConstructor: clientConstructor{vectors: ns},
|
|
||||||
},
|
|
||||||
placementOpts: append(
|
|
||||||
[]placement.Option{placement.UseBuilder(builder), placement.ForContainer(cnr)},
|
|
||||||
placement.WithCopyNumbers(nil)), // copies number ignored for EC
|
|
||||||
container: cnr,
|
|
||||||
key: &ownerKey.PrivateKey,
|
|
||||||
relay: nil,
|
|
||||||
objMetaValid: true,
|
|
||||||
}
|
|
||||||
|
|
||||||
err = ecw.WriteObject(context.Background(), obj)
|
|
||||||
require.NoError(t, err)
|
|
||||||
}
|
|
||||||
|
|
||||||
func testNodeMatrix(t testing.TB, dim []int) ([][]netmap.NodeInfo, [][]string) {
|
|
||||||
mNodes := make([][]netmap.NodeInfo, len(dim))
|
|
||||||
mAddr := make([][]string, len(dim))
|
|
||||||
|
|
||||||
for i := range dim {
|
|
||||||
ns := make([]netmap.NodeInfo, dim[i])
|
|
||||||
as := make([]string, dim[i])
|
|
||||||
|
|
||||||
for j := range dim[i] {
|
|
||||||
a := fmt.Sprintf("/ip4/192.168.0.%s/tcp/%s",
|
|
||||||
strconv.Itoa(i),
|
|
||||||
strconv.Itoa(60000+j),
|
|
||||||
)
|
|
||||||
|
|
||||||
var ni netmap.NodeInfo
|
|
||||||
ni.SetNetworkEndpoints(a)
|
|
||||||
ni.SetPublicKey([]byte(a))
|
|
||||||
|
|
||||||
var na network.AddressGroup
|
|
||||||
|
|
||||||
err := na.FromIterator(netmapcore.Node(ni))
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
as[j] = network.StringifyGroup(na)
|
|
||||||
|
|
||||||
ns[j] = ni
|
|
||||||
}
|
|
||||||
|
|
||||||
mNodes[i] = ns
|
|
||||||
mAddr[i] = as
|
|
||||||
}
|
|
||||||
|
|
||||||
return mNodes, mAddr
|
|
||||||
}
|
|
|
@ -166,7 +166,6 @@ func (s *Service) saveToNodes(ctx context.Context, obj *objectSDK.Object, req *o
|
||||||
func (s *Service) saveToREPReplicas(ctx context.Context, placement putSinglePlacement, obj *objectSDK.Object, localOnly bool, req *objectAPI.PutSingleRequest, meta object.ContentMeta) error {
|
func (s *Service) saveToREPReplicas(ctx context.Context, placement putSinglePlacement, obj *objectSDK.Object, localOnly bool, req *objectAPI.PutSingleRequest, meta object.ContentMeta) error {
|
||||||
iter := s.cfg.newNodeIterator(placement.placementOptions)
|
iter := s.cfg.newNodeIterator(placement.placementOptions)
|
||||||
iter.extraBroadcastEnabled = needAdditionalBroadcast(obj, localOnly)
|
iter.extraBroadcastEnabled = needAdditionalBroadcast(obj, localOnly)
|
||||||
iter.resetSuccessAfterOnBroadcast = placement.resetSuccessAfterOnBroadcast
|
|
||||||
|
|
||||||
signer := &putSingleRequestSigner{
|
signer := &putSingleRequestSigner{
|
||||||
req: req,
|
req: req,
|
||||||
|
@ -210,10 +209,9 @@ func (s *Service) saveToECReplicas(ctx context.Context, placement putSinglePlace
|
||||||
}
|
}
|
||||||
|
|
||||||
type putSinglePlacement struct {
|
type putSinglePlacement struct {
|
||||||
placementOptions []placement.Option
|
placementOptions []placement.Option
|
||||||
isEC bool
|
isEC bool
|
||||||
container containerSDK.Container
|
container containerSDK.Container
|
||||||
resetSuccessAfterOnBroadcast bool
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Service) getPutSinglePlacementOptions(obj *objectSDK.Object, copiesNumber []uint32, localOnly bool) (putSinglePlacement, error) {
|
func (s *Service) getPutSinglePlacementOptions(obj *objectSDK.Object, copiesNumber []uint32, localOnly bool) (putSinglePlacement, error) {
|
||||||
|
@ -234,7 +232,6 @@ func (s *Service) getPutSinglePlacementOptions(obj *objectSDK.Object, copiesNumb
|
||||||
}
|
}
|
||||||
if container.IsECContainer(cnrInfo.Value) && !object.IsECSupported(obj) && !localOnly {
|
if container.IsECContainer(cnrInfo.Value) && !object.IsECSupported(obj) && !localOnly {
|
||||||
result.placementOptions = append(result.placementOptions, placement.SuccessAfter(uint32(policy.ECParityCount(cnrInfo.Value.PlacementPolicy())+1)))
|
result.placementOptions = append(result.placementOptions, placement.SuccessAfter(uint32(policy.ECParityCount(cnrInfo.Value.PlacementPolicy())+1)))
|
||||||
result.resetSuccessAfterOnBroadcast = true
|
|
||||||
}
|
}
|
||||||
result.placementOptions = append(result.placementOptions, placement.ForContainer(cnrInfo.Value))
|
result.placementOptions = append(result.placementOptions, placement.ForContainer(cnrInfo.Value))
|
||||||
|
|
||||||
|
|
|
@ -233,19 +233,16 @@ func (p *Streamer) newDefaultObjectWriter(prm *PutInitPrm, forECPlacement bool)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
var resetSuccessAfterOnBroadcast bool
|
|
||||||
traverseOpts := prm.traverseOpts
|
traverseOpts := prm.traverseOpts
|
||||||
if forECPlacement && !prm.common.LocalOnly() {
|
if forECPlacement && !prm.common.LocalOnly() {
|
||||||
// save non-regular and linking object to EC container.
|
// save non-regular and linking object to EC container.
|
||||||
// EC 2.1 -> REP 2, EC 2.2 -> REP 3 etc.
|
// EC 2.1 -> REP 2, EC 2.2 -> REP 3 etc.
|
||||||
traverseOpts = append(traverseOpts, placement.SuccessAfter(uint32(policy.ECParityCount(prm.cnr.PlacementPolicy())+1)))
|
traverseOpts = append(traverseOpts, placement.SuccessAfter(uint32(policy.ECParityCount(prm.cnr.PlacementPolicy())+1)))
|
||||||
resetSuccessAfterOnBroadcast = true
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return &distributedTarget{
|
return &distributedTarget{
|
||||||
cfg: p.cfg,
|
cfg: p.cfg,
|
||||||
placementOpts: traverseOpts,
|
placementOpts: traverseOpts,
|
||||||
resetSuccessAfterOnBroadcast: resetSuccessAfterOnBroadcast,
|
|
||||||
nodeTargetInitializer: func(node nodeDesc) preparedObjectTarget {
|
nodeTargetInitializer: func(node nodeDesc) preparedObjectTarget {
|
||||||
if node.local {
|
if node.local {
|
||||||
return localTarget{
|
return localTarget{
|
||||||
|
|
|
@ -303,13 +303,6 @@ func SuccessAfter(v uint32) Option {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// ResetSuccessAfter resets flat success number setting option.
|
|
||||||
func ResetSuccessAfter() Option {
|
|
||||||
return func(c *cfg) {
|
|
||||||
c.flatSuccess = nil
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// WithoutSuccessTracking disables success tracking in traversal.
|
// WithoutSuccessTracking disables success tracking in traversal.
|
||||||
func WithoutSuccessTracking() Option {
|
func WithoutSuccessTracking() Option {
|
||||||
return func(c *cfg) {
|
return func(c *cfg) {
|
||||||
|
|
|
@ -138,7 +138,7 @@ func (s *Service) checkAPE(ctx context.Context, bt *bearer.Token,
|
||||||
|
|
||||||
request, err := s.newAPERequest(ctx, namespace, cid, operation, role, publicKey)
|
request, err := s.newAPERequest(ctx, namespace, cid, operation, role, publicKey)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to create ape request: %w", err)
|
return apeErr(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
var cr engine.ChainRouter
|
var cr engine.ChainRouter
|
||||||
|
@ -167,7 +167,7 @@ func (s *Service) checkAPE(ctx context.Context, bt *bearer.Token,
|
||||||
rt := engine.NewRequestTargetExtended(namespace, cid.EncodeToString(), fmt.Sprintf("%s:%s", namespace, publicKey.Address()), groups)
|
rt := engine.NewRequestTargetExtended(namespace, cid.EncodeToString(), fmt.Sprintf("%s:%s", namespace, publicKey.Address()), groups)
|
||||||
status, found, err := cr.IsAllowed(apechain.Ingress, rt, request)
|
status, found, err := cr.IsAllowed(apechain.Ingress, rt, request)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return apeErr(err)
|
||||||
}
|
}
|
||||||
if found && status == apechain.Allow {
|
if found && status == apechain.Allow {
|
||||||
return nil
|
return nil
|
||||||
|
|
|
@ -1 +0,0 @@
|
||||||
checks = ["inherit", "-SA1019"]
|
|
Loading…
Reference in a new issue