Merge pull request #3742 from nspcc-dev/default-blockfetcher-config

services: add default values for NeoFSBlockFetcher configuration
This commit is contained in:
Anna Shaleva 2024-12-13 12:17:56 +03:00 committed by GitHub
commit 38f635bce0
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
11 changed files with 115 additions and 128 deletions

View file

@ -11,6 +11,7 @@ import (
"github.com/nspcc-dev/neo-go/cli/options" "github.com/nspcc-dev/neo-go/cli/options"
"github.com/nspcc-dev/neo-go/cli/txctx" "github.com/nspcc-dev/neo-go/cli/txctx"
vmcli "github.com/nspcc-dev/neo-go/cli/vm" vmcli "github.com/nspcc-dev/neo-go/cli/vm"
"github.com/nspcc-dev/neo-go/pkg/services/helpers/neofs"
"github.com/nspcc-dev/neo-go/pkg/vm" "github.com/nspcc-dev/neo-go/pkg/vm"
"github.com/urfave/cli/v2" "github.com/urfave/cli/v2"
) )
@ -74,7 +75,7 @@ func NewCommands() []*cli.Command {
&cli.UintFlag{ &cli.UintFlag{
Name: "index-file-size", Name: "index-file-size",
Usage: "Size of index file", Usage: "Size of index file",
Value: 128000, Value: neofs.DefaultIndexFileSize,
}, },
&cli.UintFlag{ &cli.UintFlag{
Name: "workers", Name: "workers",
@ -89,7 +90,7 @@ func NewCommands() []*cli.Command {
&cli.UintFlag{ &cli.UintFlag{
Name: "retries", Name: "retries",
Usage: "Maximum number of Neo/NeoFS node request retries", Usage: "Maximum number of Neo/NeoFS node request retries",
Value: 5, Value: neofs.MaxRetries,
Action: func(context *cli.Context, u uint) error { Action: func(context *cli.Context, u uint) error {
if u < 1 { if u < 1 {
return cli.Exit("retries should be greater than 0", 1) return cli.Exit("retries should be greater than 0", 1)

View file

@ -2,7 +2,6 @@ package util
import ( import (
"context" "context"
"crypto/sha256"
"fmt" "fmt"
"slices" "slices"
"strconv" "strconv"
@ -14,7 +13,7 @@ import (
"github.com/nspcc-dev/neo-go/pkg/core/block" "github.com/nspcc-dev/neo-go/pkg/core/block"
"github.com/nspcc-dev/neo-go/pkg/io" "github.com/nspcc-dev/neo-go/pkg/io"
"github.com/nspcc-dev/neo-go/pkg/rpcclient" "github.com/nspcc-dev/neo-go/pkg/rpcclient"
"github.com/nspcc-dev/neo-go/pkg/services/oracle/neofs" "github.com/nspcc-dev/neo-go/pkg/services/helpers/neofs"
"github.com/nspcc-dev/neo-go/pkg/util" "github.com/nspcc-dev/neo-go/pkg/util"
"github.com/nspcc-dev/neo-go/pkg/wallet" "github.com/nspcc-dev/neo-go/pkg/wallet"
"github.com/nspcc-dev/neofs-sdk-go/checksum" "github.com/nspcc-dev/neofs-sdk-go/checksum"
@ -30,35 +29,6 @@ import (
"github.com/urfave/cli/v2" "github.com/urfave/cli/v2"
) )
const (
// Number of objects to search in a batch. We need to search with EQ filter to
// avoid partially-completed SEARCH responses. If EQ search haven't found object
// the object will be uploaded one more time which may lead to duplicating objects.
// We will have a risk of duplicates until #3645 is resolved (NeoFS guarantees
// search results).
searchBatchSize = 1
// Size of object ID.
oidSize = sha256.Size
)
// Constants related to retry mechanism.
const (
// Initial backoff duration.
initialBackoff = 500 * time.Millisecond
// Backoff multiplier.
backoffFactor = 2
// Maximum backoff duration.
maxBackoff = 20 * time.Second
)
// Constants related to NeoFS pool request timeouts.
// Such big values are used to avoid NeoFS pool timeouts during block search and upload.
const (
defaultDialTimeout = 10 * time.Minute
defaultStreamTimeout = 10 * time.Minute
defaultHealthcheckTimeout = 10 * time.Second
)
// poolWrapper wraps a NeoFS pool to adapt its Close method to return an error. // poolWrapper wraps a NeoFS pool to adapt its Close method to return an error.
type poolWrapper struct { type poolWrapper struct {
*pool.Pool *pool.Pool
@ -103,9 +73,9 @@ func uploadBin(ctx *cli.Context) error {
signer := user.NewAutoIDSignerRFC6979(acc.PrivateKey().PrivateKey) signer := user.NewAutoIDSignerRFC6979(acc.PrivateKey().PrivateKey)
params := pool.DefaultOptions() params := pool.DefaultOptions()
params.SetHealthcheckTimeout(defaultHealthcheckTimeout) params.SetHealthcheckTimeout(neofs.DefaultHealthcheckTimeout)
params.SetNodeDialTimeout(defaultDialTimeout) params.SetNodeDialTimeout(neofs.DefaultDialTimeout)
params.SetNodeStreamTimeout(defaultStreamTimeout) params.SetNodeStreamTimeout(neofs.DefaultStreamTimeout)
p, err := pool.New(pool.NewFlatNodeParams(rpcNeoFS), signer, params) p, err := pool.New(pool.NewFlatNodeParams(rpcNeoFS), signer, params)
if err != nil { if err != nil {
return cli.Exit(fmt.Sprintf("failed to create NeoFS pool: %v", err), 1) return cli.Exit(fmt.Sprintf("failed to create NeoFS pool: %v", err), 1)
@ -166,15 +136,15 @@ func uploadBin(ctx *cli.Context) error {
// retry function with exponential backoff. // retry function with exponential backoff.
func retry(action func() error, maxRetries uint) error { func retry(action func() error, maxRetries uint) error {
var err error var err error
backoff := initialBackoff backoff := neofs.InitialBackoff
for range maxRetries { for range maxRetries {
if err = action(); err == nil { if err = action(); err == nil {
return nil // Success, no retry needed. return nil // Success, no retry needed.
} }
time.Sleep(backoff) // Backoff before retrying. time.Sleep(backoff) // Backoff before retrying.
backoff *= time.Duration(backoffFactor) backoff *= time.Duration(neofs.BackoffFactor)
if backoff > maxBackoff { if backoff > neofs.MaxBackoff {
backoff = maxBackoff backoff = neofs.MaxBackoff
} }
} }
return err // Return the last error after exhausting retries. return err // Return the last error after exhausting retries.
@ -193,7 +163,7 @@ func uploadBlocksAndIndexFiles(ctx *cli.Context, p poolWrapper, rpc *rpcclient.C
errCh = make(chan error) errCh = make(chan error)
doneCh = make(chan struct{}) doneCh = make(chan struct{})
wg sync.WaitGroup wg sync.WaitGroup
emptyOID = make([]byte, oidSize) emptyOID = make([]byte, neofs.OIDSize)
) )
fmt.Fprintf(ctx.App.Writer, "Processing batch from %d to %d\n", indexFileStart, indexFileEnd-1) fmt.Fprintf(ctx.App.Writer, "Processing batch from %d to %d\n", indexFileStart, indexFileEnd-1)
wg.Add(int(numWorkers)) wg.Add(int(numWorkers))
@ -201,7 +171,7 @@ func uploadBlocksAndIndexFiles(ctx *cli.Context, p poolWrapper, rpc *rpcclient.C
go func(i uint) { go func(i uint) {
defer wg.Done() defer wg.Done()
for blockIndex := indexFileStart + i; blockIndex < indexFileEnd; blockIndex += numWorkers { for blockIndex := indexFileStart + i; blockIndex < indexFileEnd; blockIndex += numWorkers {
if slices.Compare(buf[blockIndex%indexFileSize*oidSize:blockIndex%indexFileSize*oidSize+oidSize], emptyOID) != 0 { if slices.Compare(buf[blockIndex%indexFileSize*neofs.OIDSize:blockIndex%indexFileSize*neofs.OIDSize+neofs.OIDSize], emptyOID) != 0 {
if debug { if debug {
fmt.Fprintf(ctx.App.Writer, "Block %d is already uploaded\n", blockIndex) fmt.Fprintf(ctx.App.Writer, "Block %d is already uploaded\n", blockIndex)
} }
@ -263,7 +233,7 @@ func uploadBlocksAndIndexFiles(ctx *cli.Context, p poolWrapper, rpc *rpcclient.C
} }
return return
} }
resOid.Encode(buf[blockIndex%indexFileSize*oidSize:]) resOid.Encode(buf[blockIndex%indexFileSize*neofs.OIDSize:])
} }
}(i) }(i)
} }
@ -281,9 +251,9 @@ func uploadBlocksAndIndexFiles(ctx *cli.Context, p poolWrapper, rpc *rpcclient.C
fmt.Fprintf(ctx.App.Writer, "Successfully processed batch of blocks: from %d to %d\n", indexFileStart, indexFileEnd-1) fmt.Fprintf(ctx.App.Writer, "Successfully processed batch of blocks: from %d to %d\n", indexFileStart, indexFileEnd-1)
// Additional check for empty OIDs in the buffer. // Additional check for empty OIDs in the buffer.
for k := uint(0); k < (indexFileEnd-indexFileStart)*oidSize; k += oidSize { for k := uint(0); k < (indexFileEnd-indexFileStart)*neofs.OIDSize; k += neofs.OIDSize {
if slices.Compare(buf[k:k+oidSize], emptyOID) == 0 { if slices.Compare(buf[k:k+neofs.OIDSize], emptyOID) == 0 {
return fmt.Errorf("empty OID found in index file %d at position %d (block index %d)", indexFileStart/indexFileSize, k/oidSize, indexFileStart/indexFileSize*indexFileSize+k/oidSize) return fmt.Errorf("empty OID found in index file %d at position %d (block index %d)", indexFileStart/indexFileSize, k/neofs.OIDSize, indexFileStart/indexFileSize*indexFileSize+k/neofs.OIDSize)
} }
} }
if indexFileEnd-indexFileStart == indexFileSize { if indexFileEnd-indexFileStart == indexFileSize {
@ -310,7 +280,7 @@ func uploadBlocksAndIndexFiles(ctx *cli.Context, p poolWrapper, rpc *rpcclient.C
func searchIndexFile(ctx *cli.Context, p poolWrapper, containerID cid.ID, account *wallet.Account, signer user.Signer, indexFileSize uint, blockAttributeKey, attributeKey string, maxParallelSearches, maxRetries uint) (uint, []byte, error) { func searchIndexFile(ctx *cli.Context, p poolWrapper, containerID cid.ID, account *wallet.Account, signer user.Signer, indexFileSize uint, blockAttributeKey, attributeKey string, maxParallelSearches, maxRetries uint) (uint, []byte, error) {
var ( var (
// buf is used to store OIDs of the uploaded blocks. // buf is used to store OIDs of the uploaded blocks.
buf = make([]byte, indexFileSize*oidSize) buf = make([]byte, indexFileSize*neofs.OIDSize)
doneCh = make(chan struct{}) doneCh = make(chan struct{})
errCh = make(chan error) errCh = make(chan error)
@ -377,7 +347,7 @@ func searchIndexFile(ctx *cli.Context, p poolWrapper, containerID cid.ID, accoun
} }
pos := uint(blockIndex) % indexFileSize pos := uint(blockIndex) % indexFileSize
if _, ok := processedIndices.LoadOrStore(pos, blockIndex); !ok { if _, ok := processedIndices.LoadOrStore(pos, blockIndex); !ok {
id.Encode(buf[pos*oidSize:]) id.Encode(buf[pos*neofs.OIDSize:])
} }
} }
}() }()
@ -404,15 +374,15 @@ func searchIndexFile(ctx *cli.Context, p poolWrapper, containerID cid.ID, accoun
// endIndex. It returns a buffered channel of resulting object IDs and closes it once // endIndex. It returns a buffered channel of resulting object IDs and closes it once
// OID search is finished. Errors are sent to errCh in a non-blocking way. // OID search is finished. Errors are sent to errCh in a non-blocking way.
func searchObjects(ctx context.Context, p poolWrapper, containerID cid.ID, account *wallet.Account, blockAttributeKey string, startIndex, endIndex, maxParallelSearches, maxRetries uint, errCh chan error, additionalFilters ...object.SearchFilters) chan oid.ID { func searchObjects(ctx context.Context, p poolWrapper, containerID cid.ID, account *wallet.Account, blockAttributeKey string, startIndex, endIndex, maxParallelSearches, maxRetries uint, errCh chan error, additionalFilters ...object.SearchFilters) chan oid.ID {
var res = make(chan oid.ID, 2*searchBatchSize) var res = make(chan oid.ID, 2*neofs.DefaultSearchBatchSize)
go func() { go func() {
var wg sync.WaitGroup var wg sync.WaitGroup
defer close(res) defer close(res)
for i := startIndex; i < endIndex; i += searchBatchSize * maxParallelSearches { for i := startIndex; i < endIndex; i += neofs.DefaultSearchBatchSize * maxParallelSearches {
for j := range maxParallelSearches { for j := range maxParallelSearches {
start := i + j*searchBatchSize start := i + j*neofs.DefaultSearchBatchSize
end := start + searchBatchSize end := start + neofs.DefaultSearchBatchSize
if start >= endIndex { if start >= endIndex {
break break

View file

@ -206,20 +206,6 @@ func TestNeoFSBlockFetcherValidation(t *testing.T) {
shouldFail: true, shouldFail: true,
errMsg: "BQueueSize (5) is lower than OIDBatchSize (10)", errMsg: "BQueueSize (5) is lower than OIDBatchSize (10)",
}, },
{
cfg: NeoFSBlockFetcher{
InternalService: InternalService{Enabled: true},
Timeout: time.Second,
ContainerID: validContainerID,
Addresses: []string{"127.0.0.1"},
OIDBatchSize: 10,
BQueueSize: 20,
SkipIndexFilesSearch: false,
IndexFileSize: 0,
},
shouldFail: true,
errMsg: "IndexFileSize is not set",
},
} }
for _, c := range cases { for _, c := range cases {

View file

@ -38,14 +38,11 @@ func (cfg *NeoFSBlockFetcher) Validate() error {
if err != nil { if err != nil {
return fmt.Errorf("invalid container ID: %w", err) return fmt.Errorf("invalid container ID: %w", err)
} }
if cfg.BQueueSize < cfg.OIDBatchSize { if cfg.BQueueSize > 0 && cfg.BQueueSize < cfg.OIDBatchSize {
return fmt.Errorf("BQueueSize (%d) is lower than OIDBatchSize (%d)", cfg.BQueueSize, cfg.OIDBatchSize) return fmt.Errorf("BQueueSize (%d) is lower than OIDBatchSize (%d)", cfg.BQueueSize, cfg.OIDBatchSize)
} }
if len(cfg.Addresses) == 0 { if len(cfg.Addresses) == 0 {
return errors.New("addresses are not set") return errors.New("addresses are not set")
} }
if !cfg.SkipIndexFilesSearch && cfg.IndexFileSize == 0 {
return errors.New("IndexFileSize is not set")
}
return nil return nil
} }

View file

@ -224,6 +224,9 @@ func newServerFromConstructors(config ServerConfig, chain Ledger, stSync StateSy
}, bqueue.DefaultCacheSize, updateBlockQueueLenMetric, bqueue.NonBlocking) }, bqueue.DefaultCacheSize, updateBlockQueueLenMetric, bqueue.NonBlocking)
s.bSyncQueue = bqueue.New(s.stateSync, log, nil, bqueue.DefaultCacheSize, updateBlockQueueLenMetric, bqueue.NonBlocking) s.bSyncQueue = bqueue.New(s.stateSync, log, nil, bqueue.DefaultCacheSize, updateBlockQueueLenMetric, bqueue.NonBlocking)
if s.NeoFSBlockFetcherCfg.BQueueSize <= 0 {
s.NeoFSBlockFetcherCfg.BQueueSize = blockfetcher.DefaultQueueCacheSize
}
s.bFetcherQueue = bqueue.New(chain, log, nil, s.NeoFSBlockFetcherCfg.BQueueSize, updateBlockQueueLenMetric, bqueue.Blocking) s.bFetcherQueue = bqueue.New(chain, log, nil, s.NeoFSBlockFetcherCfg.BQueueSize, updateBlockQueueLenMetric, bqueue.Blocking)
var err error var err error
s.blockFetcher, err = blockfetcher.New(chain, s.NeoFSBlockFetcherCfg, log, s.bFetcherQueue.PutBlock, s.blockFetcher, err = blockfetcher.New(chain, s.NeoFSBlockFetcherCfg, log, s.bFetcherQueue.PutBlock,

View file

@ -2,7 +2,6 @@ package blockfetcher
import ( import (
"context" "context"
"crypto/sha256"
"errors" "errors"
"fmt" "fmt"
"io" "io"
@ -16,7 +15,7 @@ import (
"github.com/nspcc-dev/neo-go/pkg/config" "github.com/nspcc-dev/neo-go/pkg/config"
"github.com/nspcc-dev/neo-go/pkg/core/block" "github.com/nspcc-dev/neo-go/pkg/core/block"
gio "github.com/nspcc-dev/neo-go/pkg/io" gio "github.com/nspcc-dev/neo-go/pkg/io"
"github.com/nspcc-dev/neo-go/pkg/services/oracle/neofs" "github.com/nspcc-dev/neo-go/pkg/services/helpers/neofs"
"github.com/nspcc-dev/neo-go/pkg/wallet" "github.com/nspcc-dev/neo-go/pkg/wallet"
"github.com/nspcc-dev/neofs-sdk-go/client" "github.com/nspcc-dev/neofs-sdk-go/client"
"github.com/nspcc-dev/neofs-sdk-go/container" "github.com/nspcc-dev/neofs-sdk-go/container"
@ -29,40 +28,8 @@ import (
) )
const ( const (
// oidSize is the size of the object ID in NeoFS. // DefaultQueueCacheSize is the default size of the queue cache.
oidSize = sha256.Size DefaultQueueCacheSize = 16000
// defaultTimeout is the default timeout for NeoFS requests.
defaultTimeout = 5 * time.Minute
// defaultOIDBatchSize is the default number of OIDs to search and fetch at once.
defaultOIDBatchSize = 8000
// defaultDownloaderWorkersCount is the default number of workers downloading blocks.
defaultDownloaderWorkersCount = 100
)
// Constants related to NeoFS pool request timeouts.
const (
// defaultDialTimeout is a default timeout used to establish connection with
// NeoFS storage nodes.
defaultDialTimeout = 30 * time.Second
// defaultStreamTimeout is a default timeout used for NeoFS streams processing.
// It has significantly large value to reliably avoid timeout problems with heavy
// SEARCH requests.
defaultStreamTimeout = 10 * time.Minute
// defaultHealthcheckTimeout is a timeout for request to NeoFS storage node to
// decide if it is alive.
defaultHealthcheckTimeout = 10 * time.Second
)
// Constants related to retry mechanism.
const (
// maxRetries is the maximum number of retries for a single operation.
maxRetries = 5
// initialBackoff is the initial backoff duration.
initialBackoff = 500 * time.Millisecond
// backoffFactor is the factor by which the backoff duration is multiplied.
backoffFactor = 2
// maxBackoff is the maximum backoff duration.
maxBackoff = 20 * time.Second
) )
// Ledger is an interface to Blockchain sufficient for Service. // Ledger is an interface to Blockchain sufficient for Service.
@ -143,22 +110,28 @@ func New(chain Ledger, cfg config.NeoFSBlockFetcher, logger *zap.Logger, putBloc
} }
} }
if cfg.Timeout <= 0 { if cfg.Timeout <= 0 {
cfg.Timeout = defaultTimeout cfg.Timeout = neofs.DefaultTimeout
} }
if cfg.OIDBatchSize <= 0 { if cfg.OIDBatchSize <= 0 {
cfg.OIDBatchSize = defaultOIDBatchSize cfg.OIDBatchSize = cfg.BQueueSize / 2
} }
if cfg.DownloaderWorkersCount <= 0 { if cfg.DownloaderWorkersCount <= 0 {
cfg.DownloaderWorkersCount = defaultDownloaderWorkersCount cfg.DownloaderWorkersCount = neofs.DefaultDownloaderWorkersCount
} }
if len(cfg.Addresses) == 0 { if cfg.IndexFileSize <= 0 {
return nil, errors.New("no addresses provided") cfg.IndexFileSize = neofs.DefaultIndexFileSize
}
if cfg.BlockAttribute == "" {
cfg.BlockAttribute = neofs.DefaultBlockAttribute
}
if cfg.IndexFileAttribute == "" {
cfg.IndexFileAttribute = neofs.DefaultIndexFileAttribute
} }
params := pool.DefaultOptions() params := pool.DefaultOptions()
params.SetHealthcheckTimeout(defaultHealthcheckTimeout) params.SetHealthcheckTimeout(neofs.DefaultHealthcheckTimeout)
params.SetNodeDialTimeout(defaultDialTimeout) params.SetNodeDialTimeout(neofs.DefaultDialTimeout)
params.SetNodeStreamTimeout(defaultStreamTimeout) params.SetNodeStreamTimeout(neofs.DefaultStreamTimeout)
p, err := pool.New(pool.NewFlatNodeParams(cfg.Addresses), user.NewAutoIDSignerRFC6979(account.PrivateKey().PrivateKey), params) p, err := pool.New(pool.NewFlatNodeParams(cfg.Addresses), user.NewAutoIDSignerRFC6979(account.PrivateKey().PrivateKey), params)
if err != nil { if err != nil {
return nil, err return nil, err
@ -357,7 +330,7 @@ func (bfs *Service) fetchOIDsFromIndexFiles() error {
// streamBlockOIDs reads block OIDs from the read closer and sends them to the OIDs channel. // streamBlockOIDs reads block OIDs from the read closer and sends them to the OIDs channel.
func (bfs *Service) streamBlockOIDs(rc io.ReadCloser, skip int) error { func (bfs *Service) streamBlockOIDs(rc io.ReadCloser, skip int) error {
defer rc.Close() defer rc.Close()
oidBytes := make([]byte, oidSize) oidBytes := make([]byte, neofs.OIDSize)
oidsProcessed := 0 oidsProcessed := 0
for { for {
@ -397,7 +370,7 @@ func (bfs *Service) streamBlockOIDs(rc io.ReadCloser, skip int) error {
func (bfs *Service) fetchOIDsBySearch() error { func (bfs *Service) fetchOIDsBySearch() error {
startIndex := bfs.chain.BlockHeight() startIndex := bfs.chain.BlockHeight()
//We need to search with EQ filter to avoid partially-completed SEARCH responses. //We need to search with EQ filter to avoid partially-completed SEARCH responses.
batchSize := uint32(1) batchSize := uint32(neofs.DefaultSearchBatchSize)
for { for {
select { select {
@ -513,7 +486,7 @@ func (bfs *Service) IsActive() bool {
func (bfs *Service) retry(action func() error) error { func (bfs *Service) retry(action func() error) error {
var ( var (
err error err error
backoff = initialBackoff backoff = neofs.InitialBackoff
timer = time.NewTimer(0) timer = time.NewTimer(0)
) )
defer func() { defer func() {
@ -525,11 +498,11 @@ func (bfs *Service) retry(action func() error) error {
} }
}() }()
for i := range maxRetries { for i := range neofs.MaxRetries {
if err = action(); err == nil { if err = action(); err == nil {
return nil return nil
} }
if i == maxRetries-1 { if i == neofs.MaxRetries-1 {
break break
} }
timer.Reset(backoff) timer.Reset(backoff)
@ -539,16 +512,16 @@ func (bfs *Service) retry(action func() error) error {
case <-bfs.ctx.Done(): case <-bfs.ctx.Done():
return bfs.ctx.Err() return bfs.ctx.Err()
} }
backoff *= time.Duration(backoffFactor) backoff *= time.Duration(neofs.BackoffFactor)
if backoff > maxBackoff { if backoff > neofs.MaxBackoff {
backoff = maxBackoff backoff = neofs.MaxBackoff
} }
} }
return err return err
} }
func (bfs *Service) objectGet(ctx context.Context, oid string) (io.ReadCloser, error) { func (bfs *Service) objectGet(ctx context.Context, oid string) (io.ReadCloser, error) {
u, err := url.Parse(fmt.Sprintf("neofs:%s/%s", bfs.cfg.ContainerID, oid)) u, err := url.Parse(fmt.Sprintf("%s:%s/%s", neofs.URIScheme, bfs.cfg.ContainerID, oid))
if err != nil { if err != nil {
return nil, err return nil, err
} }

View file

@ -5,6 +5,7 @@ import (
"github.com/nspcc-dev/neo-go/pkg/config" "github.com/nspcc-dev/neo-go/pkg/config"
"github.com/nspcc-dev/neo-go/pkg/core/block" "github.com/nspcc-dev/neo-go/pkg/core/block"
"github.com/nspcc-dev/neo-go/pkg/services/helpers/neofs"
"github.com/stretchr/testify/require" "github.com/stretchr/testify/require"
"go.uber.org/zap" "go.uber.org/zap"
) )
@ -65,16 +66,17 @@ func TestServiceConstructor(t *testing.T) {
InternalService: config.InternalService{ InternalService: config.InternalService{
Enabled: true, Enabled: true,
}, },
Addresses: []string{"localhost:8080"}, Addresses: []string{"localhost:8080"},
BQueueSize: DefaultQueueCacheSize,
} }
service, err := New(ledger, cfg, logger, mockPut.putBlock, shutdownCallback) service, err := New(ledger, cfg, logger, mockPut.putBlock, shutdownCallback)
require.NoError(t, err) require.NoError(t, err)
require.NotNil(t, service) require.NotNil(t, service)
require.Equal(t, service.IsActive(), false) require.Equal(t, service.IsActive(), false)
require.Equal(t, service.cfg.Timeout, defaultTimeout) require.Equal(t, service.cfg.Timeout, neofs.DefaultTimeout)
require.Equal(t, service.cfg.OIDBatchSize, defaultOIDBatchSize) require.Equal(t, service.cfg.OIDBatchSize, DefaultQueueCacheSize/2)
require.Equal(t, service.cfg.DownloaderWorkersCount, defaultDownloaderWorkersCount) require.Equal(t, service.cfg.DownloaderWorkersCount, neofs.DefaultDownloaderWorkersCount)
require.Equal(t, service.IsActive(), false) require.Equal(t, service.IsActive(), false)
}) })

View file

@ -0,0 +1,55 @@
package neofs
import (
"crypto/sha256"
"time"
)
// Constants related to NeoFS block storage.
const (
// OIDSize is the size of the object ID in NeoFS.
OIDSize = sha256.Size
// DefaultTimeout is the default timeout for NeoFS requests.
DefaultTimeout = 10 * time.Minute
// DefaultDownloaderWorkersCount is the default number of workers downloading blocks.
DefaultDownloaderWorkersCount = 500
// DefaultIndexFileSize is the default size of the index file.
DefaultIndexFileSize = 128000
// DefaultBlockAttribute is the default attribute name for block objects.
DefaultBlockAttribute = "Block"
// DefaultIndexFileAttribute is the default attribute name for index file objects.
DefaultIndexFileAttribute = "Index"
// DefaultSearchBatchSize is a number of objects to search in a batch. We need to
// search with EQ filter to avoid partially-completed SEARCH responses. If EQ search
// hasn't found object the object will be uploaded one more time which may lead to
// duplicating objects. We will have a risk of duplicates until #3645 is resolved
// (NeoFS guarantees search results).
DefaultSearchBatchSize = 1
)
// Constants related to NeoFS pool request timeouts.
const (
// DefaultDialTimeout is a default timeout used to establish connection with
// NeoFS storage nodes.
DefaultDialTimeout = 30 * time.Second
// DefaultStreamTimeout is a default timeout used for NeoFS streams processing.
// It has significantly large value to reliably avoid timeout problems with heavy
// SEARCH requests.
DefaultStreamTimeout = 10 * time.Minute
// DefaultHealthcheckTimeout is a timeout for request to NeoFS storage node to
// decide if it is alive.
DefaultHealthcheckTimeout = 10 * time.Second
)
// Constants related to retry mechanism.
const (
// MaxRetries is the maximum number of retries for a single operation.
MaxRetries = 5
// InitialBackoff is the initial backoff duration.
InitialBackoff = 500 * time.Millisecond
// BackoffFactor is the factor by which the backoff duration is multiplied.
BackoffFactor = 2
// MaxBackoff is the maximum backoff duration.
MaxBackoff = 20 * time.Second
)

View file

@ -13,7 +13,7 @@ import (
"github.com/nspcc-dev/neo-go/pkg/core/storage" "github.com/nspcc-dev/neo-go/pkg/core/storage"
"github.com/nspcc-dev/neo-go/pkg/core/transaction" "github.com/nspcc-dev/neo-go/pkg/core/transaction"
"github.com/nspcc-dev/neo-go/pkg/crypto/keys" "github.com/nspcc-dev/neo-go/pkg/crypto/keys"
"github.com/nspcc-dev/neo-go/pkg/services/oracle/neofs" "github.com/nspcc-dev/neo-go/pkg/services/helpers/neofs"
"go.uber.org/zap" "go.uber.org/zap"
) )