forked from TrueCloudLab/frostfs-node
[#1689] linter: Fix staticcheck warning: 'embedded field can be simplified'
Change-Id: I8f454f7d09973cdea096495c3949b88cdd01102e Signed-off-by: Alexander Chuprov <a.chuprov@yadro.com>
This commit is contained in:
parent
923f0acf8f
commit
6f7b6b65f3
34 changed files with 121 additions and 121 deletions
|
@ -29,7 +29,7 @@ func (c *Config) NewNodeIterator(opts []placement.Option) *NodeIterator {
|
|||
}
|
||||
|
||||
func (n *NodeIterator) ForEachNode(ctx context.Context, f func(context.Context, NodeDescriptor) error) error {
|
||||
traverser, err := placement.NewTraverser(ctx, n.Traversal.Opts...)
|
||||
traverser, err := placement.NewTraverser(ctx, n.Opts...)
|
||||
if err != nil {
|
||||
return fmt.Errorf("could not create object placement traverser: %w", err)
|
||||
}
|
||||
|
@ -56,7 +56,7 @@ func (n *NodeIterator) ForEachNode(ctx context.Context, f func(context.Context,
|
|||
}
|
||||
|
||||
// perform additional container broadcast if needed
|
||||
if n.Traversal.submitPrimaryPlacementFinish() {
|
||||
if n.submitPrimaryPlacementFinish() {
|
||||
err := n.ForEachNode(ctx, f)
|
||||
if err != nil {
|
||||
n.cfg.Logger.Error(ctx, logs.PutAdditionalContainerBroadcastFailure, zap.Error(err))
|
||||
|
@ -101,7 +101,7 @@ func (n *NodeIterator) forEachAddress(ctx context.Context, traverser *placement.
|
|||
// in subsequent container broadcast. Note that we don't
|
||||
// process this node during broadcast if primary placement
|
||||
// on it failed.
|
||||
n.Traversal.submitProcessed(addr, item)
|
||||
n.submitProcessed(addr, item)
|
||||
}
|
||||
|
||||
wg.Wait()
|
||||
|
|
|
@ -146,5 +146,5 @@ func (r *request) getObjectWithIndependentRequest(ctx context.Context, prm Reque
|
|||
|
||||
detachedExecutor.execute(ctx)
|
||||
|
||||
return detachedExecutor.statusError.err
|
||||
return detachedExecutor.err
|
||||
}
|
||||
|
|
|
@ -87,7 +87,7 @@ func (s *Service) get(ctx context.Context, prm RequestParameters) error {
|
|||
|
||||
exec.execute(ctx)
|
||||
|
||||
return exec.statusError.err
|
||||
return exec.err
|
||||
}
|
||||
|
||||
func (exec *request) execute(ctx context.Context) {
|
||||
|
|
|
@ -24,14 +24,14 @@ func (s *streamObjectWriter) WriteHeader(_ context.Context, obj *objectSDK.Objec
|
|||
p.SetHeader(objV2.GetHeader())
|
||||
p.SetSignature(objV2.GetSignature())
|
||||
|
||||
return s.GetObjectStream.Send(newResponse(p))
|
||||
return s.Send(newResponse(p))
|
||||
}
|
||||
|
||||
func (s *streamObjectWriter) WriteChunk(_ context.Context, chunk []byte) error {
|
||||
p := new(objectV2.GetObjectPartChunk)
|
||||
p.SetChunk(chunk)
|
||||
|
||||
return s.GetObjectStream.Send(newResponse(p))
|
||||
return s.Send(newResponse(p))
|
||||
}
|
||||
|
||||
func newResponse(p objectV2.GetObjectPart) *objectV2.GetResponse {
|
||||
|
@ -46,7 +46,7 @@ func newResponse(p objectV2.GetObjectPart) *objectV2.GetResponse {
|
|||
}
|
||||
|
||||
func (s *streamObjectRangeWriter) WriteChunk(_ context.Context, chunk []byte) error {
|
||||
return s.GetObjectRangeStream.Send(newRangeResponse(chunk))
|
||||
return s.Send(newRangeResponse(chunk))
|
||||
}
|
||||
|
||||
func newRangeResponse(p []byte) *objectV2.GetRangeResponse {
|
||||
|
|
|
@ -28,7 +28,7 @@ func NewService(cfg *objectwriter.Config,
|
|||
|
||||
// Patch calls internal service and returns v2 object streamer.
|
||||
func (s *Service) Patch() (object.PatchObjectStream, error) {
|
||||
nodeKey, err := s.Config.KeyStorage.GetKey(nil)
|
||||
nodeKey, err := s.KeyStorage.GetKey(nil)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
|
|
@ -102,7 +102,7 @@ func (s *Service) validarePutSingleSize(ctx context.Context, obj *objectSDK.Obje
|
|||
return target.ErrWrongPayloadSize
|
||||
}
|
||||
|
||||
maxAllowedSize := s.Config.MaxSizeSrc.MaxObjectSize(ctx)
|
||||
maxAllowedSize := s.MaxSizeSrc.MaxObjectSize(ctx)
|
||||
if obj.PayloadSize() > maxAllowedSize {
|
||||
return target.ErrExceedingMaxSize
|
||||
}
|
||||
|
@ -166,13 +166,13 @@ func (s *Service) saveToNodes(ctx context.Context, obj *objectSDK.Object, req *o
|
|||
}
|
||||
|
||||
func (s *Service) saveToREPReplicas(ctx context.Context, placement putSinglePlacement, obj *objectSDK.Object, localOnly bool, req *objectAPI.PutSingleRequest, meta object.ContentMeta) error {
|
||||
iter := s.Config.NewNodeIterator(placement.placementOptions)
|
||||
iter := s.NewNodeIterator(placement.placementOptions)
|
||||
iter.ExtraBroadcastEnabled = objectwriter.NeedAdditionalBroadcast(obj, localOnly)
|
||||
iter.ResetSuccessAfterOnBroadcast = placement.resetSuccessAfterOnBroadcast
|
||||
|
||||
signer := &putSingleRequestSigner{
|
||||
req: req,
|
||||
keyStorage: s.Config.KeyStorage,
|
||||
keyStorage: s.KeyStorage,
|
||||
signer: &sync.Once{},
|
||||
}
|
||||
|
||||
|
@ -186,13 +186,13 @@ func (s *Service) saveToECReplicas(ctx context.Context, placement putSinglePlace
|
|||
if err != nil {
|
||||
return err
|
||||
}
|
||||
key, err := s.Config.KeyStorage.GetKey(nil)
|
||||
key, err := s.KeyStorage.GetKey(nil)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
signer := &putSingleRequestSigner{
|
||||
req: req,
|
||||
keyStorage: s.Config.KeyStorage,
|
||||
keyStorage: s.KeyStorage,
|
||||
signer: &sync.Once{},
|
||||
}
|
||||
|
||||
|
@ -225,7 +225,7 @@ func (s *Service) getPutSinglePlacementOptions(ctx context.Context, obj *objectS
|
|||
if !ok {
|
||||
return result, errors.New("missing container ID")
|
||||
}
|
||||
cnrInfo, err := s.Config.ContainerSource.Get(ctx, cnrID)
|
||||
cnrInfo, err := s.ContainerSource.Get(ctx, cnrID)
|
||||
if err != nil {
|
||||
return result, fmt.Errorf("could not get container by ID: %w", err)
|
||||
}
|
||||
|
@ -249,14 +249,14 @@ func (s *Service) getPutSinglePlacementOptions(ctx context.Context, obj *objectS
|
|||
}
|
||||
result.placementOptions = append(result.placementOptions, placement.ForObject(objID))
|
||||
|
||||
latestNetmap, err := netmap.GetLatestNetworkMap(ctx, s.Config.NetmapSource)
|
||||
latestNetmap, err := netmap.GetLatestNetworkMap(ctx, s.NetmapSource)
|
||||
if err != nil {
|
||||
return result, fmt.Errorf("could not get latest network map: %w", err)
|
||||
}
|
||||
builder := placement.NewNetworkMapBuilder(latestNetmap)
|
||||
if localOnly {
|
||||
result.placementOptions = append(result.placementOptions, placement.SuccessAfter(1))
|
||||
builder = svcutil.NewLocalPlacement(builder, s.Config.NetmapKeys)
|
||||
builder = svcutil.NewLocalPlacement(builder, s.NetmapKeys)
|
||||
}
|
||||
result.placementOptions = append(result.placementOptions, placement.UseBuilder(builder))
|
||||
return result, nil
|
||||
|
@ -273,7 +273,7 @@ func (s *Service) saveToPlacementNode(ctx context.Context, nodeDesc *objectwrite
|
|||
|
||||
client.NodeInfoFromNetmapElement(&info, nodeDesc.Info)
|
||||
|
||||
c, err := s.Config.ClientConstructor.Get(info)
|
||||
c, err := s.ClientConstructor.Get(info)
|
||||
if err != nil {
|
||||
return fmt.Errorf("could not create SDK client %s: %w", info.AddressGroup(), err)
|
||||
}
|
||||
|
@ -283,7 +283,7 @@ func (s *Service) saveToPlacementNode(ctx context.Context, nodeDesc *objectwrite
|
|||
|
||||
func (s *Service) saveLocal(ctx context.Context, obj *objectSDK.Object, meta object.ContentMeta, container containerSDK.Container) error {
|
||||
localTarget := &objectwriter.LocalTarget{
|
||||
Storage: s.Config.LocalStore,
|
||||
Storage: s.LocalStore,
|
||||
Container: container,
|
||||
}
|
||||
return localTarget.WriteObject(ctx, obj, meta)
|
||||
|
@ -317,7 +317,7 @@ func (s *Service) redirectPutSingleRequest(ctx context.Context,
|
|||
if err != nil {
|
||||
objID, _ := obj.ID()
|
||||
cnrID, _ := obj.ContainerID()
|
||||
s.Config.Logger.Warn(ctx, logs.PutSingleRedirectFailure,
|
||||
s.Logger.Warn(ctx, logs.PutSingleRedirectFailure,
|
||||
zap.Error(err),
|
||||
zap.Stringer("address", addr),
|
||||
zap.Stringer("object_id", objID),
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue