forked from TrueCloudLab/frostfs-node
[#1437] node: Use ctx for logging
Signed-off-by: Dmitrii Stepanov <d.stepanov@yadro.com>
This commit is contained in:
parent
c16dae8b4d
commit
6db46257c0
157 changed files with 764 additions and 713 deletions
|
@ -92,7 +92,7 @@ func (s *Service) synchronizeAllTrees(ctx context.Context, cid cid.ID) error {
|
|||
for _, tid := range treesToSync {
|
||||
h, err := s.forest.TreeLastSyncHeight(ctx, cid, tid)
|
||||
if err != nil && !errors.Is(err, pilorama.ErrTreeNotFound) {
|
||||
s.log.Warn(logs.TreeCouldNotGetLastSynchronizedHeightForATree,
|
||||
s.log.Warn(ctx, logs.TreeCouldNotGetLastSynchronizedHeightForATree,
|
||||
zap.Stringer("cid", cid),
|
||||
zap.String("tree", tid))
|
||||
continue
|
||||
|
@ -100,7 +100,7 @@ func (s *Service) synchronizeAllTrees(ctx context.Context, cid cid.ID) error {
|
|||
newHeight := s.synchronizeTree(ctx, cid, h, tid, nodes)
|
||||
if h < newHeight {
|
||||
if err := s.forest.TreeUpdateLastSyncHeight(ctx, cid, tid, newHeight); err != nil {
|
||||
s.log.Warn(logs.TreeCouldNotUpdateLastSynchronizedHeightForATree,
|
||||
s.log.Warn(ctx, logs.TreeCouldNotUpdateLastSynchronizedHeightForATree,
|
||||
zap.Stringer("cid", cid),
|
||||
zap.String("tree", tid))
|
||||
}
|
||||
|
@ -251,7 +251,7 @@ func (s *Service) startStream(ctx context.Context, cid cid.ID, treeID string,
|
|||
func (s *Service) synchronizeTree(ctx context.Context, cid cid.ID, from uint64,
|
||||
treeID string, nodes []netmapSDK.NodeInfo,
|
||||
) uint64 {
|
||||
s.log.Debug(logs.TreeSynchronizeTree, zap.Stringer("cid", cid), zap.String("tree", treeID), zap.Uint64("from", from))
|
||||
s.log.Debug(ctx, logs.TreeSynchronizeTree, zap.Stringer("cid", cid), zap.String("tree", treeID), zap.Uint64("from", from))
|
||||
|
||||
errGroup, egCtx := errgroup.WithContext(ctx)
|
||||
const workersCount = 1024
|
||||
|
@ -282,20 +282,20 @@ func (s *Service) synchronizeTree(ctx context.Context, cid cid.ID, from uint64,
|
|||
n.IterateNetworkEndpoints(func(addr string) bool {
|
||||
var a network.Address
|
||||
if err := a.FromString(addr); err != nil {
|
||||
s.log.Warn(logs.TreeFailedToParseAddressForTreeSynchronization, zap.Error(err), zap.String("address", addr))
|
||||
s.log.Warn(ctx, logs.TreeFailedToParseAddressForTreeSynchronization, zap.Error(err), zap.String("address", addr))
|
||||
return false
|
||||
}
|
||||
|
||||
cc, err := s.createConnection(a)
|
||||
if err != nil {
|
||||
s.log.Warn(logs.TreeFailedToConnectForTreeSynchronization, zap.Error(err), zap.String("address", addr))
|
||||
s.log.Warn(ctx, logs.TreeFailedToConnectForTreeSynchronization, zap.Error(err), zap.String("address", addr))
|
||||
return false
|
||||
}
|
||||
defer cc.Close()
|
||||
|
||||
err = s.startStream(egCtx, cid, treeID, from, cc, nodeOperationStreams[i])
|
||||
if err != nil {
|
||||
s.log.Warn(logs.TreeFailedToRunTreeSynchronizationForSpecificNode, zap.Error(err), zap.String("address", addr))
|
||||
s.log.Warn(ctx, logs.TreeFailedToRunTreeSynchronizationForSpecificNode, zap.Error(err), zap.String("address", addr))
|
||||
}
|
||||
nodeSynced = err == nil
|
||||
return true
|
||||
|
@ -309,7 +309,7 @@ func (s *Service) synchronizeTree(ctx context.Context, cid cid.ID, from uint64,
|
|||
}
|
||||
if err := errGroup.Wait(); err != nil {
|
||||
allNodesSynced.Store(false)
|
||||
s.log.Warn(logs.TreeFailedToRunTreeSynchronizationOverAllNodes, zap.Error(err))
|
||||
s.log.Warn(ctx, logs.TreeFailedToRunTreeSynchronizationOverAllNodes, zap.Error(err))
|
||||
}
|
||||
|
||||
newHeight := minStreamedLastHeight
|
||||
|
@ -376,13 +376,13 @@ func (s *Service) syncLoop(ctx context.Context) {
|
|||
return
|
||||
case <-s.syncChan:
|
||||
ctx, span := tracing.StartSpanFromContext(ctx, "TreeService.sync")
|
||||
s.log.Info(logs.TreeSyncingTrees)
|
||||
s.log.Info(ctx, logs.TreeSyncingTrees)
|
||||
|
||||
start := time.Now()
|
||||
|
||||
cnrs, err := s.cfg.cnrSource.List()
|
||||
if err != nil {
|
||||
s.log.Error(logs.TreeCouldNotFetchContainers, zap.Error(err))
|
||||
s.log.Error(ctx, logs.TreeCouldNotFetchContainers, zap.Error(err))
|
||||
s.metrics.AddSyncDuration(time.Since(start), false)
|
||||
span.End()
|
||||
break
|
||||
|
@ -394,7 +394,7 @@ func (s *Service) syncLoop(ctx context.Context) {
|
|||
|
||||
s.removeContainers(ctx, newMap)
|
||||
|
||||
s.log.Info(logs.TreeTreesHaveBeenSynchronized)
|
||||
s.log.Info(ctx, logs.TreeTreesHaveBeenSynchronized)
|
||||
|
||||
s.metrics.AddSyncDuration(time.Since(start), true)
|
||||
span.End()
|
||||
|
@ -414,19 +414,19 @@ func (s *Service) syncContainers(ctx context.Context, cnrs []cid.ID) {
|
|||
|
||||
err := s.syncPool.Submit(func() {
|
||||
defer wg.Done()
|
||||
s.log.Debug(logs.TreeSyncingContainerTrees, zap.Stringer("cid", cnr))
|
||||
s.log.Debug(ctx, logs.TreeSyncingContainerTrees, zap.Stringer("cid", cnr))
|
||||
|
||||
err := s.synchronizeAllTrees(ctx, cnr)
|
||||
if err != nil {
|
||||
s.log.Error(logs.TreeCouldNotSyncTrees, zap.Stringer("cid", cnr), zap.Error(err))
|
||||
s.log.Error(ctx, logs.TreeCouldNotSyncTrees, zap.Stringer("cid", cnr), zap.Error(err))
|
||||
return
|
||||
}
|
||||
|
||||
s.log.Debug(logs.TreeContainerTreesHaveBeenSynced, zap.Stringer("cid", cnr))
|
||||
s.log.Debug(ctx, logs.TreeContainerTreesHaveBeenSynced, zap.Stringer("cid", cnr))
|
||||
})
|
||||
if err != nil {
|
||||
wg.Done()
|
||||
s.log.Error(logs.TreeCouldNotQueryTreesForSynchronization,
|
||||
s.log.Error(ctx, logs.TreeCouldNotQueryTreesForSynchronization,
|
||||
zap.Stringer("cid", cnr),
|
||||
zap.Error(err))
|
||||
if errors.Is(err, ants.ErrPoolClosed) {
|
||||
|
@ -452,7 +452,7 @@ func (s *Service) removeContainers(ctx context.Context, newContainers map[cid.ID
|
|||
|
||||
existed, err := containerCore.WasRemoved(s.cnrSource, cnr)
|
||||
if err != nil {
|
||||
s.log.Error(logs.TreeCouldNotCheckIfContainerExisted,
|
||||
s.log.Error(ctx, logs.TreeCouldNotCheckIfContainerExisted,
|
||||
zap.Stringer("cid", cnr),
|
||||
zap.Error(err))
|
||||
} else if existed {
|
||||
|
@ -464,11 +464,11 @@ func (s *Service) removeContainers(ctx context.Context, newContainers map[cid.ID
|
|||
}
|
||||
|
||||
for _, cnr := range removed {
|
||||
s.log.Debug(logs.TreeRemovingRedundantTrees, zap.Stringer("cid", cnr))
|
||||
s.log.Debug(ctx, logs.TreeRemovingRedundantTrees, zap.Stringer("cid", cnr))
|
||||
|
||||
err := s.DropTree(ctx, cnr, "")
|
||||
if err != nil {
|
||||
s.log.Error(logs.TreeCouldNotRemoveRedundantTree,
|
||||
s.log.Error(ctx, logs.TreeCouldNotRemoveRedundantTree,
|
||||
zap.Stringer("cid", cnr),
|
||||
zap.Error(err))
|
||||
}
|
||||
|
@ -482,7 +482,7 @@ func (s *Service) containersToSync(cnrs []cid.ID) (map[cid.ID]struct{}, []cid.ID
|
|||
for _, cnr := range cnrs {
|
||||
_, pos, err := s.getContainerNodes(cnr)
|
||||
if err != nil {
|
||||
s.log.Error(logs.TreeCouldNotCalculateContainerNodes,
|
||||
s.log.Error(context.Background(), logs.TreeCouldNotCalculateContainerNodes,
|
||||
zap.Stringer("cid", cnr),
|
||||
zap.Error(err))
|
||||
continue
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue