forked from TrueCloudLab/frostfs-node
e1be0180f6
Signed-off-by: Pavel Karpy <carpawell@nspcc.ru>
312 lines
7.4 KiB
Go
312 lines
7.4 KiB
Go
package tree
|
|
|
|
import (
|
|
"context"
|
|
"crypto/sha256"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
|
|
"github.com/nspcc-dev/neofs-node/pkg/local_object_storage/pilorama"
|
|
"github.com/nspcc-dev/neofs-node/pkg/network"
|
|
cid "github.com/nspcc-dev/neofs-sdk-go/container/id"
|
|
netmapSDK "github.com/nspcc-dev/neofs-sdk-go/netmap"
|
|
"go.uber.org/zap"
|
|
"google.golang.org/grpc"
|
|
"google.golang.org/grpc/credentials/insecure"
|
|
)
|
|
|
|
// ErrNotInContainer is returned when operation could not be performed
|
|
// because the node is not included in the container.
|
|
var ErrNotInContainer = errors.New("node is not in container")
|
|
var errNoOtherNodes = errors.New("no nodes to fetch trees from")
|
|
|
|
// SynchronizeAllTrees synchronizes all the trees of the container. It fetches
|
|
// tree IDs from the other container nodes. Returns ErrNotInContainer if the node
|
|
// is not included in the container.
|
|
func (s *Service) SynchronizeAllTrees(ctx context.Context, cid cid.ID) error {
|
|
nodes, pos, err := s.getContainerNodes(cid)
|
|
if err != nil {
|
|
return fmt.Errorf("can't get container nodes: %w", err)
|
|
}
|
|
|
|
if pos < 0 {
|
|
return ErrNotInContainer
|
|
}
|
|
|
|
var d pilorama.CIDDescriptor
|
|
d.CID = cid
|
|
d.Position = pos
|
|
d.Size = len(nodes)
|
|
|
|
nodes = append(nodes[:pos], nodes[pos+1:]...) // exclude that node
|
|
if len(nodes) == 0 {
|
|
return errNoOtherNodes
|
|
}
|
|
|
|
rawCID := make([]byte, sha256.Size)
|
|
cid.Encode(rawCID)
|
|
|
|
req := &TreeListRequest{
|
|
Body: &TreeListRequest_Body{
|
|
ContainerId: rawCID,
|
|
},
|
|
}
|
|
|
|
err = SignMessage(req, s.key)
|
|
if err != nil {
|
|
return fmt.Errorf("could not sign request: %w", err)
|
|
}
|
|
|
|
var resp *TreeListResponse
|
|
var treesToSync []string
|
|
var outErr error
|
|
|
|
err = s.forEachNode(ctx, nodes, func(c TreeServiceClient) bool {
|
|
resp, outErr = c.TreeList(ctx, req)
|
|
if outErr != nil {
|
|
return false
|
|
}
|
|
|
|
treesToSync = resp.GetBody().GetIds()
|
|
|
|
return true
|
|
})
|
|
if err != nil {
|
|
outErr = err
|
|
}
|
|
|
|
if outErr != nil {
|
|
return fmt.Errorf("could not fetch tree ID list: %w", outErr)
|
|
}
|
|
|
|
for _, tid := range treesToSync {
|
|
err = s.synchronizeTree(ctx, d, tid, nodes)
|
|
if err != nil {
|
|
s.log.Error("could not sync tree",
|
|
zap.Stringer("cid", cid),
|
|
zap.String("treeID", tid))
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// SynchronizeTree tries to synchronize log starting from the last stored height.
|
|
func (s *Service) SynchronizeTree(ctx context.Context, cid cid.ID, treeID string) error {
|
|
nodes, pos, err := s.getContainerNodes(cid)
|
|
if err != nil {
|
|
return fmt.Errorf("can't get container nodes: %w", err)
|
|
}
|
|
|
|
if pos < 0 {
|
|
return ErrNotInContainer
|
|
}
|
|
|
|
var d pilorama.CIDDescriptor
|
|
d.CID = cid
|
|
d.Position = pos
|
|
d.Size = len(nodes)
|
|
|
|
nodes = append(nodes[:pos], nodes[pos+1:]...) // exclude that node
|
|
if len(nodes) == 0 {
|
|
return errNoOtherNodes
|
|
}
|
|
|
|
return s.synchronizeTree(ctx, d, treeID, nodes)
|
|
}
|
|
|
|
func (s *Service) synchronizeTree(ctx context.Context, d pilorama.CIDDescriptor,
|
|
treeID string, nodes []netmapSDK.NodeInfo) error {
|
|
lm, err := s.forest.TreeGetOpLog(d.CID, treeID, 0)
|
|
if err != nil && !errors.Is(err, pilorama.ErrTreeNotFound) {
|
|
return err
|
|
}
|
|
|
|
height := lm.Time + 1
|
|
for _, n := range nodes {
|
|
n.IterateNetworkEndpoints(func(addr string) bool {
|
|
var a network.Address
|
|
if err := a.FromString(addr); err != nil {
|
|
return false
|
|
}
|
|
|
|
cc, err := grpc.DialContext(ctx, a.URIAddr(), grpc.WithTransportCredentials(insecure.NewCredentials()))
|
|
if err != nil {
|
|
// Failed to connect, try the next address.
|
|
return false
|
|
}
|
|
|
|
treeClient := NewTreeServiceClient(cc)
|
|
for {
|
|
h, err := s.synchronizeSingle(ctx, d, treeID, height, treeClient)
|
|
if height < h {
|
|
height = h
|
|
}
|
|
if err != nil || h <= height {
|
|
// Error with the response, try the next node.
|
|
return true
|
|
}
|
|
}
|
|
})
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (s *Service) synchronizeSingle(ctx context.Context, d pilorama.CIDDescriptor, treeID string, height uint64, treeClient TreeServiceClient) (uint64, error) {
|
|
rawCID := make([]byte, sha256.Size)
|
|
d.CID.Encode(rawCID)
|
|
|
|
for {
|
|
newHeight := height
|
|
req := &GetOpLogRequest{
|
|
Body: &GetOpLogRequest_Body{
|
|
ContainerId: rawCID,
|
|
TreeId: treeID,
|
|
Height: newHeight,
|
|
},
|
|
}
|
|
if err := SignMessage(req, s.key); err != nil {
|
|
return newHeight, err
|
|
}
|
|
|
|
c, err := treeClient.GetOpLog(ctx, req)
|
|
if err != nil {
|
|
return newHeight, fmt.Errorf("can't initialize client: %w", err)
|
|
}
|
|
|
|
res, err := c.Recv()
|
|
for ; err == nil; res, err = c.Recv() {
|
|
lm := res.GetBody().GetOperation()
|
|
m := &pilorama.Move{
|
|
Parent: lm.ParentId,
|
|
Child: lm.ChildId,
|
|
}
|
|
if err := m.Meta.FromBytes(lm.Meta); err != nil {
|
|
return newHeight, err
|
|
}
|
|
if err := s.forest.TreeApply(d, treeID, m); err != nil {
|
|
return newHeight, err
|
|
}
|
|
if m.Time > newHeight {
|
|
newHeight = m.Time + 1
|
|
} else {
|
|
newHeight++
|
|
}
|
|
}
|
|
if height == newHeight || err != nil && !errors.Is(err, io.EOF) {
|
|
return newHeight, err
|
|
}
|
|
height = newHeight
|
|
}
|
|
}
|
|
|
|
// ErrAlreadySyncing is returned when a service synchronization has already
|
|
// been started.
|
|
var ErrAlreadySyncing = errors.New("service is being synchronized")
|
|
|
|
// ErrShuttingDown is returned when the service is shitting down and could not
|
|
// accept any calls.
|
|
var ErrShuttingDown = errors.New("service is shutting down")
|
|
|
|
// SynchronizeAll forces tree service to synchronize all the trees according to
|
|
// netmap information. Must not be called before Service.Start.
|
|
// Returns ErrAlreadySyncing if synchronization has been started and blocked
|
|
// by another routine.
|
|
// Note: non-blocking operation.
|
|
func (s *Service) SynchronizeAll() error {
|
|
select {
|
|
case <-s.closeCh:
|
|
return ErrShuttingDown
|
|
default:
|
|
}
|
|
|
|
select {
|
|
case s.syncChan <- struct{}{}:
|
|
return nil
|
|
default:
|
|
return ErrAlreadySyncing
|
|
}
|
|
}
|
|
|
|
func (s *Service) syncLoop(ctx context.Context) {
|
|
for {
|
|
select {
|
|
case <-s.closeCh:
|
|
return
|
|
case <-ctx.Done():
|
|
return
|
|
case <-s.syncChan:
|
|
s.log.Debug("syncing trees...")
|
|
|
|
cnrs, err := s.cfg.cnrSource.List()
|
|
if err != nil {
|
|
s.log.Error("could not fetch containers", zap.Error(err))
|
|
continue
|
|
}
|
|
|
|
newMap := make(map[cid.ID]struct{}, len(s.cnrMap))
|
|
cnrsToSync := make([]cid.ID, 0, len(cnrs))
|
|
|
|
for _, cnr := range cnrs {
|
|
_, pos, err := s.getContainerNodes(cnr)
|
|
if err != nil {
|
|
s.log.Error("could not calculate container nodes",
|
|
zap.Stringer("cid", cnr),
|
|
zap.Error(err))
|
|
delete(s.cnrMap, cnr)
|
|
|
|
continue
|
|
}
|
|
|
|
if pos < 0 {
|
|
// node is not included in the container.
|
|
continue
|
|
}
|
|
|
|
_, ok := s.cnrMap[cnr]
|
|
if ok {
|
|
// known container; already in sync.
|
|
delete(s.cnrMap, cnr)
|
|
newMap[cnr] = struct{}{}
|
|
} else {
|
|
// unknown container; need to sync.
|
|
cnrsToSync = append(cnrsToSync, cnr)
|
|
}
|
|
}
|
|
|
|
// sync new containers
|
|
for _, cnr := range cnrsToSync {
|
|
s.log.Debug("syncing container trees...", zap.Stringer("cid", cnr))
|
|
|
|
err = s.SynchronizeAllTrees(ctx, cnr)
|
|
if err != nil {
|
|
s.log.Error("could not sync trees", zap.Stringer("cid", cnr), zap.Error(err))
|
|
continue
|
|
}
|
|
|
|
// mark as synced
|
|
newMap[cnr] = struct{}{}
|
|
|
|
s.log.Debug("container trees have been synced", zap.Stringer("cid", cnr))
|
|
}
|
|
|
|
// remove stored redundant trees
|
|
for cnr := range s.cnrMap {
|
|
s.log.Debug("removing redundant trees...", zap.Stringer("cid", cnr))
|
|
|
|
err = s.DropTree(ctx, cnr, "") // TODO: #1940 drop all the trees here
|
|
if err != nil {
|
|
s.log.Error("could not remove redundant tree",
|
|
zap.Stringer("cid", cnr),
|
|
zap.Error(err))
|
|
continue
|
|
}
|
|
}
|
|
|
|
s.cnrMap = newMap
|
|
|
|
s.log.Debug("trees have been synchronized")
|
|
}
|
|
}
|
|
}
|