[#271] service/audit: Implement PoR

Signed-off-by: Alex Vanin <alexey@nspcc.ru>
This commit is contained in:
Alex Vanin 2020-12-23 12:54:34 +03:00 committed by Alex Vanin
parent 5f65ec0265
commit a5320408a5
4 changed files with 115 additions and 4 deletions

View file

@ -90,6 +90,7 @@ func (c *ClientCache) GetSG(task *audit.Task, id *object.ID) (*storagegroup.Stor
cctx, cancel := context.WithTimeout(task.AuditContext(), c.sgTimeout)
obj, err := cli.GetObject(cctx, getParams)
cancel()
if err != nil {
@ -135,6 +136,7 @@ func (c *ClientCache) GetHeader(task *audit.Task, node *netmap.Node, id *object.
cctx, cancel := context.WithTimeout(task.AuditContext(), c.headTimeout)
head, err := cli.GetObjectHeader(cctx, headParams, client.WithTTL(1))
cancel()
if err != nil {

View file

@ -17,6 +17,12 @@ type Context struct {
task *audit.Task
report *audit.Report
// consider adding mutex to access caches
sgMembersCache map[int][]*object.ID
placementCache map[string][]netmap.Nodes
}
// ContextPrm groups components required to conduct data audit checks.
@ -77,6 +83,10 @@ func (c *Context) containerID() *container.ID {
func (c *Context) init() {
c.report = audit.NewReport(c.containerID())
c.sgMembersCache = make(map[int][]*object.ID)
c.placementCache = make(map[string][]netmap.Nodes)
c.log = c.log.With(
zap.Stringer("container ID", c.task.ContainerID()),
)

View file

@ -34,10 +34,6 @@ func (c *Context) Execute() {
c.writeReport()
}
func (c *Context) executePoR() {
// TODO: implement me
}
func (c *Context) executePoP() {
// TODO: implement me
}

View file

@ -0,0 +1,103 @@
package auditor
import (
"bytes"
"github.com/nspcc-dev/neofs-api-go/pkg/object"
"github.com/nspcc-dev/neofs-node/pkg/services/object_manager/placement"
"github.com/nspcc-dev/tzhash/tz"
"go.uber.org/zap"
)
func (c *Context) executePoR() {
for i, sg := range c.task.StorageGroupList() {
c.checkStorageGroupPoR(i, sg) // consider parallel it
}
}
func (c *Context) checkStorageGroupPoR(ind int, sg *object.ID) {
storageGroup, err := c.cnrCom.GetSG(c.task, sg) // get storage group
if err != nil {
c.log.Warn("can't get storage group",
zap.Stringer("sgid", sg),
zap.String("error", err.Error()))
return
}
members := storageGroup.Members()
c.sgMembersCache[ind] = members
var (
tzHash []byte
totalSize uint64
)
for i := range members {
objectPlacement, err := placement.BuildObjectPlacement(
c.task.NetworkMap(),
c.task.ContainerNodes(),
members[i],
)
if err != nil {
c.log.Info("can't build placement for storage group member",
zap.Stringer("sg", sg),
zap.Stringer("member_id", members[i]),
)
continue
}
c.placementCache[members[i].String()] = objectPlacement
for _, node := range placement.FlattenNodes(objectPlacement) {
hdr, err := c.cnrCom.GetHeader(c.task, node, members[i])
if err != nil {
c.log.Debug("can't head object",
zap.String("remote_node", node.Address()),
zap.Stringer("oid", members[i]))
continue
}
if len(tzHash) == 0 {
tzHash = hdr.PayloadHomomorphicHash().Sum()
} else {
tzHash, err = tz.Concat([][]byte{
tzHash,
hdr.PayloadHomomorphicHash().Sum(),
})
if err != nil {
c.log.Debug("can't concatenate tz hash",
zap.Stringer("oid", members[i]),
zap.String("error", err.Error()))
break
}
}
totalSize += hdr.PayloadSize()
break
}
}
sizeCheck := storageGroup.ValidationDataSize() == totalSize
tzCheck := bytes.Equal(tzHash, storageGroup.ValidationDataHash().Sum())
if sizeCheck && tzCheck {
c.report.PassedPoR(sg) // write report
} else {
if !sizeCheck {
c.log.Debug("storage group size check failed",
zap.Uint64("expected", storageGroup.ValidationDataSize()),
zap.Uint64("got", totalSize))
}
if !tzCheck {
c.log.Debug("storage group tz hash check failed")
}
c.report.FailedPoR(sg) // write report
}
}