mirror of
https://github.com/nspcc-dev/neo-go.git
synced 2024-11-26 19:42:23 +00:00
core: clone MPT node while restoring it multiple times
We need this to avoid collapse collisions. Example of such collapse described in https://github.com/nspcc-dev/neo-go/pull/2019#discussion_r689629704.
This commit is contained in:
parent
5cd78c31af
commit
36808b8904
8 changed files with 138 additions and 1 deletions
|
@ -89,6 +89,12 @@ func (b *BranchNode) UnmarshalJSON(data []byte) error {
|
|||
return errors.New("expected branch node")
|
||||
}
|
||||
|
||||
// Clone implements Node interface.
|
||||
func (b *BranchNode) Clone() Node {
|
||||
res := *b
|
||||
return &res
|
||||
}
|
||||
|
||||
// splitPath splits path for a branch node.
|
||||
func splitPath(path []byte) (byte, []byte) {
|
||||
if len(path) != 0 {
|
||||
|
|
|
@ -54,3 +54,6 @@ func (e EmptyNode) Type() NodeType {
|
|||
func (e EmptyNode) Bytes() []byte {
|
||||
return nil
|
||||
}
|
||||
|
||||
// Clone implements Node interface.
|
||||
func (EmptyNode) Clone() Node { return EmptyNode{} }
|
||||
|
|
|
@ -98,3 +98,9 @@ func (e *ExtensionNode) UnmarshalJSON(data []byte) error {
|
|||
}
|
||||
return errors.New("expected extension node")
|
||||
}
|
||||
|
||||
// Clone implements Node interface.
|
||||
func (e *ExtensionNode) Clone() Node {
|
||||
res := *e
|
||||
return &res
|
||||
}
|
||||
|
|
|
@ -77,3 +77,10 @@ func (h *HashNode) UnmarshalJSON(data []byte) error {
|
|||
}
|
||||
return errors.New("expected hash node")
|
||||
}
|
||||
|
||||
// Clone implements Node interface.
|
||||
func (h *HashNode) Clone() Node {
|
||||
res := *h
|
||||
res.Collapsed = false
|
||||
return &res
|
||||
}
|
||||
|
|
|
@ -77,3 +77,9 @@ func (n *LeafNode) UnmarshalJSON(data []byte) error {
|
|||
}
|
||||
return errors.New("expected leaf node")
|
||||
}
|
||||
|
||||
// Clone implements Node interface.
|
||||
func (n *LeafNode) Clone() Node {
|
||||
res := *n
|
||||
return &res
|
||||
}
|
||||
|
|
|
@ -34,6 +34,7 @@ type Node interface {
|
|||
json.Marshaler
|
||||
json.Unmarshaler
|
||||
Size() int
|
||||
Clone() Node
|
||||
BaseNodeIface
|
||||
}
|
||||
|
||||
|
|
|
@ -371,7 +371,9 @@ func (s *Module) restoreNode(n mpt.Node) error {
|
|||
}
|
||||
var childrenPaths = make(map[util.Uint256][][]byte)
|
||||
for _, path := range nPaths {
|
||||
err := s.billet.RestoreHashNode(path, n)
|
||||
// Must clone here in order to avoid future collapse collisions. If the node's refcount>1 then MPT pool
|
||||
// will manage all paths for this node and call RestoreHashNode separately for each of the paths.
|
||||
err := s.billet.RestoreHashNode(path, n.Clone())
|
||||
if err != nil {
|
||||
return fmt.Errorf("failed to restore MPT node with hash %s and path %s: %w", n.Hash().StringBE(), hex.EncodeToString(path), err)
|
||||
}
|
||||
|
|
106
pkg/core/statesync/module_test.go
Normal file
106
pkg/core/statesync/module_test.go
Normal file
|
@ -0,0 +1,106 @@
|
|||
package statesync
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"testing"
|
||||
|
||||
"github.com/nspcc-dev/neo-go/pkg/core/dao"
|
||||
"github.com/nspcc-dev/neo-go/pkg/core/mpt"
|
||||
"github.com/nspcc-dev/neo-go/pkg/core/storage"
|
||||
"github.com/nspcc-dev/neo-go/pkg/util"
|
||||
"github.com/nspcc-dev/neo-go/pkg/util/slice"
|
||||
"github.com/stretchr/testify/require"
|
||||
"go.uber.org/zap/zaptest"
|
||||
)
|
||||
|
||||
func TestModule_PR2019_discussion_r689629704(t *testing.T) {
|
||||
expectedStorage := storage.NewMemCachedStore(storage.NewMemoryStore())
|
||||
tr := mpt.NewTrie(nil, true, expectedStorage)
|
||||
require.NoError(t, tr.Put([]byte{0x03}, []byte("leaf1")))
|
||||
require.NoError(t, tr.Put([]byte{0x01, 0xab, 0x02}, []byte("leaf2")))
|
||||
require.NoError(t, tr.Put([]byte{0x01, 0xab, 0x04}, []byte("leaf3")))
|
||||
require.NoError(t, tr.Put([]byte{0x06, 0x01, 0xde, 0x02}, []byte("leaf2"))) // <-- the same `leaf2` and `leaf3` values are put in the storage,
|
||||
require.NoError(t, tr.Put([]byte{0x06, 0x01, 0xde, 0x04}, []byte("leaf3"))) // <-- but the path should differ.
|
||||
require.NoError(t, tr.Put([]byte{0x06, 0x03}, []byte("leaf4")))
|
||||
|
||||
sr := tr.StateRoot()
|
||||
tr.Flush()
|
||||
|
||||
// Keep MPT nodes in a map in order not to repeat them. We'll use `nodes` map to ask
|
||||
// state sync module to restore the nodes.
|
||||
var (
|
||||
nodes = make(map[util.Uint256][]byte)
|
||||
expectedItems []storage.KeyValue
|
||||
)
|
||||
expectedStorage.Seek(storage.DataMPT.Bytes(), func(k, v []byte) {
|
||||
key := slice.Copy(k)
|
||||
value := slice.Copy(v)
|
||||
expectedItems = append(expectedItems, storage.KeyValue{
|
||||
Key: key,
|
||||
Value: value,
|
||||
})
|
||||
hash, err := util.Uint256DecodeBytesBE(key[1:])
|
||||
require.NoError(t, err)
|
||||
nodeBytes := value[:len(value)-4]
|
||||
nodes[hash] = nodeBytes
|
||||
})
|
||||
|
||||
actualStorage := storage.NewMemCachedStore(storage.NewMemoryStore())
|
||||
// These actions are done in module.Init(), but it's not the point of the test.
|
||||
// Here we want to test only MPT restoring process.
|
||||
stateSync := &Module{
|
||||
log: zaptest.NewLogger(t),
|
||||
syncPoint: 1000500,
|
||||
syncStage: headersSynced,
|
||||
syncInterval: 100500,
|
||||
dao: dao.NewSimple(actualStorage, true, false),
|
||||
mptpool: NewPool(),
|
||||
}
|
||||
stateSync.billet = mpt.NewBillet(sr, true, actualStorage)
|
||||
stateSync.mptpool.Add(sr, []byte{})
|
||||
|
||||
// The test itself: we'll ask state sync module to restore each node exactly once.
|
||||
// After that storage content (including storage items and refcounts) must
|
||||
// match exactly the one got from real MPT trie. MPT pool must be empty.
|
||||
// State sync module must have mptSynced state in the end.
|
||||
// MPT Billet root must become a collapsed hashnode (it was checked manually).
|
||||
requested := make(map[util.Uint256]struct{})
|
||||
for {
|
||||
unknownHashes := stateSync.GetUnknownMPTNodesBatch(1) // restore nodes one-by-one
|
||||
if len(unknownHashes) == 0 {
|
||||
break
|
||||
}
|
||||
h := unknownHashes[0]
|
||||
node, ok := nodes[h]
|
||||
if !ok {
|
||||
if _, ok = requested[h]; ok {
|
||||
t.Fatal("node was requested twice")
|
||||
}
|
||||
t.Fatal("unknown node was requested")
|
||||
}
|
||||
require.NotPanics(t, func() {
|
||||
err := stateSync.AddMPTNodes([][]byte{node})
|
||||
require.NoError(t, err)
|
||||
}, fmt.Errorf("hash=%s, value=%s", h.StringBE(), string(node)))
|
||||
requested[h] = struct{}{}
|
||||
delete(nodes, h)
|
||||
if len(nodes) == 0 {
|
||||
break
|
||||
}
|
||||
}
|
||||
require.Equal(t, headersSynced|mptSynced, stateSync.syncStage, "all nodes were sent exactly ones, but MPT wasn't restored")
|
||||
require.Equal(t, 0, len(nodes), "not all nodes were requested by state sync module")
|
||||
require.Equal(t, 0, stateSync.mptpool.Count(), "MPT was restored, but MPT pool still contains items")
|
||||
|
||||
// Compare resulting storage items and refcounts.
|
||||
var actualItems []storage.KeyValue
|
||||
expectedStorage.Seek(storage.DataMPT.Bytes(), func(k, v []byte) {
|
||||
key := slice.Copy(k)
|
||||
value := slice.Copy(v)
|
||||
actualItems = append(actualItems, storage.KeyValue{
|
||||
Key: key,
|
||||
Value: value,
|
||||
})
|
||||
})
|
||||
require.ElementsMatch(t, expectedItems, actualItems)
|
||||
}
|
Loading…
Reference in a new issue