mirror of
https://github.com/nspcc-dev/neo-go.git
synced 2024-12-24 19:25:21 +00:00
Merge pull request #1871 from nspcc-dev/mpt/serialisation-fix
mpt: refactor serialisation
This commit is contained in:
commit
2c8f8bf3ec
14 changed files with 110 additions and 105 deletions
|
@ -9,6 +9,7 @@ import (
|
||||||
"github.com/nspcc-dev/neo-go/pkg/core/block"
|
"github.com/nspcc-dev/neo-go/pkg/core/block"
|
||||||
"github.com/nspcc-dev/neo-go/pkg/core/interop"
|
"github.com/nspcc-dev/neo-go/pkg/core/interop"
|
||||||
"github.com/nspcc-dev/neo-go/pkg/core/native"
|
"github.com/nspcc-dev/neo-go/pkg/core/native"
|
||||||
|
"github.com/nspcc-dev/neo-go/pkg/core/storage"
|
||||||
"github.com/nspcc-dev/neo-go/pkg/core/transaction"
|
"github.com/nspcc-dev/neo-go/pkg/core/transaction"
|
||||||
"github.com/nspcc-dev/neo-go/pkg/crypto/hash"
|
"github.com/nspcc-dev/neo-go/pkg/crypto/hash"
|
||||||
"github.com/nspcc-dev/neo-go/pkg/crypto/keys"
|
"github.com/nspcc-dev/neo-go/pkg/crypto/keys"
|
||||||
|
@ -16,14 +17,6 @@ import (
|
||||||
"github.com/nspcc-dev/neo-go/pkg/vm/stackitem"
|
"github.com/nspcc-dev/neo-go/pkg/vm/stackitem"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
|
||||||
// MaxStorageKeyLen is the maximum length of a key for storage items.
|
|
||||||
MaxStorageKeyLen = 64
|
|
||||||
// MaxStorageValueLen is the maximum length of a value for storage items.
|
|
||||||
// It is set to be the maximum value for uint16.
|
|
||||||
MaxStorageValueLen = 65535
|
|
||||||
)
|
|
||||||
|
|
||||||
// StorageContext contains storing id and read/write flag, it's used as
|
// StorageContext contains storing id and read/write flag, it's used as
|
||||||
// a context for storage manipulation functions.
|
// a context for storage manipulation functions.
|
||||||
type StorageContext struct {
|
type StorageContext struct {
|
||||||
|
@ -31,17 +24,6 @@ type StorageContext struct {
|
||||||
ReadOnly bool
|
ReadOnly bool
|
||||||
}
|
}
|
||||||
|
|
||||||
// StorageFlag represents storage flag which denotes whether the stored value is
|
|
||||||
// a constant.
|
|
||||||
type StorageFlag byte
|
|
||||||
|
|
||||||
const (
|
|
||||||
// None is a storage flag for non-constant items.
|
|
||||||
None StorageFlag = 0
|
|
||||||
// Constant is a storage flag for constant items.
|
|
||||||
Constant StorageFlag = 0x01
|
|
||||||
)
|
|
||||||
|
|
||||||
// engineGetScriptContainer returns transaction or block that contains the script
|
// engineGetScriptContainer returns transaction or block that contains the script
|
||||||
// being run.
|
// being run.
|
||||||
func engineGetScriptContainer(ic *interop.Context) error {
|
func engineGetScriptContainer(ic *interop.Context) error {
|
||||||
|
@ -115,10 +97,10 @@ func storageGetContextInternal(ic *interop.Context, isReadOnly bool) error {
|
||||||
}
|
}
|
||||||
|
|
||||||
func putWithContext(ic *interop.Context, stc *StorageContext, key []byte, value []byte) error {
|
func putWithContext(ic *interop.Context, stc *StorageContext, key []byte, value []byte) error {
|
||||||
if len(key) > MaxStorageKeyLen {
|
if len(key) > storage.MaxStorageKeyLen {
|
||||||
return errors.New("key is too big")
|
return errors.New("key is too big")
|
||||||
}
|
}
|
||||||
if len(value) > MaxStorageValueLen {
|
if len(value) > storage.MaxStorageValueLen {
|
||||||
return errors.New("value is too big")
|
return errors.New("value is too big")
|
||||||
}
|
}
|
||||||
if stc.ReadOnly {
|
if stc.ReadOnly {
|
||||||
|
|
|
@ -13,6 +13,7 @@ import (
|
||||||
"github.com/nspcc-dev/neo-go/pkg/core/interop/runtime"
|
"github.com/nspcc-dev/neo-go/pkg/core/interop/runtime"
|
||||||
"github.com/nspcc-dev/neo-go/pkg/core/native"
|
"github.com/nspcc-dev/neo-go/pkg/core/native"
|
||||||
"github.com/nspcc-dev/neo-go/pkg/core/state"
|
"github.com/nspcc-dev/neo-go/pkg/core/state"
|
||||||
|
"github.com/nspcc-dev/neo-go/pkg/core/storage"
|
||||||
"github.com/nspcc-dev/neo-go/pkg/core/transaction"
|
"github.com/nspcc-dev/neo-go/pkg/core/transaction"
|
||||||
"github.com/nspcc-dev/neo-go/pkg/crypto/hash"
|
"github.com/nspcc-dev/neo-go/pkg/crypto/hash"
|
||||||
"github.com/nspcc-dev/neo-go/pkg/crypto/keys"
|
"github.com/nspcc-dev/neo-go/pkg/crypto/keys"
|
||||||
|
@ -198,7 +199,7 @@ func TestStoragePut(t *testing.T) {
|
||||||
})
|
})
|
||||||
|
|
||||||
t.Run("check limits", func(t *testing.T) {
|
t.Run("check limits", func(t *testing.T) {
|
||||||
initVM(t, make([]byte, MaxStorageKeyLen), make([]byte, MaxStorageValueLen), -1)
|
initVM(t, make([]byte, storage.MaxStorageKeyLen), make([]byte, storage.MaxStorageValueLen), -1)
|
||||||
require.NoError(t, storagePut(ic))
|
require.NoError(t, storagePut(ic))
|
||||||
})
|
})
|
||||||
|
|
||||||
|
@ -209,11 +210,11 @@ func TestStoragePut(t *testing.T) {
|
||||||
require.Error(t, storagePut(ic))
|
require.Error(t, storagePut(ic))
|
||||||
})
|
})
|
||||||
t.Run("big key", func(t *testing.T) {
|
t.Run("big key", func(t *testing.T) {
|
||||||
initVM(t, make([]byte, MaxStorageKeyLen+1), []byte{1}, -1)
|
initVM(t, make([]byte, storage.MaxStorageKeyLen+1), []byte{1}, -1)
|
||||||
require.Error(t, storagePut(ic))
|
require.Error(t, storagePut(ic))
|
||||||
})
|
})
|
||||||
t.Run("big value", func(t *testing.T) {
|
t.Run("big value", func(t *testing.T) {
|
||||||
initVM(t, []byte{1}, make([]byte, MaxStorageValueLen+1), -1)
|
initVM(t, []byte{1}, make([]byte, storage.MaxStorageValueLen+1), -1)
|
||||||
require.Error(t, storagePut(ic))
|
require.Error(t, storagePut(ic))
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
|
@ -23,6 +23,7 @@ type BaseNodeIface interface {
|
||||||
Hash() util.Uint256
|
Hash() util.Uint256
|
||||||
Type() NodeType
|
Type() NodeType
|
||||||
Bytes() []byte
|
Bytes() []byte
|
||||||
|
EncodeBinaryAsChild(w *io.BinWriter)
|
||||||
}
|
}
|
||||||
|
|
||||||
type flushedNode interface {
|
type flushedNode interface {
|
||||||
|
@ -77,7 +78,17 @@ func (b *BaseNode) invalidateCache() {
|
||||||
|
|
||||||
// encodeNodeWithType encodes node together with it's type.
|
// encodeNodeWithType encodes node together with it's type.
|
||||||
func encodeNodeWithType(n Node, w *io.BinWriter) {
|
func encodeNodeWithType(n Node, w *io.BinWriter) {
|
||||||
w.WriteB(byte(n.Type()))
|
switch t := n.Type(); t {
|
||||||
|
case HashT:
|
||||||
|
hn := n.(*HashNode)
|
||||||
|
if !hn.hashValid {
|
||||||
|
w.WriteB(byte(EmptyT))
|
||||||
|
break
|
||||||
|
}
|
||||||
|
fallthrough
|
||||||
|
default:
|
||||||
|
w.WriteB(byte(t))
|
||||||
|
}
|
||||||
n.EncodeBinary(w)
|
n.EncodeBinary(w)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -93,9 +104,19 @@ func DecodeNodeWithType(r *io.BinReader) Node {
|
||||||
case ExtensionT:
|
case ExtensionT:
|
||||||
n = new(ExtensionNode)
|
n = new(ExtensionNode)
|
||||||
case HashT:
|
case HashT:
|
||||||
n = new(HashNode)
|
n = &HashNode{
|
||||||
|
BaseNode: BaseNode{
|
||||||
|
hashValid: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
case LeafT:
|
case LeafT:
|
||||||
n = new(LeafNode)
|
n = new(LeafNode)
|
||||||
|
case EmptyT:
|
||||||
|
n = &HashNode{
|
||||||
|
BaseNode: BaseNode{
|
||||||
|
hashValid: false,
|
||||||
|
},
|
||||||
|
}
|
||||||
default:
|
default:
|
||||||
r.Err = fmt.Errorf("invalid node type: %x", typ)
|
r.Err = fmt.Errorf("invalid node type: %x", typ)
|
||||||
return nil
|
return nil
|
||||||
|
|
|
@ -48,20 +48,22 @@ func (b *BranchNode) Bytes() []byte {
|
||||||
// EncodeBinary implements io.Serializable.
|
// EncodeBinary implements io.Serializable.
|
||||||
func (b *BranchNode) EncodeBinary(w *io.BinWriter) {
|
func (b *BranchNode) EncodeBinary(w *io.BinWriter) {
|
||||||
for i := 0; i < childrenCount; i++ {
|
for i := 0; i < childrenCount; i++ {
|
||||||
if hn, ok := b.Children[i].(*HashNode); ok {
|
b.Children[i].EncodeBinaryAsChild(w)
|
||||||
hn.EncodeBinary(w)
|
|
||||||
continue
|
|
||||||
}
|
}
|
||||||
n := NewHashNode(b.Children[i].Hash())
|
}
|
||||||
|
|
||||||
|
// EncodeBinaryAsChild implements BaseNode interface.
|
||||||
|
func (b *BranchNode) EncodeBinaryAsChild(w *io.BinWriter) {
|
||||||
|
n := &NodeObject{Node: NewHashNode(b.Hash())} // with type
|
||||||
n.EncodeBinary(w)
|
n.EncodeBinary(w)
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
// DecodeBinary implements io.Serializable.
|
// DecodeBinary implements io.Serializable.
|
||||||
func (b *BranchNode) DecodeBinary(r *io.BinReader) {
|
func (b *BranchNode) DecodeBinary(r *io.BinReader) {
|
||||||
for i := 0; i < childrenCount; i++ {
|
for i := 0; i < childrenCount; i++ {
|
||||||
b.Children[i] = new(HashNode)
|
no := new(NodeObject)
|
||||||
b.Children[i].DecodeBinary(r)
|
no.DecodeBinary(r)
|
||||||
|
b.Children[i] = no.Node
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -6,12 +6,13 @@ import (
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/nspcc-dev/neo-go/pkg/core/storage"
|
||||||
"github.com/nspcc-dev/neo-go/pkg/io"
|
"github.com/nspcc-dev/neo-go/pkg/io"
|
||||||
"github.com/nspcc-dev/neo-go/pkg/util"
|
"github.com/nspcc-dev/neo-go/pkg/util"
|
||||||
)
|
)
|
||||||
|
|
||||||
// MaxKeyLength is the max length of the extension node key.
|
// MaxKeyLength is the max length of the extension node key.
|
||||||
const MaxKeyLength = 1125
|
const MaxKeyLength = (storage.MaxStorageKeyLen + 4) * 2
|
||||||
|
|
||||||
// ExtensionNode represents MPT's extension node.
|
// ExtensionNode represents MPT's extension node.
|
||||||
type ExtensionNode struct {
|
type ExtensionNode struct {
|
||||||
|
@ -53,15 +54,21 @@ func (e *ExtensionNode) DecodeBinary(r *io.BinReader) {
|
||||||
}
|
}
|
||||||
e.key = make([]byte, sz)
|
e.key = make([]byte, sz)
|
||||||
r.ReadBytes(e.key)
|
r.ReadBytes(e.key)
|
||||||
e.next = new(HashNode)
|
no := new(NodeObject)
|
||||||
e.next.DecodeBinary(r)
|
no.DecodeBinary(r)
|
||||||
|
e.next = no.Node
|
||||||
e.invalidateCache()
|
e.invalidateCache()
|
||||||
}
|
}
|
||||||
|
|
||||||
// EncodeBinary implements io.Serializable.
|
// EncodeBinary implements io.Serializable.
|
||||||
func (e ExtensionNode) EncodeBinary(w *io.BinWriter) {
|
func (e ExtensionNode) EncodeBinary(w *io.BinWriter) {
|
||||||
w.WriteVarBytes(e.key)
|
w.WriteVarBytes(e.key)
|
||||||
n := NewHashNode(e.next.Hash())
|
e.next.EncodeBinaryAsChild(w)
|
||||||
|
}
|
||||||
|
|
||||||
|
// EncodeBinaryAsChild implements BaseNode interface.
|
||||||
|
func (e *ExtensionNode) EncodeBinaryAsChild(w *io.BinWriter) {
|
||||||
|
n := &NodeObject{Node: NewHashNode(e.Hash())} // with type
|
||||||
n.EncodeBinary(w)
|
n.EncodeBinary(w)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -2,7 +2,6 @@ package mpt
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
|
||||||
|
|
||||||
"github.com/nspcc-dev/neo-go/pkg/io"
|
"github.com/nspcc-dev/neo-go/pkg/io"
|
||||||
"github.com/nspcc-dev/neo-go/pkg/util"
|
"github.com/nspcc-dev/neo-go/pkg/util"
|
||||||
|
@ -46,25 +45,23 @@ func (h *HashNode) Bytes() []byte {
|
||||||
|
|
||||||
// DecodeBinary implements io.Serializable.
|
// DecodeBinary implements io.Serializable.
|
||||||
func (h *HashNode) DecodeBinary(r *io.BinReader) {
|
func (h *HashNode) DecodeBinary(r *io.BinReader) {
|
||||||
sz := r.ReadVarUint()
|
if h.hashValid {
|
||||||
switch sz {
|
h.hash.DecodeBinary(r)
|
||||||
case 0:
|
|
||||||
h.hashValid = false
|
|
||||||
case util.Uint256Size:
|
|
||||||
h.hashValid = true
|
|
||||||
r.ReadBytes(h.hash[:])
|
|
||||||
default:
|
|
||||||
r.Err = fmt.Errorf("invalid hash node size: %d", sz)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// EncodeBinary implements io.Serializable.
|
// EncodeBinary implements io.Serializable.
|
||||||
func (h HashNode) EncodeBinary(w *io.BinWriter) {
|
func (h HashNode) EncodeBinary(w *io.BinWriter) {
|
||||||
if !h.hashValid {
|
if !h.hashValid {
|
||||||
w.WriteVarUint(0)
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
w.WriteVarBytes(h.hash[:])
|
w.WriteBytes(h.hash[:])
|
||||||
|
}
|
||||||
|
|
||||||
|
// EncodeBinaryAsChild implements BaseNode interface.
|
||||||
|
func (h *HashNode) EncodeBinaryAsChild(w *io.BinWriter) {
|
||||||
|
no := &NodeObject{Node: h} // with type
|
||||||
|
no.EncodeBinary(w)
|
||||||
}
|
}
|
||||||
|
|
||||||
// MarshalJSON implements json.Marshaler.
|
// MarshalJSON implements json.Marshaler.
|
||||||
|
|
|
@ -5,12 +5,13 @@ import (
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/nspcc-dev/neo-go/pkg/core/storage"
|
||||||
"github.com/nspcc-dev/neo-go/pkg/io"
|
"github.com/nspcc-dev/neo-go/pkg/io"
|
||||||
"github.com/nspcc-dev/neo-go/pkg/util"
|
"github.com/nspcc-dev/neo-go/pkg/util"
|
||||||
)
|
)
|
||||||
|
|
||||||
// MaxValueLength is a max length of a leaf node value.
|
// MaxValueLength is a max length of a leaf node value.
|
||||||
const MaxValueLength = 1024 * 1024
|
const MaxValueLength = 3 + storage.MaxStorageValueLen + 1
|
||||||
|
|
||||||
// LeafNode represents MPT's leaf node.
|
// LeafNode represents MPT's leaf node.
|
||||||
type LeafNode struct {
|
type LeafNode struct {
|
||||||
|
@ -55,6 +56,12 @@ func (n LeafNode) EncodeBinary(w *io.BinWriter) {
|
||||||
w.WriteVarBytes(n.value)
|
w.WriteVarBytes(n.value)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// EncodeBinaryAsChild implements BaseNode interface.
|
||||||
|
func (n *LeafNode) EncodeBinaryAsChild(w *io.BinWriter) {
|
||||||
|
no := &NodeObject{Node: NewHashNode(n.Hash())} // with type
|
||||||
|
no.EncodeBinary(w)
|
||||||
|
}
|
||||||
|
|
||||||
// MarshalJSON implements json.Marshaler.
|
// MarshalJSON implements json.Marshaler.
|
||||||
func (n *LeafNode) MarshalJSON() ([]byte, error) {
|
func (n *LeafNode) MarshalJSON() ([]byte, error) {
|
||||||
return []byte(`{"value":"` + hex.EncodeToString(n.value) + `"}`), nil
|
return []byte(`{"value":"` + hex.EncodeToString(n.value) + `"}`), nil
|
||||||
|
|
|
@ -16,8 +16,9 @@ type NodeType byte
|
||||||
const (
|
const (
|
||||||
BranchT NodeType = 0x00
|
BranchT NodeType = 0x00
|
||||||
ExtensionT NodeType = 0x01
|
ExtensionT NodeType = 0x01
|
||||||
HashT NodeType = 0x02
|
LeafT NodeType = 0x02
|
||||||
LeafT NodeType = 0x03
|
HashT NodeType = 0x03
|
||||||
|
EmptyT NodeType = 0x04
|
||||||
)
|
)
|
||||||
|
|
||||||
// NodeObject represents Node together with it's type.
|
// NodeObject represents Node together with it's type.
|
||||||
|
|
|
@ -16,6 +16,9 @@ func getTestFuncEncode(ok bool, expected, actual Node) func(t *testing.T) {
|
||||||
t.Run("IO", func(t *testing.T) {
|
t.Run("IO", func(t *testing.T) {
|
||||||
bs, err := testserdes.EncodeBinary(expected)
|
bs, err := testserdes.EncodeBinary(expected)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
if hn, ok := actual.(*HashNode); ok {
|
||||||
|
hn.hashValid = true // this field is set during NodeObject decoding
|
||||||
|
}
|
||||||
err = testserdes.DecodeBinary(bs, actual)
|
err = testserdes.DecodeBinary(bs, actual)
|
||||||
if !ok {
|
if !ok {
|
||||||
require.Error(t, err)
|
require.Error(t, err)
|
||||||
|
@ -80,8 +83,8 @@ func TestNode_Serializable(t *testing.T) {
|
||||||
})
|
})
|
||||||
t.Run("InvalidSize", func(t *testing.T) {
|
t.Run("InvalidSize", func(t *testing.T) {
|
||||||
buf := io.NewBufBinWriter()
|
buf := io.NewBufBinWriter()
|
||||||
buf.BinWriter.WriteVarBytes(make([]byte, 13))
|
buf.BinWriter.WriteBytes(make([]byte, 13))
|
||||||
require.Error(t, testserdes.DecodeBinary(buf.Bytes(), new(HashNode)))
|
require.Error(t, testserdes.DecodeBinary(buf.Bytes(), &HashNode{BaseNode: BaseNode{hashValid: true}}))
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
|
@ -151,6 +154,6 @@ func TestRootHash(t *testing.T) {
|
||||||
b.Children[9] = l2
|
b.Children[9] = l2
|
||||||
|
|
||||||
r1 := NewExtensionNode([]byte{0x0A, 0x0C, 0x00, 0x01}, v1)
|
r1 := NewExtensionNode([]byte{0x0A, 0x0C, 0x00, 0x01}, v1)
|
||||||
require.Equal(t, "dea3ab46e9461e885ed7091c1e533e0a8030b248d39cbc638962394eaca0fbb3", r1.Hash().StringLE())
|
require.Equal(t, "cedd9897dd1559fbd5dfe5cfb223464da6de438271028afb8d647e950cbd18e0", r1.Hash().StringLE())
|
||||||
require.Equal(t, "93e8e1ffe2f83dd92fca67330e273bcc811bf64b8f8d9d1b25d5e7366b47d60d", r.Hash().StringLE())
|
require.Equal(t, "1037e779c8a0313bd0d99c4151fa70a277c43c53a549b6444079f2e67e8ffb7b", r.Hash().StringLE())
|
||||||
}
|
}
|
||||||
|
|
|
@ -23,6 +23,14 @@ const (
|
||||||
SYSVersion KeyPrefix = 0xf0
|
SYSVersion KeyPrefix = 0xf0
|
||||||
)
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
// MaxStorageKeyLen is the maximum length of a key for storage items.
|
||||||
|
MaxStorageKeyLen = 64
|
||||||
|
// MaxStorageValueLen is the maximum length of a value for storage items.
|
||||||
|
// It is set to be the maximum value for uint16.
|
||||||
|
MaxStorageValueLen = 65535
|
||||||
|
)
|
||||||
|
|
||||||
// ErrKeyNotFound is an error returned by Store implementations
|
// ErrKeyNotFound is an error returned by Store implementations
|
||||||
// when a certain key is not found.
|
// when a certain key is not found.
|
||||||
var ErrKeyNotFound = errors.New("key not found")
|
var ErrKeyNotFound = errors.New("key not found")
|
||||||
|
|
|
@ -2,9 +2,8 @@ package result
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"encoding/hex"
|
"encoding/base64"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"errors"
|
|
||||||
|
|
||||||
"github.com/nspcc-dev/neo-go/pkg/io"
|
"github.com/nspcc-dev/neo-go/pkg/io"
|
||||||
)
|
)
|
||||||
|
@ -21,12 +20,6 @@ type ProofWithKey struct {
|
||||||
Proof [][]byte
|
Proof [][]byte
|
||||||
}
|
}
|
||||||
|
|
||||||
// GetProof is a result of getproof RPC.
|
|
||||||
type GetProof struct {
|
|
||||||
Result ProofWithKey `json:"proof"`
|
|
||||||
Success bool `json:"success"`
|
|
||||||
}
|
|
||||||
|
|
||||||
// VerifyProof is a result of verifyproof RPC.
|
// VerifyProof is a result of verifyproof RPC.
|
||||||
// nil Value is considered invalid.
|
// nil Value is considered invalid.
|
||||||
type VerifyProof struct {
|
type VerifyProof struct {
|
||||||
|
@ -40,7 +33,7 @@ func (p *ProofWithKey) MarshalJSON() ([]byte, error) {
|
||||||
if w.Err != nil {
|
if w.Err != nil {
|
||||||
return nil, w.Err
|
return nil, w.Err
|
||||||
}
|
}
|
||||||
return []byte(`"` + hex.EncodeToString(w.Bytes()) + `"`), nil
|
return []byte(`"` + base64.StdEncoding.EncodeToString(w.Bytes()) + `"`), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// EncodeBinary implements io.Serializable.
|
// EncodeBinary implements io.Serializable.
|
||||||
|
@ -74,12 +67,12 @@ func (p *ProofWithKey) UnmarshalJSON(data []byte) error {
|
||||||
func (p *ProofWithKey) String() string {
|
func (p *ProofWithKey) String() string {
|
||||||
w := io.NewBufBinWriter()
|
w := io.NewBufBinWriter()
|
||||||
p.EncodeBinary(w.BinWriter)
|
p.EncodeBinary(w.BinWriter)
|
||||||
return hex.EncodeToString(w.Bytes())
|
return base64.StdEncoding.EncodeToString(w.Bytes())
|
||||||
}
|
}
|
||||||
|
|
||||||
// FromString decodes p from hex-encoded string.
|
// FromString decodes p from hex-encoded string.
|
||||||
func (p *ProofWithKey) FromString(s string) error {
|
func (p *ProofWithKey) FromString(s string) error {
|
||||||
rawProof, err := hex.DecodeString(s)
|
rawProof, err := base64.StdEncoding.DecodeString(s)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
@ -93,7 +86,7 @@ func (p *VerifyProof) MarshalJSON() ([]byte, error) {
|
||||||
if p.Value == nil {
|
if p.Value == nil {
|
||||||
return []byte(`"invalid"`), nil
|
return []byte(`"invalid"`), nil
|
||||||
}
|
}
|
||||||
return []byte(`{"value":"` + hex.EncodeToString(p.Value) + `"}`), nil
|
return []byte(`"` + base64.StdEncoding.EncodeToString(p.Value) + `"`), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// UnmarshalJSON implements json.Unmarshaler.
|
// UnmarshalJSON implements json.Unmarshaler.
|
||||||
|
@ -102,18 +95,11 @@ func (p *VerifyProof) UnmarshalJSON(data []byte) error {
|
||||||
p.Value = nil
|
p.Value = nil
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
var m map[string]string
|
var m string
|
||||||
if err := json.Unmarshal(data, &m); err != nil {
|
if err := json.Unmarshal(data, &m); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
if len(m) != 1 {
|
b, err := base64.StdEncoding.DecodeString(m)
|
||||||
return errors.New("must have single key")
|
|
||||||
}
|
|
||||||
v, ok := m["value"]
|
|
||||||
if !ok {
|
|
||||||
return errors.New("invalid json")
|
|
||||||
}
|
|
||||||
b, err := hex.DecodeString(v)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
|
@ -24,23 +24,17 @@ func testProofWithKey() *ProofWithKey {
|
||||||
|
|
||||||
func TestGetProof_MarshalJSON(t *testing.T) {
|
func TestGetProof_MarshalJSON(t *testing.T) {
|
||||||
t.Run("Good", func(t *testing.T) {
|
t.Run("Good", func(t *testing.T) {
|
||||||
p := &GetProof{
|
p := testProofWithKey()
|
||||||
Result: *testProofWithKey(),
|
testserdes.MarshalUnmarshalJSON(t, p, new(ProofWithKey))
|
||||||
Success: true,
|
|
||||||
}
|
|
||||||
testserdes.MarshalUnmarshalJSON(t, p, new(GetProof))
|
|
||||||
})
|
})
|
||||||
t.Run("Compatibility", func(t *testing.T) {
|
t.Run("Compatibility", func(t *testing.T) {
|
||||||
js := []byte(`{
|
js := []byte(`"Bfn///8SBiQBAQ8D6yfHa4wV24kQ9eXarzY5Bw55VFzysUbkJjrz5FipqkjSAAQEBAQEBAMcbFvhto6QJgYoJs/uzqTrZNrPxpkgNiF5Z/ME98copwPQ4q6ZqLA8S7XUXNCrJNF68vMu8Gx3W8Ooo3qwMomm0gQDiT6zHh/siCZ0c2bfBEymPmRNTiXSAKFIammjmnnBnJYD+CNwgcEzBJqYfnc7RMhr8cPhffKN0281w0M7XLQ9BO4D7W+t3cleDNdiNc6tqWR8jyIP+bolh5QnZIyKXPwGHjsEBAQDcpxkuWYJr6g3ilENTh1sztlZsXZvt6Eedmyy6kI2gQoEKQEGDw8PDw8PA33qzf1Q5ILAwmYxBnM2N80A8JtFHKR7UHhVEqo5nQ0eUgADbChDXdc7hSDZpD9xbhYGuJxVxRWqhsVRTR2dE+18gd4DG5gRFexXofB0aNb6G2kzQUSTD+aWVsfmnKGf4HHivzAEBAQEBAQEBAQEBAQEBARSAAQEA2IMPmRKP0b2BqhMB6IgtfpPeuXKJMdMze7Cr1TeJqbmA1vvqQgR5DN9ew+Zp/nc5SBQbjV5gEq7F/tIipWaQJ1hBAQEBAQEBAQEBAQEBAMCAR4="`)
|
||||||
"proof" : "25ddeb9aa1bfc353c9c54e21dffb470f65d9c22a0662616c616e63654f70000000000000000708fd12020020666eaa8a6e75d43a97d76e72b605c7e05189f0c57ec19d84acdb75810f18239d202c83028ce3d7abcf4e4f95d05fbfdfa5e18bde3a8fbb65a57559d6b5ea09425c2090c40d440744a848e3b407a00e4efb692a957245a1efc9cb8496cb05fd328ee620dd2652bf25dfc3ad5fee7b200ccf3e3ae50772ff8ed58907e4dab8e7d4b2489720d8a5d5ed75b5b0f256d0a2cf5c220b4ddae2a228ef0fc0212b689f3811dfa94620342cc0d73fabd2440ed2cc735a9608391a510e1981b321a9f4258682706adc9620ced036e52f39387b9c58ade7bf8c3ca8959b64d8031d36d9b1c62f3f1c51c7cb2031072c7c801b5c1614dae441383a65344acd238f13db28ff0a39c0626e597f002062552d64c616d8b2a6a93d22936055110c0065728aa2b4fbf4d76b108390b474203322d3c93c741674a307cf6455e77c02ceeda307d4ec23fd809a2a420b4243f82052ab92a9cedc6716ad4c66a8a3e423b195b05bdebde456f992bff48f2561e99720e6379995e7053823b8ba8fb8af9623cf48e89f60c989598445df5e711db42a6f20192894ed637e86561ff6a4b8dea4539dee8bddb2fb20bf4ae3499852985c88b120e0005edd09f2335aa6b59ff4723e1262b2192adaa5e3e56f79e662f07041f04c2033577f3e2c5bb0e58746980a07cdfad2f872e2b9a10bcc27b7c678c85576df8420f0f04180d15b6eaa0c43e62380084c75ad773d790700a7120c6c4da1fc51693000fd720100209648e8f10a5ff4c209009b9a09697babbe1b2150d0948c1970a560282a1bfa4720988af8f34859dd8309bffea0b1dff9c8cef0b9b0d6a1852d40786627729ae7be00206ebf4f1b7861bca041cbb8feca75158511ca43a1810d17e1e3017468e8cef0de20cac93064090a7da09f8202c17d1e6cbb9a16eb43afcb032e80719cbf05b3446d2019b76a10b91fb99ec08814e8108e5490b879fb09a190cb2c129dfd98335bd5de000020b1da1198bacacf2adc0d863929d77c285ce3a26e736203d0c0a69a1312255fb2207ee8aa092f49348bd89f9c4bf004b0bee2241a2d0acfe7b3ce08e414b04a5717205b0dda71eac8a4e4cdc6a7b939748c0a78abb54f2547a780e6df67b25530330f000020fc358fb9d1e0d36461e015ac8e35f97072a9f9e750a3c25722a2b1a858fcb82d203c52c9fac6d4694b351390158334a9166bc3478ceb9bea2b0b244915f918239e20d526344a24ff19ee6a9f5c5beb833f4eb6d51191590350e26fa50b138493473f005200000000000000000000002077c404fec0a4265568951dbd096572787d109fab105213f4f292a5f53ce72fca00000020b8d1c7a386eaba83ce83ee0700d4ca9b86e75d147d670ea05123e438231d895000004801250b090a0a010b0f0c0305030c090c05040e02010d0f0f0b0407000f06050d090c02020a0006202af2097cf9d3f42e49f6b3c3dd254e7cbdab3485b029721cbbbf1ad0455a810852000000000000002055170506f4b18bc573a909b51cb21bdd5d303ec511f6cdfb1c6a1ab8d8a1dad020ee774c1b9fe1d8ea8d05823837d959da48af74f384d52f06c42c9d146c5258e300000000000000000072000000204457a6fe530ee953ad1f9caf63daf7f86719c9986df2d0b6917021eb379800f00020406bfc79da4ba6f37452a679d13cca252585d34f7e94a480b047bad9427f233e00000000201ce15a2373d28e0dc5f2000cf308f155d06f72070a29e5af1528c8f05f29d248000000000000004301200601060c0601060e06030605040f0700000000000000000000000000000000072091b83866bbd7450115b462e8d48601af3c3e9a35e7018d2b98a23e107c15c200090307000410a328e800",
|
|
||||||
"success" : true
|
|
||||||
}`)
|
|
||||||
|
|
||||||
var p GetProof
|
var p ProofWithKey
|
||||||
require.NoError(t, json.Unmarshal(js, &p))
|
require.NoError(t, json.Unmarshal(js, &p))
|
||||||
require.Equal(t, 8, len(p.Result.Proof))
|
require.Equal(t, 6, len(p.Proof))
|
||||||
for i := range p.Result.Proof { // smoke test that every chunk is correctly encoded node
|
for i := range p.Proof { // smoke test that every chunk is correctly encoded node
|
||||||
r := io.NewBinReaderFromBuf(p.Result.Proof[i])
|
r := io.NewBinReaderFromBuf(p.Proof[i])
|
||||||
var n mpt.NodeObject
|
var n mpt.NodeObject
|
||||||
n.DecodeBinary(r)
|
n.DecodeBinary(r)
|
||||||
require.NoError(t, r.Err)
|
require.NoError(t, r.Err)
|
||||||
|
|
|
@ -905,7 +905,7 @@ func (s *Server) getProof(ps request.Params) (interface{}, *response.Error) {
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, response.ErrInvalidParams
|
return nil, response.ErrInvalidParams
|
||||||
}
|
}
|
||||||
key, err := ps.Value(2).GetBytesHex()
|
key, err := ps.Value(2).GetBytesBase64()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, response.ErrInvalidParams
|
return nil, response.ErrInvalidParams
|
||||||
}
|
}
|
||||||
|
@ -915,12 +915,9 @@ func (s *Server) getProof(ps request.Params) (interface{}, *response.Error) {
|
||||||
}
|
}
|
||||||
skey := makeStorageKey(cs.ID, key)
|
skey := makeStorageKey(cs.ID, key)
|
||||||
proof, err := s.chain.GetStateModule().GetStateProof(root, skey)
|
proof, err := s.chain.GetStateModule().GetStateProof(root, skey)
|
||||||
return &result.GetProof{
|
return &result.ProofWithKey{
|
||||||
Result: result.ProofWithKey{
|
|
||||||
Key: skey,
|
Key: skey,
|
||||||
Proof: proof,
|
Proof: proof,
|
||||||
},
|
|
||||||
Success: err == nil,
|
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1335,20 +1335,19 @@ func testRPCProtocol(t *testing.T, doRPCCall func(string, string, *testing.T) []
|
||||||
r, err := chain.GetStateModule().GetStateRoot(3)
|
r, err := chain.GetStateModule().GetStateRoot(3)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
rpc := fmt.Sprintf(`{"jsonrpc": "2.0", "id": 1, "method": "getproof", "params": ["%s", "%s", "%x"]}`,
|
rpc := fmt.Sprintf(`{"jsonrpc": "2.0", "id": 1, "method": "getproof", "params": ["%s", "%s", "%s"]}`,
|
||||||
r.Root.StringLE(), testContractHash, []byte("testkey"))
|
r.Root.StringLE(), testContractHash, base64.StdEncoding.EncodeToString([]byte("testkey")))
|
||||||
body := doRPCCall(rpc, httpSrv.URL, t)
|
body := doRPCCall(rpc, httpSrv.URL, t)
|
||||||
rawRes := checkErrGetResult(t, body, false)
|
rawRes := checkErrGetResult(t, body, false)
|
||||||
res := new(result.GetProof)
|
res := new(result.ProofWithKey)
|
||||||
require.NoError(t, json.Unmarshal(rawRes, res))
|
require.NoError(t, json.Unmarshal(rawRes, res))
|
||||||
require.True(t, res.Success)
|
|
||||||
h, _ := util.Uint160DecodeStringLE(testContractHash)
|
h, _ := util.Uint160DecodeStringLE(testContractHash)
|
||||||
skey := makeStorageKey(chain.GetContractState(h).ID, []byte("testkey"))
|
skey := makeStorageKey(chain.GetContractState(h).ID, []byte("testkey"))
|
||||||
require.Equal(t, skey, res.Result.Key)
|
require.Equal(t, skey, res.Key)
|
||||||
require.True(t, len(res.Result.Proof) > 0)
|
require.True(t, len(res.Proof) > 0)
|
||||||
|
|
||||||
rpc = fmt.Sprintf(`{"jsonrpc": "2.0", "id": 1, "method": "verifyproof", "params": ["%s", "%s"]}`,
|
rpc = fmt.Sprintf(`{"jsonrpc": "2.0", "id": 1, "method": "verifyproof", "params": ["%s", "%s"]}`,
|
||||||
r.Root.StringLE(), res.Result.String())
|
r.Root.StringLE(), res.String())
|
||||||
body = doRPCCall(rpc, httpSrv.URL, t)
|
body = doRPCCall(rpc, httpSrv.URL, t)
|
||||||
rawRes = checkErrGetResult(t, body, false)
|
rawRes = checkErrGetResult(t, body, false)
|
||||||
vp := new(result.VerifyProof)
|
vp := new(result.VerifyProof)
|
||||||
|
|
Loading…
Reference in a new issue