mirror of
https://github.com/nspcc-dev/neo-go.git
synced 2024-11-23 23:30:36 +00:00
ce09c82b25
It's used in two places now: * Blockchain.AddBlock() This one does transaction duplication check of its own, doing it in Verify() is just a waste of time. Merkle tree root hash value check is still relevant though * Block.DecodeBinary() We're decoding blocks for the following purposes: - on restore from dump The block will be added to the chain via AddBlock() and that will do a full check of it (if configured to do so) - on retrieving the block from the DB (DAO) We trust the DB, if it's gone wild, this check won't really help - on receiving the block via P2P It's gonna be put into block queue and then end up in AddBlock() which will check it - on receiving the block via RPC (submitblock) It is to be passed into AddBlock() - on receiving the block via RPC in a client That's the only problematic case probably, but RPC client has to trust the server and it can check for the signature if it really cares. Or a separate in-client check might be added. As we can see nothing really requires this verification to be done the way it is now, AddBlock can just have a Merkle check and DecodeBinary can do fine without it at all.
234 lines
5.8 KiB
Go
234 lines
5.8 KiB
Go
package block
|
|
|
|
import (
|
|
"encoding/json"
|
|
"errors"
|
|
|
|
"github.com/Workiva/go-datastructures/queue"
|
|
"github.com/nspcc-dev/neo-go/pkg/config/netmode"
|
|
"github.com/nspcc-dev/neo-go/pkg/core/transaction"
|
|
"github.com/nspcc-dev/neo-go/pkg/crypto/hash"
|
|
"github.com/nspcc-dev/neo-go/pkg/io"
|
|
"github.com/nspcc-dev/neo-go/pkg/util"
|
|
)
|
|
|
|
// Block represents one block in the chain.
|
|
type Block struct {
|
|
// The base of the block.
|
|
Base
|
|
|
|
// Primary index and nonce
|
|
ConsensusData ConsensusData `json:"consensusdata"`
|
|
|
|
// Transaction list.
|
|
Transactions []*transaction.Transaction
|
|
|
|
// True if this block is created from trimmed data.
|
|
Trimmed bool
|
|
}
|
|
|
|
// auxBlockOut is used for JSON i/o.
|
|
type auxBlockOut struct {
|
|
ConsensusData ConsensusData `json:"consensusdata"`
|
|
Transactions []*transaction.Transaction `json:"tx"`
|
|
}
|
|
|
|
// auxBlockIn is used for JSON i/o.
|
|
type auxBlockIn struct {
|
|
ConsensusData ConsensusData `json:"consensusdata"`
|
|
Transactions []json.RawMessage `json:"tx"`
|
|
}
|
|
|
|
// Header returns the Header of the Block.
|
|
func (b *Block) Header() *Header {
|
|
return &Header{
|
|
Base: b.Base,
|
|
}
|
|
}
|
|
|
|
// ComputeMerkleRoot computes Merkle tree root hash based on actual block's data.
|
|
func (b *Block) ComputeMerkleRoot() util.Uint256 {
|
|
hashes := make([]util.Uint256, len(b.Transactions)+1)
|
|
hashes[0] = b.ConsensusData.Hash()
|
|
for i, tx := range b.Transactions {
|
|
hashes[i+1] = tx.Hash()
|
|
}
|
|
|
|
return hash.CalcMerkleRoot(hashes)
|
|
}
|
|
|
|
// RebuildMerkleRoot rebuilds the merkleroot of the block.
|
|
func (b *Block) RebuildMerkleRoot() {
|
|
b.MerkleRoot = b.ComputeMerkleRoot()
|
|
}
|
|
|
|
// NewBlockFromTrimmedBytes returns a new block from trimmed data.
|
|
// This is commonly used to create a block from stored data.
|
|
// Blocks created from trimmed data will have their Trimmed field
|
|
// set to true.
|
|
func NewBlockFromTrimmedBytes(network netmode.Magic, b []byte) (*Block, error) {
|
|
block := &Block{
|
|
Base: Base{
|
|
Network: network,
|
|
},
|
|
Trimmed: true,
|
|
}
|
|
|
|
br := io.NewBinReaderFromBuf(b)
|
|
block.decodeHashableFields(br)
|
|
|
|
_ = br.ReadB()
|
|
|
|
block.Script.DecodeBinary(br)
|
|
|
|
lenHashes := br.ReadVarUint()
|
|
if lenHashes > 0 {
|
|
var consensusDataHash util.Uint256
|
|
consensusDataHash.DecodeBinary(br)
|
|
lenTX := lenHashes - 1
|
|
block.Transactions = make([]*transaction.Transaction, lenTX)
|
|
for i := 0; i < int(lenTX); i++ {
|
|
var hash util.Uint256
|
|
hash.DecodeBinary(br)
|
|
block.Transactions[i] = transaction.NewTrimmedTX(hash)
|
|
}
|
|
block.ConsensusData.DecodeBinary(br)
|
|
}
|
|
|
|
return block, br.Err
|
|
}
|
|
|
|
// New creates a new blank block tied to the specific network.
|
|
func New(network netmode.Magic) *Block {
|
|
return &Block{
|
|
Base: Base{
|
|
Network: network,
|
|
},
|
|
}
|
|
}
|
|
|
|
// Trim returns a subset of the block data to save up space
|
|
// in storage.
|
|
// Notice that only the hashes of the transactions are stored.
|
|
func (b *Block) Trim() ([]byte, error) {
|
|
buf := io.NewBufBinWriter()
|
|
b.encodeHashableFields(buf.BinWriter)
|
|
buf.WriteB(1)
|
|
b.Script.EncodeBinary(buf.BinWriter)
|
|
|
|
buf.WriteVarUint(uint64(len(b.Transactions)) + 1)
|
|
hash := b.ConsensusData.Hash()
|
|
hash.EncodeBinary(buf.BinWriter)
|
|
|
|
for _, tx := range b.Transactions {
|
|
h := tx.Hash()
|
|
h.EncodeBinary(buf.BinWriter)
|
|
}
|
|
|
|
b.ConsensusData.EncodeBinary(buf.BinWriter)
|
|
if buf.Err != nil {
|
|
return nil, buf.Err
|
|
}
|
|
|
|
return buf.Bytes(), nil
|
|
}
|
|
|
|
// DecodeBinary decodes the block from the given BinReader, implementing
|
|
// Serializable interface.
|
|
func (b *Block) DecodeBinary(br *io.BinReader) {
|
|
b.Base.DecodeBinary(br)
|
|
contentsCount := br.ReadVarUint()
|
|
if contentsCount == 0 {
|
|
br.Err = errors.New("invalid block format")
|
|
return
|
|
}
|
|
b.ConsensusData.DecodeBinary(br)
|
|
txes := make([]*transaction.Transaction, contentsCount-1)
|
|
for i := 0; i < int(contentsCount)-1; i++ {
|
|
tx := &transaction.Transaction{Network: b.Network}
|
|
tx.DecodeBinary(br)
|
|
txes[i] = tx
|
|
}
|
|
b.Transactions = txes
|
|
if br.Err != nil {
|
|
return
|
|
}
|
|
}
|
|
|
|
// EncodeBinary encodes the block to the given BinWriter, implementing
|
|
// Serializable interface.
|
|
func (b *Block) EncodeBinary(bw *io.BinWriter) {
|
|
b.Base.EncodeBinary(bw)
|
|
bw.WriteVarUint(uint64(len(b.Transactions) + 1))
|
|
b.ConsensusData.EncodeBinary(bw)
|
|
for i := 0; i < len(b.Transactions); i++ {
|
|
b.Transactions[i].EncodeBinary(bw)
|
|
}
|
|
}
|
|
|
|
// Compare implements the queue Item interface.
|
|
func (b *Block) Compare(item queue.Item) int {
|
|
other := item.(*Block)
|
|
switch {
|
|
case b.Index > other.Index:
|
|
return 1
|
|
case b.Index == other.Index:
|
|
return 0
|
|
default:
|
|
return -1
|
|
}
|
|
}
|
|
|
|
// MarshalJSON implements json.Marshaler interface.
|
|
func (b Block) MarshalJSON() ([]byte, error) {
|
|
auxb, err := json.Marshal(auxBlockOut{
|
|
ConsensusData: b.ConsensusData,
|
|
Transactions: b.Transactions,
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
baseBytes, err := json.Marshal(b.Base)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// Stitch them together.
|
|
if baseBytes[len(baseBytes)-1] != '}' || auxb[0] != '{' {
|
|
return nil, errors.New("can't merge internal jsons")
|
|
}
|
|
baseBytes[len(baseBytes)-1] = ','
|
|
baseBytes = append(baseBytes, auxb[1:]...)
|
|
return baseBytes, nil
|
|
}
|
|
|
|
// UnmarshalJSON implements json.Unmarshaler interface.
|
|
func (b *Block) UnmarshalJSON(data []byte) error {
|
|
// As Base and auxb are at the same level in json,
|
|
// do unmarshalling separately for both structs.
|
|
auxb := new(auxBlockIn)
|
|
err := json.Unmarshal(data, auxb)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
err = json.Unmarshal(data, &b.Base)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if len(auxb.Transactions) != 0 {
|
|
b.Transactions = make([]*transaction.Transaction, 0, len(auxb.Transactions))
|
|
for _, txBytes := range auxb.Transactions {
|
|
tx := &transaction.Transaction{Network: b.Network}
|
|
err = tx.UnmarshalJSON(txBytes)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
b.Transactions = append(b.Transactions, tx)
|
|
}
|
|
}
|
|
b.ConsensusData = auxb.ConsensusData
|
|
// Some tests rely on hash presence and we're usually precomputing
|
|
// other hashes upon deserialization.
|
|
_ = b.ConsensusData.Hash()
|
|
return nil
|
|
}
|