package block import ( "encoding/json" "errors" "github.com/Workiva/go-datastructures/queue" "github.com/nspcc-dev/neo-go/pkg/config/netmode" "github.com/nspcc-dev/neo-go/pkg/core/transaction" "github.com/nspcc-dev/neo-go/pkg/crypto/hash" "github.com/nspcc-dev/neo-go/pkg/io" "github.com/nspcc-dev/neo-go/pkg/util" ) // Block represents one block in the chain. type Block struct { // The base of the block. Base // Primary index and nonce ConsensusData ConsensusData `json:"consensusdata"` // Transaction list. Transactions []*transaction.Transaction // True if this block is created from trimmed data. Trimmed bool } // auxBlockOut is used for JSON i/o. type auxBlockOut struct { ConsensusData ConsensusData `json:"consensusdata"` Transactions []*transaction.Transaction `json:"tx"` } // auxBlockIn is used for JSON i/o. type auxBlockIn struct { ConsensusData ConsensusData `json:"consensusdata"` Transactions []json.RawMessage `json:"tx"` } // Header returns the Header of the Block. func (b *Block) Header() *Header { return &Header{ Base: b.Base, } } // computeMerkleTree computes Merkle tree based on actual block's data. func (b *Block) computeMerkleTree() util.Uint256 { hashes := make([]util.Uint256, len(b.Transactions)+1) hashes[0] = b.ConsensusData.Hash() for i, tx := range b.Transactions { hashes[i+1] = tx.Hash() } return hash.CalcMerkleRoot(hashes) } // RebuildMerkleRoot rebuilds the merkleroot of the block. func (b *Block) RebuildMerkleRoot() { b.MerkleRoot = b.computeMerkleTree() } // Verify verifies the integrity of the block. func (b *Block) Verify() error { if b.Transactions != nil { hashes := map[util.Uint256]bool{} for _, tx := range b.Transactions { if !hashes[tx.Hash()] { hashes[tx.Hash()] = true } else { return errors.New("transaction duplication is not allowed") } } } merkle := b.computeMerkleTree() if !b.MerkleRoot.Equals(merkle) { return errors.New("MerkleRoot mismatch") } return nil } // NewBlockFromTrimmedBytes returns a new block from trimmed data. // This is commonly used to create a block from stored data. // Blocks created from trimmed data will have their Trimmed field // set to true. func NewBlockFromTrimmedBytes(network netmode.Magic, b []byte) (*Block, error) { block := &Block{ Base: Base{ Network: network, }, Trimmed: true, } br := io.NewBinReaderFromBuf(b) block.decodeHashableFields(br) _ = br.ReadB() block.Script.DecodeBinary(br) lenHashes := br.ReadVarUint() if lenHashes > 0 { var consensusDataHash util.Uint256 consensusDataHash.DecodeBinary(br) lenTX := lenHashes - 1 block.Transactions = make([]*transaction.Transaction, lenTX) for i := 0; i < int(lenTX); i++ { var hash util.Uint256 hash.DecodeBinary(br) block.Transactions[i] = transaction.NewTrimmedTX(hash) } block.ConsensusData.DecodeBinary(br) } return block, br.Err } // New creates a new blank block tied to the specific network. func New(network netmode.Magic) *Block { return &Block{ Base: Base{ Network: network, }, } } // Trim returns a subset of the block data to save up space // in storage. // Notice that only the hashes of the transactions are stored. func (b *Block) Trim() ([]byte, error) { buf := io.NewBufBinWriter() b.encodeHashableFields(buf.BinWriter) buf.WriteB(1) b.Script.EncodeBinary(buf.BinWriter) buf.WriteVarUint(uint64(len(b.Transactions)) + 1) hash := b.ConsensusData.Hash() hash.EncodeBinary(buf.BinWriter) for _, tx := range b.Transactions { h := tx.Hash() h.EncodeBinary(buf.BinWriter) } b.ConsensusData.EncodeBinary(buf.BinWriter) if buf.Err != nil { return nil, buf.Err } return buf.Bytes(), nil } // DecodeBinary decodes the block from the given BinReader, implementing // Serializable interface. func (b *Block) DecodeBinary(br *io.BinReader) { b.Base.DecodeBinary(br) contentsCount := br.ReadVarUint() if contentsCount == 0 { br.Err = errors.New("invalid block format") return } b.ConsensusData.DecodeBinary(br) txes := make([]*transaction.Transaction, contentsCount-1) for i := 0; i < int(contentsCount)-1; i++ { tx := &transaction.Transaction{Network: b.Network} tx.DecodeBinary(br) txes[i] = tx } b.Transactions = txes if br.Err != nil { return } br.Err = b.Verify() } // EncodeBinary encodes the block to the given BinWriter, implementing // Serializable interface. func (b *Block) EncodeBinary(bw *io.BinWriter) { b.Base.EncodeBinary(bw) bw.WriteVarUint(uint64(len(b.Transactions) + 1)) b.ConsensusData.EncodeBinary(bw) for i := 0; i < len(b.Transactions); i++ { b.Transactions[i].EncodeBinary(bw) } } // Compare implements the queue Item interface. func (b *Block) Compare(item queue.Item) int { other := item.(*Block) switch { case b.Index > other.Index: return 1 case b.Index == other.Index: return 0 default: return -1 } } // MarshalJSON implements json.Marshaler interface. func (b Block) MarshalJSON() ([]byte, error) { auxb, err := json.Marshal(auxBlockOut{ ConsensusData: b.ConsensusData, Transactions: b.Transactions, }) if err != nil { return nil, err } baseBytes, err := json.Marshal(b.Base) if err != nil { return nil, err } // Stitch them together. if baseBytes[len(baseBytes)-1] != '}' || auxb[0] != '{' { return nil, errors.New("can't merge internal jsons") } baseBytes[len(baseBytes)-1] = ',' baseBytes = append(baseBytes, auxb[1:]...) return baseBytes, nil } // UnmarshalJSON implements json.Unmarshaler interface. func (b *Block) UnmarshalJSON(data []byte) error { // As Base and auxb are at the same level in json, // do unmarshalling separately for both structs. auxb := new(auxBlockIn) err := json.Unmarshal(data, auxb) if err != nil { return err } err = json.Unmarshal(data, &b.Base) if err != nil { return err } if len(auxb.Transactions) != 0 { b.Transactions = make([]*transaction.Transaction, 0, len(auxb.Transactions)) for _, txBytes := range auxb.Transactions { tx := &transaction.Transaction{Network: b.Network} err = tx.UnmarshalJSON(txBytes) if err != nil { return err } b.Transactions = append(b.Transactions, tx) } } b.ConsensusData = auxb.ConsensusData // Some tests rely on hash presence and we're usually precomputing // other hashes upon deserialization. _ = b.ConsensusData.Hash() return nil }