diff --git a/pkg/core/blockchain.go b/pkg/core/blockchain.go index 2a87a729e..27f9bd734 100644 --- a/pkg/core/blockchain.go +++ b/pkg/core/blockchain.go @@ -45,8 +45,7 @@ import ( // Tuning parameters. const ( - headerBatchCount = 2000 - version = "0.2.6" + version = "0.2.6" defaultInitialGAS = 52000000_00000000 defaultGCPeriod = 10000 @@ -115,6 +114,8 @@ var ( // the state of the ledger that can be accessed in various ways and changed by // adding new blocks or headers. type Blockchain struct { + HeaderHashes + config config.ProtocolConfiguration // The only way chain state changes is by adding blocks, so we can't @@ -151,13 +152,6 @@ type Blockchain struct { // Current persisted block count. persistedHeight uint32 - // Number of headers stored in the chain file. - storedHeaderCount uint32 - - // Header hashes list with associated lock. - headerHashesLock sync.RWMutex - headerHashes []util.Uint256 - // Stop synchronization mechanisms. stopCh chan struct{} runToExitCh chan struct{} @@ -380,8 +374,7 @@ func (bc *Blockchain) init() error { if err != nil { return err } - bc.headerHashes = []util.Uint256{genesisBlock.Hash()} - bc.dao.PutCurrentHeader(genesisBlock.Hash(), genesisBlock.Index) + bc.HeaderHashes.initGenesis(bc.dao, genesisBlock.Hash()) if err := bc.stateRoot.Init(0); err != nil { return fmt.Errorf("can't init MPT: %w", err) } @@ -414,41 +407,11 @@ func (bc *Blockchain) init() error { // and the genesis block as first block. bc.log.Info("restoring blockchain", zap.String("version", version)) - bc.headerHashes, err = bc.dao.GetHeaderHashes() + err = bc.HeaderHashes.init(bc.dao) if err != nil { return err } - bc.storedHeaderCount = uint32(len(bc.headerHashes)) - - currHeaderHeight, currHeaderHash, err := bc.dao.GetCurrentHeaderHeight() - if err != nil { - return fmt.Errorf("failed to retrieve current header info: %w", err) - } - - // There is a high chance that the Node is stopped before the next - // batch of 2000 headers was stored. Via the currentHeaders stored we can sync - // that with stored blocks. - if currHeaderHeight >= bc.storedHeaderCount { - hash := currHeaderHash - var targetHash util.Uint256 - if len(bc.headerHashes) > 0 { - targetHash = bc.headerHashes[len(bc.headerHashes)-1] - } - headers := make([]util.Uint256, 0, headerBatchCount) - - for hash != targetHash { - header, err := bc.GetHeader(hash) - if err != nil { - return fmt.Errorf("could not get header %s: %w", hash, err) - } - headers = append(headers, header.Hash()) - hash = header.PrevHash - } - hashSliceReverse(headers) - bc.headerHashes = append(bc.headerHashes, headers...) - } - // Check whether StateChangeState stage is in the storage and continue interrupted state jump / state reset if so. stateChStage, err := bc.dao.Store.Get([]byte{byte(storage.SYSStateChangeStage)}) if err == nil { @@ -539,8 +502,8 @@ func (bc *Blockchain) jumpToState(p uint32) error { // jump stage. All the data needed for the jump must be in the DB, otherwise an // error is returned. It is not protected by mutex. func (bc *Blockchain) jumpToStateInternal(p uint32, stage stateChangeStage) error { - if p+1 >= uint32(len(bc.headerHashes)) { - return fmt.Errorf("invalid state sync point %d: headerHeignt is %d", p, len(bc.headerHashes)) + if p >= bc.HeaderHeight() { + return fmt.Errorf("invalid state sync point %d: headerHeignt is %d", p, bc.HeaderHeight()) } bc.log.Info("jumping to state sync point", zap.Uint32("state sync point", p)) @@ -575,7 +538,7 @@ func (bc *Blockchain) jumpToStateInternal(p uint32, stage stateChangeStage) erro // After current state is updated, we need to remove outdated state-related data if so. // The only outdated data we might have is genesis-related data, so check it. if p-bc.config.MaxTraceableBlocks > 0 { - err := cache.DeleteBlock(bc.headerHashes[0]) + err := cache.DeleteBlock(bc.GetHeaderHash(0)) if err != nil { return fmt.Errorf("failed to remove outdated state data for the genesis block: %w", err) } @@ -588,7 +551,7 @@ func (bc *Blockchain) jumpToStateInternal(p uint32, stage stateChangeStage) erro } } // Update SYS-prefixed info. - block, err := bc.dao.GetBlock(bc.headerHashes[p]) + block, err := bc.dao.GetBlock(bc.GetHeaderHash(p)) if err != nil { return fmt.Errorf("failed to get current block: %w", err) } @@ -604,7 +567,7 @@ func (bc *Blockchain) jumpToStateInternal(p uint32, stage stateChangeStage) erro default: return fmt.Errorf("unknown state jump stage: %d", stage) } - block, err := bc.dao.GetBlock(bc.headerHashes[p+1]) + block, err := bc.dao.GetBlock(bc.GetHeaderHash(p + 1)) if err != nil { return fmt.Errorf("failed to get block to init MPT: %w", err) } @@ -625,10 +588,12 @@ func (bc *Blockchain) jumpToStateInternal(p uint32, stage stateChangeStage) erro // resetRAMState resets in-memory cached info. func (bc *Blockchain) resetRAMState(height uint32, resetHeaders bool) error { if resetHeaders { - bc.headerHashes = bc.headerHashes[:height+1] - bc.storedHeaderCount = height + 1 + err := bc.HeaderHashes.init(bc.dao) + if err != nil { + return err + } } - block, err := bc.dao.GetBlock(bc.headerHashes[height]) + block, err := bc.dao.GetBlock(bc.GetHeaderHash(height)) if err != nil { return fmt.Errorf("failed to get current block: %w", err) } @@ -685,7 +650,7 @@ func (bc *Blockchain) resetStateInternal(height uint32, stage stateChangeStage) } // Retrieve necessary state before the DB modification. - b, err := bc.GetBlock(bc.headerHashes[height]) + b, err := bc.GetBlock(bc.GetHeaderHash(height)) if err != nil { return fmt.Errorf("failed to retrieve block %d: %w", height, err) } @@ -1406,7 +1371,6 @@ func (bc *Blockchain) AddHeaders(headers ...*block.Header) error { func (bc *Blockchain) addHeaders(verify bool, headers ...*block.Header) error { var ( start = time.Now() - batch = bc.dao.GetPrivate() err error ) @@ -1436,44 +1400,14 @@ func (bc *Blockchain) addHeaders(verify bool, headers ...*block.Header) error { lastHeader = h } } - - bc.headerHashesLock.Lock() - defer bc.headerHashesLock.Unlock() - oldlen := len(bc.headerHashes) - var lastHeader *block.Header - for _, h := range headers { - if int(h.Index) != len(bc.headerHashes) { - continue - } - err = batch.StoreHeader(h) - if err != nil { - return err - } - bc.headerHashes = append(bc.headerHashes, h.Hash()) - lastHeader = h - } - - if oldlen != len(bc.headerHashes) { - for int(lastHeader.Index)-headerBatchCount >= int(bc.storedHeaderCount) { - err = batch.StoreHeaderHashes(bc.headerHashes[bc.storedHeaderCount:bc.storedHeaderCount+headerBatchCount], - bc.storedHeaderCount) - if err != nil { - return err - } - bc.storedHeaderCount += headerBatchCount - } - - batch.PutCurrentHeader(lastHeader.Hash(), lastHeader.Index) - updateHeaderHeightMetric(uint32(len(bc.headerHashes) - 1)) - if _, err = batch.Persist(); err != nil { - return err - } + res := bc.HeaderHashes.addHeaders(headers...) + if res == nil { bc.log.Debug("done processing headers", - zap.Int("headerIndex", len(bc.headerHashes)-1), + zap.Uint32("headerIndex", bc.HeaderHeight()), zap.Uint32("blockHeight", bc.BlockHeight()), zap.Duration("took", time.Since(start))) } - return nil + return res } // GetStateRoot returns state root for the given height. @@ -1528,7 +1462,7 @@ func (bc *Blockchain) storeBlock(block *block.Block, txpool *mempool.Pool) error stop = start + 1 } for index := start; index < stop; index++ { - err := kvcache.DeleteBlock(bc.headerHashes[index]) + err := kvcache.DeleteBlock(bc.GetHeaderHash(index)) if err != nil { bc.log.Warn("error while removing old block", zap.Uint32("index", index), @@ -2151,15 +2085,9 @@ func (bc *Blockchain) HasTransaction(hash util.Uint256) bool { // HasBlock returns true if the blockchain contains the given // block hash. func (bc *Blockchain) HasBlock(hash util.Uint256) bool { - var height = bc.BlockHeight() - bc.headerHashesLock.RLock() - for i := int(height); i >= int(height)-4 && i >= 0; i-- { - if hash.Equals(bc.headerHashes[i]) { - bc.headerHashesLock.RUnlock() - return true - } + if bc.HeaderHashes.haveRecentHash(hash, bc.BlockHeight()) { + return true } - bc.headerHashesLock.RUnlock() if header, err := bc.GetHeader(hash); err == nil { return header.Index <= bc.BlockHeight() @@ -2177,40 +2105,11 @@ func (bc *Blockchain) CurrentBlockHash() util.Uint256 { return bc.GetHeaderHash(bc.BlockHeight()) } -// CurrentHeaderHash returns the hash of the latest known header. -func (bc *Blockchain) CurrentHeaderHash() util.Uint256 { - bc.headerHashesLock.RLock() - hash := bc.headerHashes[len(bc.headerHashes)-1] - bc.headerHashesLock.RUnlock() - return hash -} - -// GetHeaderHash returns hash of the header/block with specified index, if -// Blockchain doesn't have a hash for this height, zero Uint256 value is returned. -func (bc *Blockchain) GetHeaderHash(i uint32) util.Uint256 { - bc.headerHashesLock.RLock() - defer bc.headerHashesLock.RUnlock() - - hashesLen := uint32(len(bc.headerHashes)) - if hashesLen <= i { - return util.Uint256{} - } - return bc.headerHashes[i] -} - // BlockHeight returns the height/index of the highest block. func (bc *Blockchain) BlockHeight() uint32 { return atomic.LoadUint32(&bc.blockHeight) } -// HeaderHeight returns the index/height of the highest header. -func (bc *Blockchain) HeaderHeight() uint32 { - bc.headerHashesLock.RLock() - n := len(bc.headerHashes) - bc.headerHashesLock.RUnlock() - return uint32(n - 1) -} - // GetContractState returns contract by its script hash. func (bc *Blockchain) GetContractState(hash util.Uint160) *state.Contract { contract, err := bc.contracts.Management.GetContract(bc.dao, hash) diff --git a/pkg/core/blockchain_core_test.go b/pkg/core/blockchain_core_test.go index 6d31d9638..529d12f49 100644 --- a/pkg/core/blockchain_core_test.go +++ b/pkg/core/blockchain_core_test.go @@ -225,7 +225,7 @@ func TestBlockchain_InitWithIncompleteStateJump(t *testing.T) { t.Run("invalid state sync point", func(t *testing.T) { bcSpout.dao.Store.Put(bPrefix, []byte{byte(stateJumpStarted)}) point := make([]byte, 4) - binary.LittleEndian.PutUint32(point, uint32(len(bcSpout.headerHashes))) + binary.LittleEndian.PutUint32(point, bcSpout.lastHeaderIndex()+1) bcSpout.dao.Store.Put([]byte{byte(storage.SYSStateSyncPoint)}, point) checkNewBlockchainErr(t, boltCfg, bcSpout.dao.Store, "invalid state sync point") }) diff --git a/pkg/core/blockchain_neotest_test.go b/pkg/core/blockchain_neotest_test.go index d1559e69c..3e23f4545 100644 --- a/pkg/core/blockchain_neotest_test.go +++ b/pkg/core/blockchain_neotest_test.go @@ -146,14 +146,15 @@ func TestBlockchain_StartFromExistingDB(t *testing.T) { // Corrupt headers hashes batch. cache := storage.NewMemCachedStore(ps) // Extra wrapper to avoid good DB corruption. - key := make([]byte, 5) - key[0] = byte(storage.IXHeaderHashList) - binary.BigEndian.PutUint32(key[1:], 1) - cache.Put(key, []byte{1, 2, 3}) + // Make the chain think we're at 2000+ which will trigger page 0 read. + buf := io.NewBufBinWriter() + buf.WriteBytes(util.Uint256{}.BytesLE()) + buf.WriteU32LE(2000) + cache.Put([]byte{byte(storage.SYSCurrentHeader)}, buf.Bytes()) _, _, _, err := chain.NewMultiWithCustomConfigAndStoreNoCheck(t, customConfig, cache) require.Error(t, err) - require.True(t, strings.Contains(err.Error(), "failed to read batch of 2000"), err) + require.True(t, strings.Contains(err.Error(), "failed to retrieve header hash page"), err) }) t.Run("corrupted current header height", func(t *testing.T) { ps = newPS(t) diff --git a/pkg/core/dao/dao.go b/pkg/core/dao/dao.go index f802dab52..8fb33efaa 100644 --- a/pkg/core/dao/dao.go +++ b/pkg/core/dao/dao.go @@ -1,7 +1,6 @@ package dao import ( - "bytes" "context" "encoding/binary" "errors" @@ -582,25 +581,23 @@ func (dao *Simple) GetStateSyncCurrentBlockHeight() (uint32, error) { return binary.LittleEndian.Uint32(b), nil } -// GetHeaderHashes returns a sorted list of header hashes retrieved from +// GetHeaderHashes returns a page of header hashes retrieved from // the given underlying store. -func (dao *Simple) GetHeaderHashes() ([]util.Uint256, error) { - var hashes = make([]util.Uint256, 0) +func (dao *Simple) GetHeaderHashes(height uint32) ([]util.Uint256, error) { + var hashes []util.Uint256 - var seekErr error - dao.Store.Seek(storage.SeekRange{ - Prefix: dao.mkKeyPrefix(storage.IXHeaderHashList), - }, func(k, v []byte) bool { - newHashes, err := read2000Uint256Hashes(v) - if err != nil { - seekErr = fmt.Errorf("failed to read batch of 2000 header hashes: %w", err) - return false - } - hashes = append(hashes, newHashes...) - return true - }) + key := dao.mkHeaderHashKey(height) + b, err := dao.Store.Get(key) + if err != nil { + return nil, err + } - return hashes, seekErr + br := io.NewBinReaderFromBuf(b) + br.ReadArray(&hashes) + if br.Err != nil { + return nil, br.Err + } + return hashes, nil } // DeleteHeaderHashes removes batches of header hashes starting from the one that @@ -683,19 +680,6 @@ func (dao *Simple) PutStateSyncCurrentBlockHeight(h uint32) { dao.Store.Put(dao.mkKeyPrefix(storage.SYSStateSyncCurrentBlockHeight), buf.Bytes()) } -// read2000Uint256Hashes attempts to read 2000 Uint256 hashes from -// the given byte array. -func read2000Uint256Hashes(b []byte) ([]util.Uint256, error) { - r := bytes.NewReader(b) - br := io.NewBinReaderFromIO(r) - hashes := make([]util.Uint256, 0) - br.ReadArray(&hashes) - if br.Err != nil { - return nil, br.Err - } - return hashes, nil -} - func (dao *Simple) mkHeaderHashKey(h uint32) []byte { b := dao.getKeyBuf(1 + 4) b[0] = byte(storage.IXHeaderHashList) diff --git a/pkg/core/headerhashes.go b/pkg/core/headerhashes.go new file mode 100644 index 000000000..4d9ce042a --- /dev/null +++ b/pkg/core/headerhashes.go @@ -0,0 +1,211 @@ +package core + +import ( + "fmt" + "sync" + + lru "github.com/hashicorp/golang-lru" + "github.com/nspcc-dev/neo-go/pkg/core/block" + "github.com/nspcc-dev/neo-go/pkg/core/dao" + "github.com/nspcc-dev/neo-go/pkg/util" +) + +const ( + headerBatchCount = 2000 + pagesCache = 8 +) + +// HeaderHashes is a header hash manager part of the Blockchain. It can't be used +// without Blockchain. +type HeaderHashes struct { + // Backing storage. + dao *dao.Simple + + // Lock for all internal state fields. + lock sync.RWMutex + + // The latest header hashes (storedHeaderCount+). + latest []util.Uint256 + + // Previously completed page of header hashes (pre-storedHeaderCount). + previous []util.Uint256 + + // Number of headers stored in the chain file. + storedHeaderCount uint32 + + // Cache for accessed pages of header hashes. + cache *lru.Cache +} + +func (h *HeaderHashes) initGenesis(dao *dao.Simple, hash util.Uint256) { + h.dao = dao + h.cache, _ = lru.New(pagesCache) // Never errors for positive size. + h.previous = make([]util.Uint256, headerBatchCount) + h.latest = make([]util.Uint256, 0, headerBatchCount) + h.latest = append(h.latest, hash) + dao.PutCurrentHeader(hash, 0) +} + +func (h *HeaderHashes) init(dao *dao.Simple) error { + h.dao = dao + h.cache, _ = lru.New(pagesCache) // Never errors for positive size. + + currHeaderHeight, currHeaderHash, err := h.dao.GetCurrentHeaderHeight() + if err != nil { + return fmt.Errorf("failed to retrieve current header info: %w", err) + } + h.storedHeaderCount = ((currHeaderHeight + 1) / headerBatchCount) * headerBatchCount + + if h.storedHeaderCount >= headerBatchCount { + h.previous, err = h.dao.GetHeaderHashes(h.storedHeaderCount - headerBatchCount) + if err != nil { + return fmt.Errorf("failed to retrieve header hash page %d: %w", h.storedHeaderCount-headerBatchCount, err) + } + } else { + h.previous = make([]util.Uint256, headerBatchCount) + } + h.latest = make([]util.Uint256, 0, headerBatchCount) + + // There is a high chance that the Node is stopped before the next + // batch of 2000 headers was stored. Via the currentHeaders stored we can sync + // that with stored blocks. + if currHeaderHeight >= h.storedHeaderCount { + hash := currHeaderHash + var targetHash util.Uint256 + if h.storedHeaderCount >= headerBatchCount { + targetHash = h.previous[len(h.previous)-1] + } + headers := make([]util.Uint256, 0, headerBatchCount) + + for hash != targetHash { + blk, err := h.dao.GetBlock(hash) + if err != nil { + return fmt.Errorf("could not get header %s: %w", hash, err) + } + headers = append(headers, blk.Hash()) + hash = blk.PrevHash + } + hashSliceReverse(headers) + h.latest = append(h.latest, headers...) + } + return nil +} + +func (h *HeaderHashes) lastHeaderIndex() uint32 { + return h.storedHeaderCount + uint32(len(h.latest)) - 1 +} + +// HeaderHeight returns the index/height of the highest header. +func (h *HeaderHashes) HeaderHeight() uint32 { + h.lock.RLock() + n := h.lastHeaderIndex() + h.lock.RUnlock() + return n +} + +func (h *HeaderHashes) addHeaders(headers ...*block.Header) error { + var ( + batch = h.dao.GetPrivate() + lastHeader *block.Header + err error + ) + + h.lock.Lock() + defer h.lock.Unlock() + + for _, head := range headers { + if head.Index != h.lastHeaderIndex()+1 { + continue + } + err = batch.StoreHeader(head) + if err != nil { + return err + } + lastHeader = head + h.latest = append(h.latest, head.Hash()) + if len(h.latest) == headerBatchCount { + err = batch.StoreHeaderHashes(h.latest, h.storedHeaderCount) + if err != nil { + return err + } + copy(h.previous, h.latest) + h.latest = h.latest[:0] + h.storedHeaderCount += headerBatchCount + } + } + if lastHeader != nil { + batch.PutCurrentHeader(lastHeader.Hash(), lastHeader.Index) + updateHeaderHeightMetric(lastHeader.Index) + if _, err = batch.Persist(); err != nil { + return err + } + } + return nil +} + +// CurrentHeaderHash returns the hash of the latest known header. +func (h *HeaderHashes) CurrentHeaderHash() util.Uint256 { + var hash util.Uint256 + + h.lock.RLock() + if len(h.latest) > 0 { + hash = h.latest[len(h.latest)-1] + } else { + hash = h.previous[len(h.previous)-1] + } + h.lock.RUnlock() + return hash +} + +// GetHeaderHash returns hash of the header/block with specified index, if +// HeaderHashes doesn't have a hash for this height, zero Uint256 value is returned. +func (h *HeaderHashes) GetHeaderHash(i uint32) util.Uint256 { + h.lock.RLock() + res, ok := h.getLocalHeaderHash(i) + h.lock.RUnlock() + if ok { + return res + } + // If it's not in the latest/previous, then it's in the cache or DB, those + // need no additional locks. + page := (i / headerBatchCount) * headerBatchCount + cache, ok := h.cache.Get(page) + if ok { + hashes := cache.([]util.Uint256) + return hashes[i-page] + } + hashes, err := h.dao.GetHeaderHashes(page) + if err != nil { + return util.Uint256{} + } + _ = h.cache.Add(page, hashes) + return hashes[i-page] +} + +// getLocalHeaderHash looks for the index in the latest and previous caches. +// Locking is left to the user. +func (h *HeaderHashes) getLocalHeaderHash(i uint32) (util.Uint256, bool) { + if i > h.lastHeaderIndex() { + return util.Uint256{}, false + } + if i >= h.storedHeaderCount { + return h.latest[i-h.storedHeaderCount], true + } + previousStored := h.storedHeaderCount - headerBatchCount + if i >= previousStored { + return h.previous[i-previousStored], true + } + return util.Uint256{}, false +} + +func (h *HeaderHashes) haveRecentHash(hash util.Uint256, i uint32) bool { + h.lock.RLock() + defer h.lock.RUnlock() + for ; i > 0; i-- { + lh, ok := h.getLocalHeaderHash(i) + if ok && hash.Equals(lh) { + return true + } + } + return false +}