Merge pull request #761 from nspcc-dev/feature/dump

cli: allow to reuse existing dumps
This commit is contained in:
Roman Khimov 2020-03-16 13:24:40 +03:00 committed by GitHub
commit d129c5c47b
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
2 changed files with 58 additions and 20 deletions

View file

@ -4,6 +4,7 @@ import (
"encoding/hex" "encoding/hex"
"encoding/json" "encoding/json"
"fmt" "fmt"
"io/ioutil"
"os" "os"
"path/filepath" "path/filepath"
@ -11,7 +12,13 @@ import (
"github.com/nspcc-dev/neo-go/pkg/util" "github.com/nspcc-dev/neo-go/pkg/util"
) )
type dump []interface{} type dump []blockDump
type blockDump struct {
Block uint32 `json:"block"`
Size int `json:"size"`
Storage []storageOp `json:"storage"`
}
type storageOp struct { type storageOp struct {
State string `json:"state"` State string `json:"state"`
@ -59,7 +66,7 @@ func toNeoStorageKey(key []byte) []byte {
// batchToMap converts batch to a map so that JSON is compatible // batchToMap converts batch to a map so that JSON is compatible
// with https://github.com/NeoResearch/neo-storage-audit/ // with https://github.com/NeoResearch/neo-storage-audit/
func batchToMap(index uint32, batch *storage.MemBatch) map[string]interface{} { func batchToMap(index uint32, batch *storage.MemBatch) blockDump {
size := len(batch.Put) + len(batch.Deleted) size := len(batch.Put) + len(batch.Deleted)
ops := make([]storageOp, 0, size) ops := make([]storageOp, 0, size)
for i := range batch.Put { for i := range batch.Put {
@ -94,10 +101,10 @@ func batchToMap(index uint32, batch *storage.MemBatch) map[string]interface{} {
}) })
} }
return map[string]interface{}{ return blockDump{
"block": index, Block: index,
"size": len(ops), Size: len(ops),
"storage": ops, Storage: ops,
} }
} }
@ -111,20 +118,28 @@ func (d *dump) add(index uint32, batch *storage.MemBatch) {
} }
func (d *dump) tryPersist(prefix string, index uint32) error { func (d *dump) tryPersist(prefix string, index uint32) error {
if index%1000 != 0 { if len(*d) == 0 {
return nil return nil
} }
path, err := getPath(prefix, index)
f, err := createFile(prefix, index) if err != nil {
return err
}
old, err := readFile(path)
if err == nil {
*old = append(*old, *d...)
} else {
old = d
}
f, err := os.Create(path)
if err != nil { if err != nil {
return err return err
} }
defer f.Close() defer f.Close()
enc := json.NewEncoder(f) enc := json.NewEncoder(f)
enc.SetIndent("", " ") enc.SetIndent("", " ")
if err := enc.Encode(*d); err != nil { if err := enc.Encode(*old); err != nil {
return err return err
} }
@ -133,14 +148,26 @@ func (d *dump) tryPersist(prefix string, index uint32) error {
return nil return nil
} }
// createFile creates directory and file in it for storing blocks up to index. func readFile(path string) (*dump, error) {
data, err := ioutil.ReadFile(path)
if err != nil {
return nil, err
}
d := newDump()
if err := json.Unmarshal(data, d); err != nil {
return nil, err
}
return d, err
}
// getPath returns filename for storing blocks up to index.
// Directory structure is the following: // Directory structure is the following:
// https://github.com/NeoResearch/neo-storage-audit#folder-organization-where-to-find-the-desired-block // https://github.com/NeoResearch/neo-storage-audit#folder-organization-where-to-find-the-desired-block
// Dir `BlockStorage_$DIRNO` contains blocks up to $DIRNO (from $DIRNO-100k) // Dir `BlockStorage_$DIRNO` contains blocks up to $DIRNO (from $DIRNO-100k)
// Inside it there are files grouped by 1k blocks. // Inside it there are files grouped by 1k blocks.
// File dump-block-$FILENO.json contains blocks from $FILENO-999, $FILENO // File dump-block-$FILENO.json contains blocks from $FILENO-999, $FILENO
// Example: file `BlockStorage_100000/dump-block-6000.json` contains blocks from 5001 to 6000. // Example: file `BlockStorage_100000/dump-block-6000.json` contains blocks from 5001 to 6000.
func createFile(prefix string, index uint32) (*os.File, error) { func getPath(prefix string, index uint32) (string, error) {
dirN := (index-1)/100000 + 1 dirN := (index-1)/100000 + 1
dir := fmt.Sprintf("BlockStorage_%d00000", dirN) dir := fmt.Sprintf("BlockStorage_%d00000", dirN)
@ -149,15 +176,13 @@ func createFile(prefix string, index uint32) (*os.File, error) {
if os.IsNotExist(err) { if os.IsNotExist(err) {
err := os.MkdirAll(path, os.ModePerm) err := os.MkdirAll(path, os.ModePerm)
if err != nil { if err != nil {
return nil, err return "", err
} }
} else if !info.IsDir() { } else if !info.IsDir() {
return nil, fmt.Errorf("file `%s` is not a directory", path) return "", fmt.Errorf("file `%s` is not a directory", path)
} }
fileN := (index-1)/1000 + 1 fileN := (index-1)/1000 + 1
file := fmt.Sprintf("dump-block-%d000.json", fileN) file := fmt.Sprintf("dump-block-%d000.json", fileN)
path = filepath.Join(path, file) return filepath.Join(path, file), nil
return os.Create(path)
} }

View file

@ -277,9 +277,19 @@ func restoreDB(ctx *cli.Context) error {
} }
} }
gctx := newGraceContext()
var lastIndex uint32
dump := newDump() dump := newDump()
defer func() {
_ = dump.tryPersist(dumpDir, lastIndex)
}()
for ; i < skip+count; i++ { for ; i < skip+count; i++ {
select {
case <-gctx.Done():
return cli.NewExitError("cancelled", 1)
default:
}
bytes, err := readBlock(reader) bytes, err := readBlock(reader)
block := &block.Block{} block := &block.Block{}
newReader := io.NewBinReaderFromBuf(bytes) newReader := io.NewBinReaderFromBuf(bytes)
@ -302,8 +312,11 @@ func restoreDB(ctx *cli.Context) error {
if dumpDir != "" { if dumpDir != "" {
batch := chain.LastBatch() batch := chain.LastBatch()
dump.add(block.Index, batch) dump.add(block.Index, batch)
if err := dump.tryPersist(dumpDir, block.Index); err != nil { lastIndex = block.Index
return cli.NewExitError(fmt.Errorf("can't dump storage to file: %v", err), 1) if block.Index%1000 == 0 {
if err := dump.tryPersist(dumpDir, block.Index); err != nil {
return cli.NewExitError(fmt.Errorf("can't dump storage to file: %v", err), 1)
}
} }
} }
} }