2014-12-05 20:45:49 +00:00
|
|
|
package restic
|
2014-09-23 20:39:12 +00:00
|
|
|
|
|
|
|
import (
|
2015-01-10 22:40:10 +00:00
|
|
|
"encoding/json"
|
2015-01-04 21:39:30 +00:00
|
|
|
"errors"
|
2014-11-30 21:49:14 +00:00
|
|
|
"fmt"
|
2014-11-17 22:28:51 +00:00
|
|
|
"io"
|
2014-09-23 20:39:12 +00:00
|
|
|
"os"
|
2014-11-16 21:50:20 +00:00
|
|
|
"sync"
|
2014-09-23 20:39:12 +00:00
|
|
|
|
2014-11-23 11:14:56 +00:00
|
|
|
"github.com/juju/arrar"
|
2014-12-05 20:45:49 +00:00
|
|
|
"github.com/restic/restic/backend"
|
|
|
|
"github.com/restic/restic/chunker"
|
2015-01-14 21:08:48 +00:00
|
|
|
"github.com/restic/restic/debug"
|
2015-02-15 13:44:54 +00:00
|
|
|
"github.com/restic/restic/pipe"
|
2014-09-23 20:39:12 +00:00
|
|
|
)
|
|
|
|
|
2014-11-16 21:50:20 +00:00
|
|
|
const (
|
2015-02-17 22:15:08 +00:00
|
|
|
maxConcurrentBlobs = 32
|
|
|
|
maxConcurrency = 10
|
|
|
|
maxConcurrencyPreload = 100
|
2015-02-09 22:38:50 +00:00
|
|
|
|
|
|
|
// chunkerBufSize is used in pool.go
|
|
|
|
chunkerBufSize = 512 * chunker.KiB
|
2014-11-16 21:50:20 +00:00
|
|
|
)
|
|
|
|
|
2014-09-23 20:39:12 +00:00
|
|
|
type Archiver struct {
|
2015-01-10 22:40:10 +00:00
|
|
|
s Server
|
|
|
|
m *Map
|
2014-09-23 20:39:12 +00:00
|
|
|
|
2014-11-22 21:05:39 +00:00
|
|
|
blobToken chan struct{}
|
2014-11-16 21:50:20 +00:00
|
|
|
|
2014-09-23 20:39:12 +00:00
|
|
|
Error func(dir string, fi os.FileInfo, err error) error
|
|
|
|
Filter func(item string, fi os.FileInfo) bool
|
2014-11-16 20:29:11 +00:00
|
|
|
|
2015-01-04 17:23:00 +00:00
|
|
|
p *Progress
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
|
2015-01-10 22:40:10 +00:00
|
|
|
func NewArchiver(s Server, p *Progress) (*Archiver, error) {
|
2014-09-23 20:39:12 +00:00
|
|
|
var err error
|
2014-11-16 21:50:20 +00:00
|
|
|
arch := &Archiver{
|
2014-12-21 16:02:49 +00:00
|
|
|
s: s,
|
2015-01-04 17:23:00 +00:00
|
|
|
p: p,
|
2014-11-22 21:05:39 +00:00
|
|
|
blobToken: make(chan struct{}, maxConcurrentBlobs),
|
2014-11-16 21:50:20 +00:00
|
|
|
}
|
|
|
|
|
2015-02-15 13:44:54 +00:00
|
|
|
// fill blob token
|
2014-11-22 21:05:39 +00:00
|
|
|
for i := 0; i < maxConcurrentBlobs; i++ {
|
|
|
|
arch.blobToken <- struct{}{}
|
|
|
|
}
|
|
|
|
|
2015-01-10 22:40:10 +00:00
|
|
|
// create new map to store all blobs in
|
|
|
|
arch.m = NewMap()
|
|
|
|
|
2014-09-23 20:39:12 +00:00
|
|
|
// abort on all errors
|
|
|
|
arch.Error = func(string, os.FileInfo, error) error { return err }
|
|
|
|
// allow all files
|
|
|
|
arch.Filter = func(string, os.FileInfo) bool { return true }
|
|
|
|
|
2015-01-10 22:40:10 +00:00
|
|
|
return arch, nil
|
|
|
|
}
|
|
|
|
|
2015-02-17 21:39:44 +00:00
|
|
|
// Preload loads all tree objects from repository and adds all blobs that are
|
|
|
|
// still available to the map for deduplication.
|
|
|
|
func (arch *Archiver) Preload() error {
|
2015-02-17 22:38:40 +00:00
|
|
|
debug.Log("Archiver.Preload", "Start loading known blobs")
|
|
|
|
|
2015-02-17 21:39:44 +00:00
|
|
|
// load all trees, in parallel
|
|
|
|
worker := func(wg *sync.WaitGroup, c <-chan backend.ID) {
|
|
|
|
for id := range c {
|
|
|
|
tree, err := LoadTree(arch.s, id)
|
|
|
|
// ignore error and advance to next tree
|
|
|
|
if err != nil {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2015-02-17 22:38:40 +00:00
|
|
|
debug.Log("Archiver.Preload", "load tree %v with %d blobs", id, tree.Map.Len())
|
|
|
|
|
2015-02-17 21:39:44 +00:00
|
|
|
arch.m.Merge(tree.Map)
|
|
|
|
}
|
|
|
|
wg.Done()
|
|
|
|
}
|
|
|
|
|
|
|
|
idCh := make(chan backend.ID)
|
|
|
|
|
|
|
|
// start workers
|
|
|
|
var wg sync.WaitGroup
|
2015-02-17 22:15:08 +00:00
|
|
|
for i := 0; i < maxConcurrencyPreload; i++ {
|
2015-02-17 21:39:44 +00:00
|
|
|
wg.Add(1)
|
|
|
|
go worker(&wg, idCh)
|
|
|
|
}
|
|
|
|
|
|
|
|
// list ids
|
2015-02-17 22:38:40 +00:00
|
|
|
trees := 0
|
2015-02-17 21:39:44 +00:00
|
|
|
err := arch.s.EachID(backend.Tree, func(id backend.ID) {
|
2015-02-17 22:38:40 +00:00
|
|
|
trees++
|
|
|
|
|
|
|
|
if trees%1000 == 0 {
|
|
|
|
debug.Log("Archiver.Preload", "Loaded %v trees", trees)
|
|
|
|
}
|
2015-02-17 21:39:44 +00:00
|
|
|
idCh <- id
|
|
|
|
})
|
|
|
|
|
|
|
|
close(idCh)
|
|
|
|
|
|
|
|
// wait for workers
|
|
|
|
wg.Wait()
|
|
|
|
|
2015-02-17 22:38:40 +00:00
|
|
|
debug.Log("Archiver.Preload", "Loaded %v blobs from %v trees", arch.m.Len(), trees)
|
|
|
|
|
2015-02-17 21:39:44 +00:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2015-02-08 21:54:45 +00:00
|
|
|
func (arch *Archiver) Save(t backend.Type, id backend.ID, length uint, rd io.Reader) (Blob, error) {
|
2015-01-14 21:08:48 +00:00
|
|
|
debug.Log("Archiver.Save", "Save(%v, %v)\n", t, id.Str())
|
2015-01-10 22:40:10 +00:00
|
|
|
|
|
|
|
// test if this blob is already known
|
|
|
|
blob, err := arch.m.FindID(id)
|
|
|
|
if err == nil {
|
2015-01-14 21:08:48 +00:00
|
|
|
debug.Log("Archiver.Save", "Save(%v, %v): reusing %v\n", t, id.Str(), blob.Storage.Str())
|
2015-01-10 22:40:10 +00:00
|
|
|
id.Free()
|
|
|
|
return blob, nil
|
2015-01-05 20:40:43 +00:00
|
|
|
}
|
2014-09-23 20:39:12 +00:00
|
|
|
|
2015-01-10 22:40:10 +00:00
|
|
|
// else encrypt and save data
|
2015-02-08 21:54:45 +00:00
|
|
|
blob, err = arch.s.SaveFrom(t, id, length, rd)
|
2015-01-10 22:40:10 +00:00
|
|
|
|
|
|
|
// store blob in storage map
|
|
|
|
smapblob := arch.m.Insert(blob)
|
|
|
|
|
|
|
|
// if the map has a different storage id for this plaintext blob, use that
|
|
|
|
// one and remove the other. This happens if the same plaintext blob was
|
|
|
|
// stored concurrently and finished earlier than this blob.
|
|
|
|
if blob.Storage.Compare(smapblob.Storage) != 0 {
|
2015-01-14 21:08:48 +00:00
|
|
|
debug.Log("Archiver.Save", "using other block, removing %v\n", blob.Storage.Str())
|
2015-01-10 22:40:10 +00:00
|
|
|
|
|
|
|
// remove the blob again
|
|
|
|
// TODO: implement a list of blobs in transport, so this doesn't happen so often
|
|
|
|
err = arch.s.Remove(t, blob.Storage)
|
|
|
|
if err != nil {
|
|
|
|
return Blob{}, err
|
|
|
|
}
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
|
2015-01-14 21:08:48 +00:00
|
|
|
debug.Log("Archiver.Save", "Save(%v, %v): new blob %v\n", t, id.Str(), blob)
|
2015-01-10 22:40:10 +00:00
|
|
|
|
|
|
|
return smapblob, nil
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
|
2015-01-10 22:40:10 +00:00
|
|
|
func (arch *Archiver) SaveTreeJSON(item interface{}) (Blob, error) {
|
|
|
|
// convert to json
|
|
|
|
data, err := json.Marshal(item)
|
2015-02-15 23:30:51 +00:00
|
|
|
// append newline
|
|
|
|
data = append(data, '\n')
|
2014-09-23 20:39:12 +00:00
|
|
|
if err != nil {
|
2014-11-21 20:21:44 +00:00
|
|
|
return Blob{}, err
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
|
2015-01-10 22:40:10 +00:00
|
|
|
// check if tree has been saved before
|
2015-02-15 23:30:51 +00:00
|
|
|
id := backend.Hash(data)
|
2015-01-10 22:40:10 +00:00
|
|
|
blob, err := arch.m.FindID(id)
|
2014-09-23 20:39:12 +00:00
|
|
|
|
2015-01-10 22:40:10 +00:00
|
|
|
// return the blob if we found it
|
|
|
|
if err == nil {
|
|
|
|
return blob, nil
|
|
|
|
}
|
2014-09-23 20:39:12 +00:00
|
|
|
|
2015-01-10 22:40:10 +00:00
|
|
|
// otherwise save the data
|
2015-02-15 23:30:51 +00:00
|
|
|
blob, err = arch.s.SaveJSON(backend.Tree, item)
|
2014-09-23 20:39:12 +00:00
|
|
|
if err != nil {
|
2014-11-21 20:21:44 +00:00
|
|
|
return Blob{}, err
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
|
2015-01-10 22:40:10 +00:00
|
|
|
// store blob in storage map
|
|
|
|
arch.m.Insert(blob)
|
2014-09-23 20:39:12 +00:00
|
|
|
|
|
|
|
return blob, nil
|
|
|
|
}
|
|
|
|
|
2014-11-17 22:28:51 +00:00
|
|
|
// SaveFile stores the content of the file on the backend as a Blob by calling
|
|
|
|
// Save for each chunk.
|
2015-01-10 22:40:10 +00:00
|
|
|
func (arch *Archiver) SaveFile(node *Node) (Blobs, error) {
|
2014-11-17 22:28:51 +00:00
|
|
|
file, err := os.Open(node.path)
|
|
|
|
defer file.Close()
|
2014-09-23 20:39:12 +00:00
|
|
|
if err != nil {
|
2015-01-10 22:40:10 +00:00
|
|
|
return nil, err
|
2015-01-04 21:39:30 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// check file again
|
|
|
|
fi, err := file.Stat()
|
|
|
|
if err != nil {
|
2015-01-10 22:40:10 +00:00
|
|
|
return nil, err
|
2015-01-04 21:39:30 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if fi.ModTime() != node.ModTime {
|
2015-01-10 22:40:10 +00:00
|
|
|
e2 := arch.Error(node.path, fi, errors.New("file was updated, using new version"))
|
2015-01-04 21:39:30 +00:00
|
|
|
|
|
|
|
if e2 == nil {
|
|
|
|
// create new node
|
|
|
|
n, err := NodeFromFileInfo(node.path, fi)
|
|
|
|
if err != nil {
|
2015-01-10 22:40:10 +00:00
|
|
|
return nil, err
|
2015-01-04 21:39:30 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// copy node
|
|
|
|
*node = *n
|
|
|
|
}
|
2014-11-17 22:28:51 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
var blobs Blobs
|
|
|
|
|
2015-02-08 21:54:45 +00:00
|
|
|
// store all chunks
|
2015-02-09 22:38:50 +00:00
|
|
|
chnker := GetChunker("archiver.SaveFile")
|
|
|
|
chnker.Reset(file)
|
2015-02-08 21:54:45 +00:00
|
|
|
chans := [](<-chan Blob){}
|
2015-02-09 22:38:50 +00:00
|
|
|
defer FreeChunker("archiver.SaveFile", chnker)
|
2014-11-23 15:48:00 +00:00
|
|
|
|
2015-02-08 21:54:45 +00:00
|
|
|
chunks := 0
|
2014-11-17 22:28:51 +00:00
|
|
|
|
2015-02-08 21:54:45 +00:00
|
|
|
for {
|
|
|
|
chunk, err := chnker.Next()
|
2014-11-30 21:16:34 +00:00
|
|
|
if err == io.EOF {
|
2015-02-08 21:54:45 +00:00
|
|
|
break
|
2014-11-17 22:28:51 +00:00
|
|
|
}
|
|
|
|
|
2015-02-08 21:54:45 +00:00
|
|
|
if err != nil {
|
|
|
|
return nil, arrar.Annotate(err, "SaveFile() chunker.Next()")
|
|
|
|
}
|
2014-11-17 22:28:51 +00:00
|
|
|
|
2015-02-08 21:54:45 +00:00
|
|
|
chunks++
|
2014-11-30 21:49:14 +00:00
|
|
|
|
2015-02-08 21:54:45 +00:00
|
|
|
// acquire token, start goroutine to save chunk
|
|
|
|
token := <-arch.blobToken
|
|
|
|
resCh := make(chan Blob, 1)
|
2014-11-22 21:05:39 +00:00
|
|
|
|
2015-02-08 21:54:45 +00:00
|
|
|
go func(ch chan<- Blob) {
|
|
|
|
blob, err := arch.Save(backend.Data, chunk.Digest, chunk.Length, chunk.Reader(file))
|
|
|
|
// TODO handle error
|
|
|
|
if err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
2014-11-17 22:28:51 +00:00
|
|
|
|
2015-02-08 21:54:45 +00:00
|
|
|
arch.p.Report(Stat{Bytes: blob.Size})
|
|
|
|
arch.blobToken <- token
|
|
|
|
ch <- blob
|
|
|
|
}(resCh)
|
2014-11-22 21:05:39 +00:00
|
|
|
|
2015-02-08 21:54:45 +00:00
|
|
|
chans = append(chans, resCh)
|
|
|
|
}
|
2014-11-17 22:28:51 +00:00
|
|
|
|
2015-02-08 21:54:45 +00:00
|
|
|
blobs = []Blob{}
|
|
|
|
for _, ch := range chans {
|
|
|
|
blobs = append(blobs, <-ch)
|
|
|
|
}
|
2014-11-30 21:49:14 +00:00
|
|
|
|
2015-02-08 21:54:45 +00:00
|
|
|
if len(blobs) != chunks {
|
|
|
|
return nil, fmt.Errorf("chunker returned %v chunks, but only %v blobs saved", chunks, len(blobs))
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
|
2014-12-07 12:30:16 +00:00
|
|
|
var bytes uint64
|
|
|
|
|
2014-09-23 20:39:12 +00:00
|
|
|
node.Content = make([]backend.ID, len(blobs))
|
2015-02-08 21:54:45 +00:00
|
|
|
debug.Log("Archiver.Save", "checking size for file %s", node.path)
|
2014-09-23 20:39:12 +00:00
|
|
|
for i, blob := range blobs {
|
|
|
|
node.Content[i] = blob.ID
|
2014-12-07 12:30:16 +00:00
|
|
|
bytes += blob.Size
|
2015-02-08 21:54:45 +00:00
|
|
|
|
|
|
|
debug.Log("Archiver.Save", " adding blob %s", blob)
|
2014-12-07 12:30:16 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if bytes != node.Size {
|
2015-01-10 22:40:10 +00:00
|
|
|
return nil, fmt.Errorf("errors saving node %q: saved %d bytes, wanted %d bytes", node.path, bytes, node.Size)
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
|
2015-01-14 21:08:48 +00:00
|
|
|
debug.Log("Archiver.SaveFile", "SaveFile(%q): %v\n", node.path, blobs)
|
2015-01-10 22:40:10 +00:00
|
|
|
|
|
|
|
return blobs, nil
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
|
2014-11-21 20:21:44 +00:00
|
|
|
func (arch *Archiver) saveTree(t *Tree) (Blob, error) {
|
2015-01-14 21:08:48 +00:00
|
|
|
debug.Log("Archiver.saveTree", "saveTree(%v)\n", t)
|
2014-11-16 21:50:20 +00:00
|
|
|
var wg sync.WaitGroup
|
|
|
|
|
2015-01-10 22:40:10 +00:00
|
|
|
// add all blobs to global map
|
|
|
|
arch.m.Merge(t.Map)
|
|
|
|
|
|
|
|
// TODO: do all this in parallel
|
|
|
|
for _, node := range t.Nodes {
|
|
|
|
if node.tree != nil {
|
2015-01-04 21:39:30 +00:00
|
|
|
b, err := arch.saveTree(node.tree)
|
2014-11-16 20:29:11 +00:00
|
|
|
if err != nil {
|
2014-11-21 20:21:44 +00:00
|
|
|
return Blob{}, err
|
2014-11-16 20:29:11 +00:00
|
|
|
}
|
|
|
|
node.Subtree = b.ID
|
2015-01-10 22:40:10 +00:00
|
|
|
t.Map.Insert(b)
|
2015-01-04 17:23:00 +00:00
|
|
|
arch.p.Report(Stat{Dirs: 1})
|
2015-01-10 22:40:10 +00:00
|
|
|
} else if node.Type == "file" {
|
|
|
|
if len(node.Content) > 0 {
|
|
|
|
removeContent := false
|
|
|
|
|
|
|
|
// check content
|
|
|
|
for _, id := range node.Content {
|
|
|
|
blob, err := t.Map.FindID(id)
|
|
|
|
if err != nil {
|
2015-01-14 21:08:48 +00:00
|
|
|
debug.Log("Archiver.saveTree", "unable to find storage id for data blob %v", id.Str())
|
2015-01-10 22:40:10 +00:00
|
|
|
arch.Error(node.path, nil, fmt.Errorf("unable to find storage id for data blob %v", id.Str()))
|
|
|
|
removeContent = true
|
|
|
|
t.Map.DeleteID(id)
|
|
|
|
arch.m.DeleteID(id)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
if ok, err := arch.s.Test(backend.Data, blob.Storage); !ok || err != nil {
|
2015-01-14 21:08:48 +00:00
|
|
|
debug.Log("Archiver.saveTree", "blob %v not in repository (error is %v)", blob, err)
|
2015-01-10 22:40:10 +00:00
|
|
|
arch.Error(node.path, nil, fmt.Errorf("blob %v not in repository (error is %v)", blob.Storage.Str(), err))
|
|
|
|
removeContent = true
|
|
|
|
t.Map.DeleteID(id)
|
|
|
|
arch.m.DeleteID(id)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if removeContent {
|
2015-01-14 21:08:48 +00:00
|
|
|
debug.Log("Archiver.saveTree", "removing content for %s", node.path)
|
2015-01-10 22:40:10 +00:00
|
|
|
node.Content = node.Content[:0]
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(node.Content) == 0 {
|
|
|
|
// start goroutine
|
|
|
|
wg.Add(1)
|
|
|
|
go func(n *Node) {
|
|
|
|
defer wg.Done()
|
|
|
|
|
|
|
|
var blobs Blobs
|
2015-01-14 16:48:23 +00:00
|
|
|
blobs, n.err = arch.SaveFile(n)
|
2015-01-10 22:40:10 +00:00
|
|
|
for _, b := range blobs {
|
|
|
|
t.Map.Insert(b)
|
|
|
|
}
|
|
|
|
|
|
|
|
arch.p.Report(Stat{Files: 1})
|
|
|
|
}(node)
|
|
|
|
}
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-11-16 21:50:20 +00:00
|
|
|
wg.Wait()
|
|
|
|
|
2015-01-10 22:40:10 +00:00
|
|
|
usedIDs := backend.NewIDSet()
|
|
|
|
|
2014-11-30 23:06:29 +00:00
|
|
|
// check for invalid file nodes
|
2015-01-10 22:40:10 +00:00
|
|
|
for _, node := range t.Nodes {
|
2015-01-04 21:58:58 +00:00
|
|
|
if node.Type == "file" && node.Content == nil && node.err == nil {
|
2014-11-30 23:06:29 +00:00
|
|
|
return Blob{}, fmt.Errorf("node %v has empty content", node.Name)
|
|
|
|
}
|
2015-01-04 21:58:58 +00:00
|
|
|
|
2015-01-10 22:40:10 +00:00
|
|
|
// remember used hashes
|
|
|
|
if node.Type == "file" && node.Content != nil {
|
|
|
|
for _, id := range node.Content {
|
|
|
|
usedIDs.Insert(id)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if node.Type == "dir" && node.Subtree != nil {
|
|
|
|
usedIDs.Insert(node.Subtree)
|
|
|
|
}
|
|
|
|
|
2015-01-04 21:58:58 +00:00
|
|
|
if node.err != nil {
|
|
|
|
err := arch.Error(node.path, nil, node.err)
|
|
|
|
if err != nil {
|
|
|
|
return Blob{}, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// save error message in node
|
|
|
|
node.Error = node.err.Error()
|
|
|
|
}
|
2014-11-30 23:06:29 +00:00
|
|
|
}
|
|
|
|
|
2015-01-10 22:40:10 +00:00
|
|
|
before := len(t.Map.IDs())
|
|
|
|
t.Map.Prune(usedIDs)
|
|
|
|
after := len(t.Map.IDs())
|
|
|
|
|
|
|
|
if before != after {
|
2015-01-14 21:08:48 +00:00
|
|
|
debug.Log("Archiver.saveTree", "pruned %d ids from map for tree %v\n", before-after, t)
|
2015-01-10 22:40:10 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
blob, err := arch.SaveTreeJSON(t)
|
2014-11-16 20:29:11 +00:00
|
|
|
if err != nil {
|
2014-11-21 20:21:44 +00:00
|
|
|
return Blob{}, err
|
2014-11-16 20:29:11 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return blob, nil
|
|
|
|
}
|
|
|
|
|
2015-02-15 13:44:54 +00:00
|
|
|
func (arch *Archiver) Snapshot(path string, parentSnapshot backend.ID) (*Snapshot, backend.ID, error) {
|
2015-01-14 21:08:48 +00:00
|
|
|
debug.Break("Archiver.Snapshot")
|
2015-01-11 13:09:44 +00:00
|
|
|
|
2015-01-04 17:23:00 +00:00
|
|
|
arch.p.Start()
|
|
|
|
defer arch.p.Done()
|
2014-11-23 08:22:18 +00:00
|
|
|
|
2015-02-15 13:44:54 +00:00
|
|
|
sn, err := NewSnapshot(path)
|
2014-12-21 16:20:49 +00:00
|
|
|
if err != nil {
|
|
|
|
return nil, nil, err
|
|
|
|
}
|
|
|
|
|
2014-11-30 21:34:21 +00:00
|
|
|
sn.Parent = parentSnapshot
|
2014-11-16 20:29:11 +00:00
|
|
|
|
2015-02-15 13:44:54 +00:00
|
|
|
done := make(chan struct{})
|
|
|
|
entCh := make(chan pipe.Entry)
|
|
|
|
dirCh := make(chan pipe.Dir)
|
|
|
|
|
|
|
|
fileWorker := func(wg *sync.WaitGroup, done <-chan struct{}, entCh <-chan pipe.Entry) {
|
|
|
|
defer wg.Done()
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case e, ok := <-entCh:
|
|
|
|
if !ok {
|
|
|
|
// channel is closed
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
node, err := NodeFromFileInfo(e.Path, e.Info)
|
|
|
|
if err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if node.Type == "file" {
|
|
|
|
node.blobs, err = arch.SaveFile(node)
|
|
|
|
if err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
e.Result <- node
|
|
|
|
case <-done:
|
|
|
|
// pipeline was cancelled
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
dirWorker := func(wg *sync.WaitGroup, done <-chan struct{}, dirCh <-chan pipe.Dir) {
|
|
|
|
defer wg.Done()
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case dir, ok := <-dirCh:
|
|
|
|
if !ok {
|
|
|
|
// channel is closed
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
tree := NewTree()
|
|
|
|
|
|
|
|
// wait for all content
|
|
|
|
for _, ch := range dir.Entries {
|
|
|
|
node := (<-ch).(*Node)
|
|
|
|
tree.Insert(node)
|
|
|
|
|
|
|
|
if node.Type == "dir" {
|
|
|
|
debug.Log("Archiver.DirWorker", "got tree node for %s: %v", node.path, node.blobs)
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, blob := range node.blobs {
|
|
|
|
tree.Map.Insert(blob)
|
|
|
|
arch.m.Insert(blob)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
node, err := NodeFromFileInfo(dir.Path, dir.Info)
|
|
|
|
if err != nil {
|
|
|
|
node.Error = err.Error()
|
|
|
|
dir.Result <- node
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
blob, err := arch.SaveTreeJSON(tree)
|
|
|
|
if err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
|
|
|
debug.Log("Archiver.DirWorker", "save tree for %s: %v", dir.Path, blob)
|
|
|
|
|
|
|
|
node.Subtree = blob.ID
|
|
|
|
node.blobs = Blobs{blob}
|
|
|
|
|
|
|
|
dir.Result <- node
|
|
|
|
case <-done:
|
|
|
|
// pipeline was cancelled
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
for i := 0; i < maxConcurrency; i++ {
|
|
|
|
wg.Add(2)
|
|
|
|
go fileWorker(&wg, done, entCh)
|
|
|
|
go dirWorker(&wg, done, dirCh)
|
|
|
|
}
|
|
|
|
|
|
|
|
resCh, err := pipe.Walk(path, done, entCh, dirCh)
|
|
|
|
if err != nil {
|
|
|
|
close(done)
|
|
|
|
}
|
|
|
|
|
|
|
|
// wait for all workers to terminate
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
return nil, nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// wait for top-level node
|
|
|
|
node := (<-resCh).(*Node)
|
|
|
|
|
|
|
|
// add tree for top-level directory
|
|
|
|
tree := NewTree()
|
|
|
|
tree.Insert(node)
|
|
|
|
for _, blob := range node.blobs {
|
|
|
|
blob = arch.m.Insert(blob)
|
|
|
|
tree.Map.Insert(blob)
|
|
|
|
}
|
|
|
|
|
|
|
|
tb, err := arch.SaveTreeJSON(tree)
|
2014-09-23 20:39:12 +00:00
|
|
|
if err != nil {
|
|
|
|
return nil, nil, err
|
|
|
|
}
|
2015-02-15 13:44:54 +00:00
|
|
|
|
|
|
|
sn.Tree = tb
|
2014-11-23 21:26:01 +00:00
|
|
|
|
2014-09-23 20:39:12 +00:00
|
|
|
// save snapshot
|
2015-02-15 13:44:54 +00:00
|
|
|
blob, err := arch.s.SaveJSON(backend.Snapshot, sn)
|
2014-09-23 20:39:12 +00:00
|
|
|
if err != nil {
|
|
|
|
return nil, nil, err
|
|
|
|
}
|
|
|
|
|
2014-11-16 20:29:11 +00:00
|
|
|
return sn, blob.Storage, nil
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|