restic/src/restic/repository/packer_manager.go

164 lines
3.4 KiB
Go
Raw Normal View History

package repository
import (
"crypto/sha256"
"io"
"io/ioutil"
"os"
2016-08-31 18:29:54 +00:00
"restic"
"sync"
2016-09-01 20:17:37 +00:00
"restic/errors"
"restic/hashing"
"restic/crypto"
"restic/debug"
"restic/fs"
"restic/pack"
)
// Saver implements saving data in a backend.
type Saver interface {
Save(restic.Handle, io.Reader) error
}
// Packer holds a pack.Packer together with a hash writer.
type Packer struct {
*pack.Packer
hw *hashing.Writer
tmpfile *os.File
}
// packerManager keeps a list of open packs and creates new on demand.
type packerManager struct {
be Saver
key *crypto.Key
pm sync.Mutex
packers []*Packer
2016-03-06 12:14:06 +00:00
pool sync.Pool
}
const minPackSize = 4 * 1024 * 1024
const maxPackSize = 16 * 1024 * 1024
const maxPackers = 200
2016-03-06 13:20:48 +00:00
// newPackerManager returns an new packer manager which writes temporary files
// to a temporary directory
2016-03-06 13:20:48 +00:00
func newPackerManager(be Saver, key *crypto.Key) *packerManager {
2016-03-06 12:14:06 +00:00
return &packerManager{
be: be,
key: key,
2016-03-06 12:14:06 +00:00
pool: sync.Pool{
New: func() interface{} {
return make([]byte, (minPackSize+maxPackSize)/2)
},
},
}
}
// findPacker returns a packer for a new blob of size bytes. Either a new one is
// created or one is returned that already has some blobs.
func (r *packerManager) findPacker(size uint) (packer *Packer, err error) {
r.pm.Lock()
defer r.pm.Unlock()
// search for a suitable packer
if len(r.packers) > 0 {
2016-09-27 20:35:08 +00:00
debug.Log("searching packer for %d bytes\n", size)
for i, p := range r.packers {
if p.Packer.Size()+size < maxPackSize {
2016-09-27 20:35:08 +00:00
debug.Log("found packer %v", p)
// remove from list
r.packers = append(r.packers[:i], r.packers[i+1:]...)
return p, nil
}
}
}
// no suitable packer found, return new
2016-09-27 20:35:08 +00:00
debug.Log("create new pack for %d bytes", size)
2016-03-06 12:14:06 +00:00
tmpfile, err := ioutil.TempFile("", "restic-temp-pack-")
if err != nil {
2016-08-29 20:16:58 +00:00
return nil, errors.Wrap(err, "ioutil.TempFile")
}
hw := hashing.NewWriter(tmpfile, sha256.New())
p := pack.NewPacker(r.key, hw)
packer = &Packer{
Packer: p,
hw: hw,
tmpfile: tmpfile,
}
return packer, nil
}
// insertPacker appends p to s.packs.
func (r *packerManager) insertPacker(p *Packer) {
r.pm.Lock()
defer r.pm.Unlock()
r.packers = append(r.packers, p)
debug.Log("%d packers\n", len(r.packers))
}
// savePacker stores p in the backend.
func (r *Repository) savePacker(p *Packer) error {
debug.Log("save packer with %d blobs\n", p.Packer.Count())
_, err := p.Packer.Finalize()
if err != nil {
return err
}
_, err = p.tmpfile.Seek(0, 0)
if err != nil {
return errors.Wrap(err, "Seek")
}
id := restic.IDFromHash(p.hw.Sum(nil))
2016-09-01 19:19:30 +00:00
h := restic.Handle{Type: restic.DataFile, Name: id.String()}
2016-01-24 18:30:14 +00:00
err = r.be.Save(h, p.tmpfile)
if err != nil {
2016-09-27 20:35:08 +00:00
debug.Log("Save(%v) error: %v", h, err)
return err
}
2016-09-27 20:35:08 +00:00
debug.Log("saved as %v", h)
err = p.tmpfile.Close()
if err != nil {
return errors.Wrap(err, "close tempfile")
}
err = fs.Remove(p.tmpfile.Name())
2016-03-06 12:14:06 +00:00
if err != nil {
2016-08-29 20:16:58 +00:00
return errors.Wrap(err, "Remove")
2016-03-06 12:14:06 +00:00
}
// update blobs in the index
for _, b := range p.Packer.Blobs() {
2016-09-27 20:35:08 +00:00
debug.Log(" updating blob %v to pack %v", b.ID.Str(), id.Str())
r.idx.Store(restic.PackedBlob{
2016-08-31 20:39:36 +00:00
Blob: restic.Blob{
Type: b.Type,
ID: b.ID,
Offset: b.Offset,
Length: uint(b.Length),
},
2016-01-24 18:30:14 +00:00
PackID: id,
})
}
return nil
}
// countPacker returns the number of open (unfinished) packers.
func (r *packerManager) countPacker() int {
r.pm.Lock()
defer r.pm.Unlock()
return len(r.packers)
}