Rename 'Repository' -> Repo

This commit is contained in:
Alexander Neumann 2015-05-09 17:41:28 +02:00
parent 5fc1583acc
commit ae21938f3e
15 changed files with 106 additions and 106 deletions

View file

@ -30,7 +30,7 @@ var archiverAllowAllFiles = func(string, os.FileInfo) bool { return true }
// Archiver is used to backup a set of directories. // Archiver is used to backup a set of directories.
type Archiver struct { type Archiver struct {
repo *repo.Repository repo *repo.Repo
blobToken chan struct{} blobToken chan struct{}
@ -39,7 +39,7 @@ type Archiver struct {
} }
// NewArchiver returns a new archiver. // NewArchiver returns a new archiver.
func NewArchiver(repo *repo.Repository) *Archiver { func NewArchiver(repo *repo.Repo) *Archiver {
arch := &Archiver{ arch := &Archiver{
repo: repo, repo: repo,
blobToken: make(chan struct{}, maxConcurrentBlobs), blobToken: make(chan struct{}, maxConcurrentBlobs),

View file

@ -18,7 +18,7 @@ type Cache struct {
base string base string
} }
func NewCache(repo *repo.Repository) (*Cache, error) { func NewCache(repo *repo.Repo) (*Cache, error) {
cacheDir, err := getCacheDir() cacheDir, err := getCacheDir()
if err != nil { if err != nil {
return nil, err return nil, err
@ -106,7 +106,7 @@ func (c *Cache) purge(t backend.Type, subtype string, id backend.ID) error {
} }
// Clear removes information from the cache that isn't present in the repository any more. // Clear removes information from the cache that isn't present in the repository any more.
func (c *Cache) Clear(repo *repo.Repository) error { func (c *Cache) Clear(repo *repo.Repo) error {
list, err := c.list(backend.Snapshot) list, err := c.list(backend.Snapshot)
if err != nil { if err != nil {
return err return err

View file

@ -59,7 +59,7 @@ func parseTime(str string) (time.Time, error) {
return time.Time{}, fmt.Errorf("unable to parse time: %q", str) return time.Time{}, fmt.Errorf("unable to parse time: %q", str)
} }
func (c CmdFind) findInTree(repo *repo.Repository, id backend.ID, path string) ([]findResult, error) { func (c CmdFind) findInTree(repo *repo.Repo, id backend.ID, path string) ([]findResult, error) {
debug.Log("restic.find", "checking tree %v\n", id) debug.Log("restic.find", "checking tree %v\n", id)
tree, err := restic.LoadTree(repo, id) tree, err := restic.LoadTree(repo, id)
if err != nil { if err != nil {
@ -105,7 +105,7 @@ func (c CmdFind) findInTree(repo *repo.Repository, id backend.ID, path string) (
return results, nil return results, nil
} }
func (c CmdFind) findInSnapshot(repo *repo.Repository, name string) error { func (c CmdFind) findInSnapshot(repo *repo.Repo, name string) error {
debug.Log("restic.find", "searching in snapshot %s\n for entries within [%s %s]", name, c.oldest, c.newest) debug.Log("restic.find", "searching in snapshot %s\n for entries within [%s %s]", name, c.oldest, c.newest)
id, err := backend.ParseID(name) id, err := backend.ParseID(name)

View file

@ -34,7 +34,7 @@ func init() {
} }
} }
func fsckFile(opts CmdFsck, repo *repo.Repository, IDs []backend.ID) (uint64, error) { func fsckFile(opts CmdFsck, repo *repo.Repo, IDs []backend.ID) (uint64, error) {
debug.Log("restic.fsckFile", "checking file %v", IDs) debug.Log("restic.fsckFile", "checking file %v", IDs)
var bytes uint64 var bytes uint64
@ -77,7 +77,7 @@ func fsckFile(opts CmdFsck, repo *repo.Repository, IDs []backend.ID) (uint64, er
return bytes, nil return bytes, nil
} }
func fsckTree(opts CmdFsck, repo *repo.Repository, id backend.ID) error { func fsckTree(opts CmdFsck, repo *repo.Repo, id backend.ID) error {
debug.Log("restic.fsckTree", "checking tree %v", id.Str()) debug.Log("restic.fsckTree", "checking tree %v", id.Str())
tree, err := restic.LoadTree(repo, id) tree, err := restic.LoadTree(repo, id)
@ -157,7 +157,7 @@ func fsckTree(opts CmdFsck, repo *repo.Repository, id backend.ID) error {
return firstErr return firstErr
} }
func fsckSnapshot(opts CmdFsck, repo *repo.Repository, id backend.ID) error { func fsckSnapshot(opts CmdFsck, repo *repo.Repo, id backend.ID) error {
debug.Log("restic.fsck", "checking snapshot %v\n", id) debug.Log("restic.fsck", "checking snapshot %v\n", id)
sn, err := restic.LoadSnapshot(repo, id) sn, err := restic.LoadSnapshot(repo, id)

View file

@ -21,7 +21,7 @@ func init() {
} }
} }
func listKeys(s *repo.Repository) error { func listKeys(s *repo.Repo) error {
tab := NewTable() tab := NewTable()
tab.Header = fmt.Sprintf(" %-10s %-10s %-10s %s", "ID", "User", "Host", "Created") tab.Header = fmt.Sprintf(" %-10s %-10s %-10s %s", "ID", "User", "Host", "Created")
tab.RowFormat = "%s%-10s %-10s %-10s %s" tab.RowFormat = "%s%-10s %-10s %-10s %s"
@ -56,7 +56,7 @@ func listKeys(s *repo.Repository) error {
return nil return nil
} }
func addKey(s *repo.Repository) error { func addKey(s *repo.Repo) error {
pw := readPassword("RESTIC_NEWPASSWORD", "enter password for new key: ") pw := readPassword("RESTIC_NEWPASSWORD", "enter password for new key: ")
pw2 := readPassword("RESTIC_NEWPASSWORD", "enter password again: ") pw2 := readPassword("RESTIC_NEWPASSWORD", "enter password again: ")
@ -74,7 +74,7 @@ func addKey(s *repo.Repository) error {
return nil return nil
} }
func deleteKey(repo *repo.Repository, name string) error { func deleteKey(repo *repo.Repo, name string) error {
if name == repo.KeyName() { if name == repo.KeyName() {
return errors.New("refusing to remove key currently used to access repository") return errors.New("refusing to remove key currently used to access repository")
} }
@ -88,7 +88,7 @@ func deleteKey(repo *repo.Repository, name string) error {
return nil return nil
} }
func changePassword(s *repo.Repository) error { func changePassword(s *repo.Repo) error {
pw := readPassword("RESTIC_NEWPASSWORD", "enter password for new key: ") pw := readPassword("RESTIC_NEWPASSWORD", "enter password for new key: ")
pw2 := readPassword("RESTIC_NEWPASSWORD", "enter password again: ") pw2 := readPassword("RESTIC_NEWPASSWORD", "enter password again: ")

View file

@ -38,7 +38,7 @@ func printNode(prefix string, n *restic.Node) string {
} }
} }
func printTree(prefix string, repo *repo.Repository, id backend.ID) error { func printTree(prefix string, repo *repo.Repo, id backend.ID) error {
tree, err := restic.LoadTree(repo, id) tree, err := restic.LoadTree(repo, id)
if err != nil { if err != nil {
return err return err

View file

@ -133,7 +133,7 @@ func create(u string) (backend.Backend, error) {
return sftp.Create(url.Path[1:], "ssh", args...) return sftp.Create(url.Path[1:], "ssh", args...)
} }
func OpenRepo() (*repo.Repository, error) { func OpenRepo() (*repo.Repo, error) {
if opts.Repo == "" { if opts.Repo == "" {
return nil, errors.New("Please specify repository location (-r)") return nil, errors.New("Please specify repository location (-r)")
} }

View file

@ -103,7 +103,7 @@ func nodeTypeFromFileInfo(fi os.FileInfo) string {
} }
// CreateAt creates the node at the given path and restores all the meta data. // CreateAt creates the node at the given path and restores all the meta data.
func (node *Node) CreateAt(path string, repo *repo.Repository) error { func (node *Node) CreateAt(path string, repo *repo.Repo) error {
switch node.Type { switch node.Type {
case "dir": case "dir":
if err := node.createDirAt(path); err != nil { if err := node.createDirAt(path); err != nil {
@ -176,7 +176,7 @@ func (node Node) createDirAt(path string) error {
return nil return nil
} }
func (node Node) createFileAt(path string, repo *repo.Repository) error { func (node Node) createFileAt(path string, repo *repo.Repo) error {
f, err := os.OpenFile(path, os.O_CREATE|os.O_WRONLY, 0600) f, err := os.OpenFile(path, os.O_CREATE|os.O_WRONLY, 0600)
defer f.Close() defer f.Close()

View file

@ -49,12 +49,12 @@ type Key struct {
// createMasterKey creates a new master key in the given backend and encrypts // createMasterKey creates a new master key in the given backend and encrypts
// it with the password. // it with the password.
func createMasterKey(s *Repository, password string) (*Key, error) { func createMasterKey(s *Repo, password string) (*Key, error) {
return AddKey(s, password, nil) return AddKey(s, password, nil)
} }
// OpenKey tries do decrypt the key specified by name with the given password. // OpenKey tries do decrypt the key specified by name with the given password.
func OpenKey(s *Repository, name string, password string) (*Key, error) { func OpenKey(s *Repo, name string, password string) (*Key, error) {
k, err := LoadKey(s, name) k, err := LoadKey(s, name)
if err != nil { if err != nil {
return nil, err return nil, err
@ -94,7 +94,7 @@ func OpenKey(s *Repository, name string, password string) (*Key, error) {
// SearchKey tries to decrypt all keys in the backend with the given password. // SearchKey tries to decrypt all keys in the backend with the given password.
// If none could be found, ErrNoKeyFound is returned. // If none could be found, ErrNoKeyFound is returned.
func SearchKey(s *Repository, password string) (*Key, error) { func SearchKey(s *Repo, password string) (*Key, error) {
// try all keys in repo // try all keys in repo
done := make(chan struct{}) done := make(chan struct{})
defer close(done) defer close(done)
@ -111,7 +111,7 @@ func SearchKey(s *Repository, password string) (*Key, error) {
} }
// LoadKey loads a key from the backend. // LoadKey loads a key from the backend.
func LoadKey(s *Repository, name string) (*Key, error) { func LoadKey(s *Repo, name string) (*Key, error) {
// extract data from repo // extract data from repo
rd, err := s.be.Get(backend.Key, name) rd, err := s.be.Get(backend.Key, name)
if err != nil { if err != nil {
@ -131,7 +131,7 @@ func LoadKey(s *Repository, name string) (*Key, error) {
} }
// AddKey adds a new key to an already existing repository. // AddKey adds a new key to an already existing repository.
func AddKey(s *Repository, password string, template *crypto.Key) (*Key, error) { func AddKey(s *Repo, password string, template *crypto.Key) (*Key, error) {
// fill meta data about key // fill meta data about key
newkey := &Key{ newkey := &Key{
Created: time.Now(), Created: time.Now(),

View file

@ -26,8 +26,8 @@ type Config struct {
ChunkerPolynomial chunker.Pol `json:"chunker_polynomial"` ChunkerPolynomial chunker.Pol `json:"chunker_polynomial"`
} }
// Repository is used to access a repository in a backend. // Repo is used to access a repository in a backend.
type Repository struct { type Repo struct {
be backend.Backend be backend.Backend
Config Config Config Config
key *crypto.Key key *crypto.Key
@ -38,8 +38,8 @@ type Repository struct {
packs []*pack.Packer packs []*pack.Packer
} }
func New(be backend.Backend) *Repository { func New(be backend.Backend) *Repo {
return &Repository{ return &Repo{
be: be, be: be,
idx: NewIndex(), idx: NewIndex(),
} }
@ -48,31 +48,31 @@ func New(be backend.Backend) *Repository {
// Find loads the list of all blobs of type t and searches for names which start // Find loads the list of all blobs of type t and searches for names which start
// with prefix. If none is found, nil and ErrNoIDPrefixFound is returned. If // with prefix. If none is found, nil and ErrNoIDPrefixFound is returned. If
// more than one is found, nil and ErrMultipleIDMatches is returned. // more than one is found, nil and ErrMultipleIDMatches is returned.
func (s *Repository) Find(t backend.Type, prefix string) (string, error) { func (s *Repo) Find(t backend.Type, prefix string) (string, error) {
return backend.Find(s.be, t, prefix) return backend.Find(s.be, t, prefix)
} }
// FindSnapshot takes a string and tries to find a snapshot whose ID matches // FindSnapshot takes a string and tries to find a snapshot whose ID matches
// the string as closely as possible. // the string as closely as possible.
func (s *Repository) FindSnapshot(name string) (string, error) { func (s *Repo) FindSnapshot(name string) (string, error) {
return backend.FindSnapshot(s.be, name) return backend.FindSnapshot(s.be, name)
} }
// PrefixLength returns the number of bytes required so that all prefixes of // PrefixLength returns the number of bytes required so that all prefixes of
// all IDs of type t are unique. // all IDs of type t are unique.
func (s *Repository) PrefixLength(t backend.Type) (int, error) { func (s *Repo) PrefixLength(t backend.Type) (int, error) {
return backend.PrefixLength(s.be, t) return backend.PrefixLength(s.be, t)
} }
// Load tries to load and decrypt content identified by t and id from the // Load tries to load and decrypt content identified by t and id from the
// backend. // backend.
func (s *Repository) Load(t backend.Type, id backend.ID) ([]byte, error) { func (s *Repo) Load(t backend.Type, id backend.ID) ([]byte, error) {
debug.Log("Repository.Load", "load %v with id %v", t, id.Str()) debug.Log("Repo.Load", "load %v with id %v", t, id.Str())
// load blob from pack // load blob from pack
rd, err := s.be.Get(t, id.String()) rd, err := s.be.Get(t, id.String())
if err != nil { if err != nil {
debug.Log("Repository.Load", "error loading %v: %v", id.Str(), err) debug.Log("Repo.Load", "error loading %v: %v", id.Str(), err)
return nil, err return nil, err
} }
@ -102,26 +102,26 @@ func (s *Repository) Load(t backend.Type, id backend.ID) ([]byte, error) {
// LoadBlob tries to load and decrypt content identified by t and id from a // LoadBlob tries to load and decrypt content identified by t and id from a
// pack from the backend. // pack from the backend.
func (s *Repository) LoadBlob(t pack.BlobType, id backend.ID) ([]byte, error) { func (s *Repo) LoadBlob(t pack.BlobType, id backend.ID) ([]byte, error) {
debug.Log("Repository.LoadBlob", "load %v with id %v", t, id.Str()) debug.Log("Repo.LoadBlob", "load %v with id %v", t, id.Str())
// lookup pack // lookup pack
packID, tpe, offset, length, err := s.idx.Lookup(id) packID, tpe, offset, length, err := s.idx.Lookup(id)
if err != nil { if err != nil {
debug.Log("Repository.LoadBlob", "id %v not found in index: %v", id.Str(), err) debug.Log("Repo.LoadBlob", "id %v not found in index: %v", id.Str(), err)
return nil, err return nil, err
} }
if tpe != t { if tpe != t {
debug.Log("Repository.LoadBlob", "wrong type returned for %v: wanted %v, got %v", id.Str(), t, tpe) debug.Log("Repo.LoadBlob", "wrong type returned for %v: wanted %v, got %v", id.Str(), t, tpe)
return nil, fmt.Errorf("blob has wrong type %v (wanted: %v)", tpe, t) return nil, fmt.Errorf("blob has wrong type %v (wanted: %v)", tpe, t)
} }
debug.Log("Repository.LoadBlob", "id %v found in pack %v at offset %v (length %d)", id.Str(), packID.Str(), offset, length) debug.Log("Repo.LoadBlob", "id %v found in pack %v at offset %v (length %d)", id.Str(), packID.Str(), offset, length)
// load blob from pack // load blob from pack
rd, err := s.be.GetReader(backend.Data, packID.String(), offset, length) rd, err := s.be.GetReader(backend.Data, packID.String(), offset, length)
if err != nil { if err != nil {
debug.Log("Repository.LoadBlob", "error loading pack %v for %v: %v", packID.Str(), id.Str(), err) debug.Log("Repo.LoadBlob", "error loading pack %v for %v: %v", packID.Str(), id.Str(), err)
return nil, err return nil, err
} }
@ -151,7 +151,7 @@ func (s *Repository) LoadBlob(t pack.BlobType, id backend.ID) ([]byte, error) {
// LoadJSONUnpacked decrypts the data and afterwards calls json.Unmarshal on // LoadJSONUnpacked decrypts the data and afterwards calls json.Unmarshal on
// the item. // the item.
func (s *Repository) LoadJSONUnpacked(t backend.Type, id backend.ID, item interface{}) error { func (s *Repo) LoadJSONUnpacked(t backend.Type, id backend.ID, item interface{}) error {
// load blob from backend // load blob from backend
rd, err := s.be.Get(t, id.String()) rd, err := s.be.Get(t, id.String())
if err != nil { if err != nil {
@ -178,7 +178,7 @@ func (s *Repository) LoadJSONUnpacked(t backend.Type, id backend.ID, item interf
// LoadJSONPack calls LoadBlob() to load a blob from the backend, decrypt the // LoadJSONPack calls LoadBlob() to load a blob from the backend, decrypt the
// data and afterwards call json.Unmarshal on the item. // data and afterwards call json.Unmarshal on the item.
func (s *Repository) LoadJSONPack(t pack.BlobType, id backend.ID, item interface{}) error { func (s *Repo) LoadJSONPack(t pack.BlobType, id backend.ID, item interface{}) error {
// lookup pack // lookup pack
packID, _, offset, length, err := s.idx.Lookup(id) packID, _, offset, length, err := s.idx.Lookup(id)
if err != nil { if err != nil {
@ -215,16 +215,16 @@ const maxPackers = 200
// findPacker returns a packer for a new blob of size bytes. Either a new one is // findPacker returns a packer for a new blob of size bytes. Either a new one is
// created or one is returned that already has some blobs. // created or one is returned that already has some blobs.
func (s *Repository) findPacker(size uint) (*pack.Packer, error) { func (s *Repo) findPacker(size uint) (*pack.Packer, error) {
s.pm.Lock() s.pm.Lock()
defer s.pm.Unlock() defer s.pm.Unlock()
// search for a suitable packer // search for a suitable packer
if len(s.packs) > 0 { if len(s.packs) > 0 {
debug.Log("Repository.findPacker", "searching packer for %d bytes\n", size) debug.Log("Repo.findPacker", "searching packer for %d bytes\n", size)
for i, p := range s.packs { for i, p := range s.packs {
if p.Size()+size < maxPackSize { if p.Size()+size < maxPackSize {
debug.Log("Repository.findPacker", "found packer %v", p) debug.Log("Repo.findPacker", "found packer %v", p)
// remove from list // remove from list
s.packs = append(s.packs[:i], s.packs[i+1:]...) s.packs = append(s.packs[:i], s.packs[i+1:]...)
return p, nil return p, nil
@ -237,22 +237,22 @@ func (s *Repository) findPacker(size uint) (*pack.Packer, error) {
if err != nil { if err != nil {
return nil, err return nil, err
} }
debug.Log("Repository.findPacker", "create new pack %p", blob) debug.Log("Repo.findPacker", "create new pack %p", blob)
return pack.NewPacker(s.key, blob), nil return pack.NewPacker(s.key, blob), nil
} }
// insertPacker appends p to s.packs. // insertPacker appends p to s.packs.
func (s *Repository) insertPacker(p *pack.Packer) { func (s *Repo) insertPacker(p *pack.Packer) {
s.pm.Lock() s.pm.Lock()
defer s.pm.Unlock() defer s.pm.Unlock()
s.packs = append(s.packs, p) s.packs = append(s.packs, p)
debug.Log("Repository.insertPacker", "%d packers\n", len(s.packs)) debug.Log("Repo.insertPacker", "%d packers\n", len(s.packs))
} }
// savePacker stores p in the backend. // savePacker stores p in the backend.
func (s *Repository) savePacker(p *pack.Packer) error { func (s *Repo) savePacker(p *pack.Packer) error {
debug.Log("Repository.savePacker", "save packer with %d blobs\n", p.Count()) debug.Log("Repo.savePacker", "save packer with %d blobs\n", p.Count())
_, err := p.Finalize() _, err := p.Finalize()
if err != nil { if err != nil {
return err return err
@ -262,15 +262,15 @@ func (s *Repository) savePacker(p *pack.Packer) error {
sid := p.ID() sid := p.ID()
err = p.Writer().(backend.Blob).Finalize(backend.Data, sid.String()) err = p.Writer().(backend.Blob).Finalize(backend.Data, sid.String())
if err != nil { if err != nil {
debug.Log("Repository.savePacker", "blob Finalize() error: %v", err) debug.Log("Repo.savePacker", "blob Finalize() error: %v", err)
return err return err
} }
debug.Log("Repository.savePacker", "saved as %v", sid.Str()) debug.Log("Repo.savePacker", "saved as %v", sid.Str())
// update blobs in the index // update blobs in the index
for _, b := range p.Blobs() { for _, b := range p.Blobs() {
debug.Log("Repository.savePacker", " updating blob %v to pack %v", b.ID.Str(), sid.Str()) debug.Log("Repo.savePacker", " updating blob %v to pack %v", b.ID.Str(), sid.Str())
s.idx.Store(b.Type, b.ID, sid, b.Offset, uint(b.Length)) s.idx.Store(b.Type, b.ID, sid, b.Offset, uint(b.Length))
} }
@ -278,7 +278,7 @@ func (s *Repository) savePacker(p *pack.Packer) error {
} }
// countPacker returns the number of open (unfinished) packers. // countPacker returns the number of open (unfinished) packers.
func (s *Repository) countPacker() int { func (s *Repo) countPacker() int {
s.pm.Lock() s.pm.Lock()
defer s.pm.Unlock() defer s.pm.Unlock()
@ -287,13 +287,13 @@ func (s *Repository) countPacker() int {
// Save encrypts data and stores it to the backend as type t. If data is small // Save encrypts data and stores it to the backend as type t. If data is small
// enough, it will be packed together with other small blobs. // enough, it will be packed together with other small blobs.
func (s *Repository) Save(t pack.BlobType, data []byte, id backend.ID) (backend.ID, error) { func (s *Repo) Save(t pack.BlobType, data []byte, id backend.ID) (backend.ID, error) {
if id == nil { if id == nil {
// compute plaintext hash // compute plaintext hash
id = backend.Hash(data) id = backend.Hash(data)
} }
debug.Log("Repository.Save", "save id %v (%v, %d bytes)", id.Str(), t, len(data)) debug.Log("Repo.Save", "save id %v (%v, %d bytes)", id.Str(), t, len(data))
// get buf from the pool // get buf from the pool
ciphertext := getBuf() ciphertext := getBuf()
@ -317,12 +317,12 @@ func (s *Repository) Save(t pack.BlobType, data []byte, id backend.ID) (backend.
// add this id to the index, although we don't know yet in which pack it // add this id to the index, although we don't know yet in which pack it
// will be saved, the entry will be updated when the pack is written. // will be saved, the entry will be updated when the pack is written.
s.idx.Store(t, id, nil, 0, 0) s.idx.Store(t, id, nil, 0, 0)
debug.Log("Repository.Save", "saving stub for %v (%v) in index", id.Str, t) debug.Log("Repo.Save", "saving stub for %v (%v) in index", id.Str, t)
// if the pack is not full enough and there are less than maxPackers // if the pack is not full enough and there are less than maxPackers
// packers, put back to the list // packers, put back to the list
if packer.Size() < minPackSize && s.countPacker() < maxPackers { if packer.Size() < minPackSize && s.countPacker() < maxPackers {
debug.Log("Repository.Save", "pack is not full enough (%d bytes)", packer.Size()) debug.Log("Repo.Save", "pack is not full enough (%d bytes)", packer.Size())
s.insertPacker(packer) s.insertPacker(packer)
return id, nil return id, nil
} }
@ -332,8 +332,8 @@ func (s *Repository) Save(t pack.BlobType, data []byte, id backend.ID) (backend.
} }
// SaveFrom encrypts data read from rd and stores it in a pack in the backend as type t. // SaveFrom encrypts data read from rd and stores it in a pack in the backend as type t.
func (s *Repository) SaveFrom(t pack.BlobType, id backend.ID, length uint, rd io.Reader) error { func (s *Repo) SaveFrom(t pack.BlobType, id backend.ID, length uint, rd io.Reader) error {
debug.Log("Repository.SaveFrom", "save id %v (%v, %d bytes)", id.Str(), t, length) debug.Log("Repo.SaveFrom", "save id %v (%v, %d bytes)", id.Str(), t, length)
if id == nil { if id == nil {
return errors.New("id is nil") return errors.New("id is nil")
} }
@ -353,8 +353,8 @@ func (s *Repository) SaveFrom(t pack.BlobType, id backend.ID, length uint, rd io
// SaveJSON serialises item as JSON and encrypts and saves it in a pack in the // SaveJSON serialises item as JSON and encrypts and saves it in a pack in the
// backend as type t. // backend as type t.
func (s *Repository) SaveJSON(t pack.BlobType, item interface{}) (backend.ID, error) { func (s *Repo) SaveJSON(t pack.BlobType, item interface{}) (backend.ID, error) {
debug.Log("Repository.SaveJSON", "save %v blob", t) debug.Log("Repo.SaveJSON", "save %v blob", t)
buf := getBuf()[:0] buf := getBuf()[:0]
defer freeBuf(buf) defer freeBuf(buf)
@ -372,13 +372,13 @@ func (s *Repository) SaveJSON(t pack.BlobType, item interface{}) (backend.ID, er
// SaveJSONUnpacked serialises item as JSON and encrypts and saves it in the // SaveJSONUnpacked serialises item as JSON and encrypts and saves it in the
// backend as type t, without a pack. It returns the storage hash. // backend as type t, without a pack. It returns the storage hash.
func (s *Repository) SaveJSONUnpacked(t backend.Type, item interface{}) (backend.ID, error) { func (s *Repo) SaveJSONUnpacked(t backend.Type, item interface{}) (backend.ID, error) {
// create file // create file
blob, err := s.be.Create() blob, err := s.be.Create()
if err != nil { if err != nil {
return nil, err return nil, err
} }
debug.Log("Repository.SaveJSONUnpacked", "create new file %p", blob) debug.Log("Repo.SaveJSONUnpacked", "create new file %p", blob)
// hash // hash
hw := backend.NewHashingWriter(blob, sha256.New()) hw := backend.NewHashingWriter(blob, sha256.New())
@ -409,11 +409,11 @@ func (s *Repository) SaveJSONUnpacked(t backend.Type, item interface{}) (backend
} }
// Flush saves all remaining packs. // Flush saves all remaining packs.
func (s *Repository) Flush() error { func (s *Repo) Flush() error {
s.pm.Lock() s.pm.Lock()
defer s.pm.Unlock() defer s.pm.Unlock()
debug.Log("Repository.Flush", "manually flushing %d packs", len(s.packs)) debug.Log("Repo.Flush", "manually flushing %d packs", len(s.packs))
for _, p := range s.packs { for _, p := range s.packs {
err := s.savePacker(p) err := s.savePacker(p)
@ -426,23 +426,23 @@ func (s *Repository) Flush() error {
return nil return nil
} }
func (s *Repository) Backend() backend.Backend { func (s *Repo) Backend() backend.Backend {
return s.be return s.be
} }
func (s *Repository) Index() *Index { func (s *Repo) Index() *Index {
return s.idx return s.idx
} }
// SetIndex instructs the repository to use the given index. // SetIndex instructs the repository to use the given index.
func (s *Repository) SetIndex(i *Index) { func (s *Repo) SetIndex(i *Index) {
s.idx = i s.idx = i
} }
// SaveIndex saves all new packs in the index in the backend, returned is the // SaveIndex saves all new packs in the index in the backend, returned is the
// storage ID. // storage ID.
func (s *Repository) SaveIndex() (backend.ID, error) { func (s *Repo) SaveIndex() (backend.ID, error) {
debug.Log("Repository.SaveIndex", "Saving index") debug.Log("Repo.SaveIndex", "Saving index")
// create blob // create blob
blob, err := s.be.Create() blob, err := s.be.Create()
@ -450,7 +450,7 @@ func (s *Repository) SaveIndex() (backend.ID, error) {
return nil, err return nil, err
} }
debug.Log("Repository.SaveIndex", "create new pack %p", blob) debug.Log("Repo.SaveIndex", "create new pack %p", blob)
// hash // hash
hw := backend.NewHashingWriter(blob, sha256.New()) hw := backend.NewHashingWriter(blob, sha256.New())
@ -476,15 +476,15 @@ func (s *Repository) SaveIndex() (backend.ID, error) {
return nil, err return nil, err
} }
debug.Log("Repository.SaveIndex", "Saved index as %v", sid.Str()) debug.Log("Repo.SaveIndex", "Saved index as %v", sid.Str())
return sid, nil return sid, nil
} }
// LoadIndex loads all index files from the backend and merges them with the // LoadIndex loads all index files from the backend and merges them with the
// current index. // current index.
func (s *Repository) LoadIndex() error { func (s *Repo) LoadIndex() error {
debug.Log("Repository.LoadIndex", "Loading index") debug.Log("Repo.LoadIndex", "Loading index")
done := make(chan struct{}) done := make(chan struct{})
defer close(done) defer close(done)
@ -498,8 +498,8 @@ func (s *Repository) LoadIndex() error {
} }
// loadIndex loads the index id and merges it with the currently used index. // loadIndex loads the index id and merges it with the currently used index.
func (s *Repository) loadIndex(id string) error { func (s *Repo) loadIndex(id string) error {
debug.Log("Repository.loadIndex", "Loading index %v", id[:8]) debug.Log("Repo.loadIndex", "Loading index %v", id[:8])
before := len(s.idx.pack) before := len(s.idx.pack)
rd, err := s.be.Get(backend.Index, id) rd, err := s.be.Get(backend.Index, id)
@ -517,22 +517,22 @@ func (s *Repository) loadIndex(id string) error {
idx, err := DecodeIndex(decryptRd) idx, err := DecodeIndex(decryptRd)
if err != nil { if err != nil {
debug.Log("Repository.loadIndex", "error while decoding index %v: %v", id, err) debug.Log("Repo.loadIndex", "error while decoding index %v: %v", id, err)
return err return err
} }
s.idx.Merge(idx) s.idx.Merge(idx)
after := len(s.idx.pack) after := len(s.idx.pack)
debug.Log("Repository.loadIndex", "Loaded index %v, added %v blobs", id[:8], after-before) debug.Log("Repo.loadIndex", "Loaded index %v, added %v blobs", id[:8], after-before)
return nil return nil
} }
const repositoryIDSize = sha256.Size const repositoryIDSize = sha256.Size
const RepositoryVersion = 1 const RepoVersion = 1
func createConfig(s *Repository) (err error) { func createConfig(s *Repo) (err error) {
s.Config.ChunkerPolynomial, err = chunker.RandomPolynomial() s.Config.ChunkerPolynomial, err = chunker.RandomPolynomial()
if err != nil { if err != nil {
return err return err
@ -545,21 +545,21 @@ func createConfig(s *Repository) (err error) {
} }
s.Config.ID = hex.EncodeToString(newID) s.Config.ID = hex.EncodeToString(newID)
s.Config.Version = RepositoryVersion s.Config.Version = RepoVersion
debug.Log("Repository.createConfig", "New config: %#v", s.Config) debug.Log("Repo.createConfig", "New config: %#v", s.Config)
_, err = s.SaveJSONUnpacked(backend.Config, s.Config) _, err = s.SaveJSONUnpacked(backend.Config, s.Config)
return err return err
} }
func (s *Repository) loadConfig(cfg *Config) error { func (s *Repo) loadConfig(cfg *Config) error {
err := s.LoadJSONUnpacked(backend.Config, nil, cfg) err := s.LoadJSONUnpacked(backend.Config, nil, cfg)
if err != nil { if err != nil {
return err return err
} }
if cfg.Version != RepositoryVersion { if cfg.Version != RepoVersion {
return errors.New("unsupported repository version") return errors.New("unsupported repository version")
} }
@ -572,7 +572,7 @@ func (s *Repository) loadConfig(cfg *Config) error {
// SearchKey finds a key with the supplied password, afterwards the config is // SearchKey finds a key with the supplied password, afterwards the config is
// read and parsed. // read and parsed.
func (s *Repository) SearchKey(password string) error { func (s *Repo) SearchKey(password string) error {
key, err := SearchKey(s, password) key, err := SearchKey(s, password)
if err != nil { if err != nil {
return err return err
@ -585,7 +585,7 @@ func (s *Repository) SearchKey(password string) error {
// Init creates a new master key with the supplied password and initializes the // Init creates a new master key with the supplied password and initializes the
// repository config. // repository config.
func (s *Repository) Init(password string) error { func (s *Repo) Init(password string) error {
has, err := s.Test(backend.Config, "") has, err := s.Test(backend.Config, "")
if err != nil { if err != nil {
return err return err
@ -604,7 +604,7 @@ func (s *Repository) Init(password string) error {
return createConfig(s) return createConfig(s)
} }
func (s *Repository) Decrypt(ciphertext []byte) ([]byte, error) { func (s *Repo) Decrypt(ciphertext []byte) ([]byte, error) {
if s.key == nil { if s.key == nil {
return nil, errors.New("key for repository not set") return nil, errors.New("key for repository not set")
} }
@ -612,7 +612,7 @@ func (s *Repository) Decrypt(ciphertext []byte) ([]byte, error) {
return crypto.Decrypt(s.key, nil, ciphertext) return crypto.Decrypt(s.key, nil, ciphertext)
} }
func (s *Repository) Encrypt(ciphertext, plaintext []byte) ([]byte, error) { func (s *Repo) Encrypt(ciphertext, plaintext []byte) ([]byte, error) {
if s.key == nil { if s.key == nil {
return nil, errors.New("key for repository not set") return nil, errors.New("key for repository not set")
} }
@ -620,16 +620,16 @@ func (s *Repository) Encrypt(ciphertext, plaintext []byte) ([]byte, error) {
return crypto.Encrypt(s.key, ciphertext, plaintext) return crypto.Encrypt(s.key, ciphertext, plaintext)
} }
func (s *Repository) Key() *crypto.Key { func (s *Repo) Key() *crypto.Key {
return s.key return s.key
} }
func (s *Repository) KeyName() string { func (s *Repo) KeyName() string {
return s.keyName return s.keyName
} }
// Count returns the number of blobs of a given type in the backend. // Count returns the number of blobs of a given type in the backend.
func (s *Repository) Count(t backend.Type) (n uint) { func (s *Repo) Count(t backend.Type) (n uint) {
for _ = range s.be.List(t, nil) { for _ = range s.be.List(t, nil) {
n++ n++
} }
@ -639,27 +639,27 @@ func (s *Repository) Count(t backend.Type) (n uint) {
// Proxy methods to backend // Proxy methods to backend
func (s *Repository) Get(t backend.Type, name string) (io.ReadCloser, error) { func (s *Repo) Get(t backend.Type, name string) (io.ReadCloser, error) {
return s.be.Get(t, name) return s.be.Get(t, name)
} }
func (s *Repository) List(t backend.Type, done <-chan struct{}) <-chan string { func (s *Repo) List(t backend.Type, done <-chan struct{}) <-chan string {
return s.be.List(t, done) return s.be.List(t, done)
} }
func (s *Repository) Test(t backend.Type, name string) (bool, error) { func (s *Repo) Test(t backend.Type, name string) (bool, error) {
return s.be.Test(t, name) return s.be.Test(t, name)
} }
func (s *Repository) Remove(t backend.Type, name string) error { func (s *Repo) Remove(t backend.Type, name string) error {
return s.be.Remove(t, name) return s.be.Remove(t, name)
} }
func (s *Repository) Close() error { func (s *Repo) Close() error {
return s.be.Close() return s.be.Close()
} }
func (s *Repository) Delete() error { func (s *Repo) Delete() error {
if b, ok := s.be.(backend.Deleter); ok { if b, ok := s.be.(backend.Deleter); ok {
return b.Delete() return b.Delete()
} }
@ -667,6 +667,6 @@ func (s *Repository) Delete() error {
return errors.New("Delete() called for backend that does not implement this method") return errors.New("Delete() called for backend that does not implement this method")
} }
func (s *Repository) Location() string { func (s *Repo) Location() string {
return s.be.Location() return s.be.Location()
} }

View file

@ -14,7 +14,7 @@ import (
// Restorer is used to restore a snapshot to a directory. // Restorer is used to restore a snapshot to a directory.
type Restorer struct { type Restorer struct {
repo *repo.Repository repo *repo.Repo
sn *Snapshot sn *Snapshot
Error func(dir string, node *Node, err error) error Error func(dir string, node *Node, err error) error
@ -24,7 +24,7 @@ type Restorer struct {
var restorerAbortOnAllErrors = func(str string, node *Node, err error) error { return err } var restorerAbortOnAllErrors = func(str string, node *Node, err error) error { return err }
// NewRestorer creates a restorer preloaded with the content from the snapshot id. // NewRestorer creates a restorer preloaded with the content from the snapshot id.
func NewRestorer(repo *repo.Repository, id backend.ID) (*Restorer, error) { func NewRestorer(repo *repo.Repo, id backend.ID) (*Restorer, error) {
r := &Restorer{repo: repo, Error: restorerAbortOnAllErrors} r := &Restorer{repo: repo, Error: restorerAbortOnAllErrors}
var err error var err error

View file

@ -50,7 +50,7 @@ func NewSnapshot(paths []string) (*Snapshot, error) {
return sn, nil return sn, nil
} }
func LoadSnapshot(repo *repo.Repository, id backend.ID) (*Snapshot, error) { func LoadSnapshot(repo *repo.Repo, id backend.ID) (*Snapshot, error) {
sn := &Snapshot{id: id} sn := &Snapshot{id: id}
err := repo.LoadJSONUnpacked(backend.Snapshot, id, sn) err := repo.LoadJSONUnpacked(backend.Snapshot, id, sn)
if err != nil { if err != nil {

View file

@ -17,7 +17,7 @@ var TestPassword = flag.String("test.password", "", `use this password for repos
var TestCleanup = flag.Bool("test.cleanup", true, "clean up after running tests (remove local backend directory with all content)") var TestCleanup = flag.Bool("test.cleanup", true, "clean up after running tests (remove local backend directory with all content)")
var TestTempDir = flag.String("test.tempdir", "", "use this directory for temporary storage (default: system temp dir)") var TestTempDir = flag.String("test.tempdir", "", "use this directory for temporary storage (default: system temp dir)")
func SetupRepo(t testing.TB) *repo.Repository { func SetupRepo(t testing.TB) *repo.Repo {
tempdir, err := ioutil.TempDir(*TestTempDir, "restic-test-") tempdir, err := ioutil.TempDir(*TestTempDir, "restic-test-")
OK(t, err) OK(t, err)
@ -34,7 +34,7 @@ func SetupRepo(t testing.TB) *repo.Repository {
return repo return repo
} }
func TeardownRepo(t testing.TB, repo *repo.Repository) { func TeardownRepo(t testing.TB, repo *repo.Repo) {
if !*TestCleanup { if !*TestCleanup {
l := repo.Backend().(*local.Local) l := repo.Backend().(*local.Local)
t.Logf("leaving local backend at %s\n", l.Location()) t.Logf("leaving local backend at %s\n", l.Location())
@ -44,7 +44,7 @@ func TeardownRepo(t testing.TB, repo *repo.Repository) {
OK(t, repo.Delete()) OK(t, repo.Delete())
} }
func SnapshotDir(t testing.TB, repo *repo.Repository, path string, parent backend.ID) *restic.Snapshot { func SnapshotDir(t testing.TB, repo *repo.Repo, path string, parent backend.ID) *restic.Snapshot {
arch := restic.NewArchiver(repo) arch := restic.NewArchiver(repo)
sn, _, err := arch.Snapshot(nil, []string{path}, parent) sn, _, err := arch.Snapshot(nil, []string{path}, parent)
OK(t, err) OK(t, err)

View file

@ -30,7 +30,7 @@ func (t Tree) String() string {
return fmt.Sprintf("Tree<%d nodes>", len(t.Nodes)) return fmt.Sprintf("Tree<%d nodes>", len(t.Nodes))
} }
func LoadTree(repo *repo.Repository, id backend.ID) (*Tree, error) { func LoadTree(repo *repo.Repo, id backend.ID) (*Tree, error) {
tree := &Tree{} tree := &Tree{}
err := repo.LoadJSONPack(pack.Tree, id, tree) err := repo.LoadJSONPack(pack.Tree, id, tree)
if err != nil { if err != nil {

View file

@ -16,7 +16,7 @@ type WalkTreeJob struct {
Tree *Tree Tree *Tree
} }
func walkTree(repo *repo.Repository, path string, treeID backend.ID, done chan struct{}, jobCh chan<- WalkTreeJob) { func walkTree(repo *repo.Repo, path string, treeID backend.ID, done chan struct{}, jobCh chan<- WalkTreeJob) {
debug.Log("walkTree", "start on %q (%v)", path, treeID.Str()) debug.Log("walkTree", "start on %q (%v)", path, treeID.Str())
t, err := LoadTree(repo, treeID) t, err := LoadTree(repo, treeID)
@ -41,7 +41,7 @@ func walkTree(repo *repo.Repository, path string, treeID backend.ID, done chan s
// WalkTree walks the tree specified by id recursively and sends a job for each // WalkTree walks the tree specified by id recursively and sends a job for each
// file and directory it finds. When the channel done is closed, processing // file and directory it finds. When the channel done is closed, processing
// stops. // stops.
func WalkTree(repo *repo.Repository, id backend.ID, done chan struct{}, jobCh chan<- WalkTreeJob) { func WalkTree(repo *repo.Repo, id backend.ID, done chan struct{}, jobCh chan<- WalkTreeJob) {
debug.Log("WalkTree", "start on %v", id.Str()) debug.Log("WalkTree", "start on %v", id.Str())
walkTree(repo, "", id, done, jobCh) walkTree(repo, "", id, done, jobCh)
close(jobCh) close(jobCh)