2016-07-04 12:12:33 +00:00
|
|
|
// Implementation of sync/copy/move
|
|
|
|
|
|
|
|
package fs
|
|
|
|
|
|
|
|
import (
|
2017-01-03 17:35:12 +00:00
|
|
|
"fmt"
|
2017-08-09 20:06:39 +00:00
|
|
|
"sort"
|
2016-07-04 12:12:33 +00:00
|
|
|
"sync"
|
|
|
|
"time"
|
2016-11-05 18:03:55 +00:00
|
|
|
|
|
|
|
"github.com/pkg/errors"
|
2016-07-04 12:12:33 +00:00
|
|
|
)
|
|
|
|
|
2017-06-13 13:35:51 +00:00
|
|
|
var oldSyncMethod = BoolP("old-sync-method", "", false, "Deprecated - use --fast-list instead")
|
2017-01-25 19:59:53 +00:00
|
|
|
|
2016-07-04 12:12:33 +00:00
|
|
|
type syncCopyMove struct {
|
|
|
|
// parameters
|
2017-01-25 19:35:14 +00:00
|
|
|
fdst Fs
|
|
|
|
fsrc Fs
|
|
|
|
deleteMode DeleteMode // how we are doing deletions
|
|
|
|
DoMove bool
|
|
|
|
dir string
|
2016-07-04 12:12:33 +00:00
|
|
|
// internal state
|
2017-01-03 17:35:12 +00:00
|
|
|
noTraverse bool // if set don't trafevers the dst
|
2017-01-25 19:59:53 +00:00
|
|
|
deletersWg sync.WaitGroup // for delete before go routine
|
|
|
|
deleteFilesCh chan Object // channel to receive deletes if delete before
|
2017-01-03 17:35:12 +00:00
|
|
|
trackRenames bool // set if we should do server side renames
|
|
|
|
dstFilesMu sync.Mutex // protect dstFiles
|
|
|
|
dstFiles map[string]Object // dst files, always filled
|
|
|
|
srcFiles map[string]Object // src files, only used if deleteBefore
|
|
|
|
srcFilesChan chan Object // passes src objects
|
|
|
|
srcFilesResult chan error // error result of src listing
|
|
|
|
dstFilesResult chan error // error result of dst listing
|
2017-08-09 20:06:39 +00:00
|
|
|
dstEmptyDirsMu sync.Mutex // protect dstEmptyDirs
|
|
|
|
dstEmptyDirs []DirEntry // potentially empty directories
|
2017-01-03 17:35:12 +00:00
|
|
|
abort chan struct{} // signal to abort the copiers
|
|
|
|
checkerWg sync.WaitGroup // wait for checkers
|
|
|
|
toBeChecked ObjectPairChan // checkers channel
|
|
|
|
transfersWg sync.WaitGroup // wait for transfers
|
|
|
|
toBeUploaded ObjectPairChan // copiers channel
|
|
|
|
errorMu sync.Mutex // Mutex covering the errors variables
|
|
|
|
err error // normal error from copy process
|
|
|
|
noRetryErr error // error with NoRetry set
|
|
|
|
fatalErr error // fatal error
|
|
|
|
commonHash HashType // common hash type between src and dst
|
|
|
|
renameMapMu sync.Mutex // mutex to protect the below
|
|
|
|
renameMap map[string][]Object // dst files by hash - only used by trackRenames
|
|
|
|
renamerWg sync.WaitGroup // wait for renamers
|
|
|
|
toBeRenamed ObjectPairChan // renamers channel
|
2017-01-24 11:04:09 +00:00
|
|
|
trackRenamesWg sync.WaitGroup // wg for background track renames
|
|
|
|
trackRenamesCh chan Object // objects are pumped in here
|
|
|
|
renameCheck []Object // accumulate files to check for rename here
|
2017-01-10 21:47:03 +00:00
|
|
|
backupDir Fs // place to store overwrites/deletes
|
2017-01-19 17:26:29 +00:00
|
|
|
suffix string // suffix to add to files placed in backupDir
|
2017-06-06 22:04:01 +00:00
|
|
|
srcListDir listDirFn // function to call to list a directory in the src
|
|
|
|
dstListDir listDirFn // function to call to list a directory in the dst
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
|
|
|
|
2017-01-25 19:35:14 +00:00
|
|
|
func newSyncCopyMove(fdst, fsrc Fs, deleteMode DeleteMode, DoMove bool) (*syncCopyMove, error) {
|
2016-07-04 12:12:33 +00:00
|
|
|
s := &syncCopyMove{
|
|
|
|
fdst: fdst,
|
|
|
|
fsrc: fsrc,
|
2017-01-25 19:35:14 +00:00
|
|
|
deleteMode: deleteMode,
|
2016-07-04 12:12:33 +00:00
|
|
|
DoMove: DoMove,
|
|
|
|
dir: "",
|
|
|
|
srcFilesChan: make(chan Object, Config.Checkers+Config.Transfers),
|
|
|
|
srcFilesResult: make(chan error, 1),
|
|
|
|
dstFilesResult: make(chan error, 1),
|
|
|
|
noTraverse: Config.NoTraverse,
|
|
|
|
abort: make(chan struct{}),
|
|
|
|
toBeChecked: make(ObjectPairChan, Config.Transfers),
|
|
|
|
toBeUploaded: make(ObjectPairChan, Config.Transfers),
|
2017-01-25 19:59:53 +00:00
|
|
|
deleteFilesCh: make(chan Object, Config.Checkers),
|
2017-01-02 16:37:59 +00:00
|
|
|
trackRenames: Config.TrackRenames,
|
2017-01-03 17:35:12 +00:00
|
|
|
commonHash: fsrc.Hashes().Overlap(fdst.Hashes()).GetOne(),
|
|
|
|
toBeRenamed: make(ObjectPairChan, Config.Transfers),
|
2017-01-24 11:04:09 +00:00
|
|
|
trackRenamesCh: make(chan Object, Config.Checkers),
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
2017-01-25 19:35:14 +00:00
|
|
|
if s.noTraverse && s.deleteMode != DeleteModeOff {
|
2017-02-09 11:01:20 +00:00
|
|
|
Errorf(nil, "Ignoring --no-traverse with sync")
|
2016-07-04 12:12:33 +00:00
|
|
|
s.noTraverse = false
|
|
|
|
}
|
2017-01-02 16:37:59 +00:00
|
|
|
if s.trackRenames {
|
2017-01-10 20:03:55 +00:00
|
|
|
// Don't track renames for remotes without server-side move support.
|
2017-01-10 21:47:03 +00:00
|
|
|
if !CanServerSideMove(fdst) {
|
2017-02-09 11:01:20 +00:00
|
|
|
Errorf(fdst, "Ignoring --track-renames as the destination does not support server-side move or copy")
|
2017-01-02 16:37:59 +00:00
|
|
|
s.trackRenames = false
|
|
|
|
}
|
2017-01-03 17:35:12 +00:00
|
|
|
if s.commonHash == HashNone {
|
2017-02-09 11:01:20 +00:00
|
|
|
Errorf(fdst, "Ignoring --track-renames as the source and destination do not have a common hash")
|
2017-01-02 16:37:59 +00:00
|
|
|
s.trackRenames = false
|
|
|
|
}
|
|
|
|
}
|
2017-01-25 19:35:14 +00:00
|
|
|
if s.trackRenames {
|
|
|
|
// track renames needs delete after
|
|
|
|
if s.deleteMode != DeleteModeOff {
|
|
|
|
s.deleteMode = DeleteModeAfter
|
|
|
|
}
|
|
|
|
if s.noTraverse {
|
2017-02-09 11:01:20 +00:00
|
|
|
Errorf(nil, "Ignoring --no-traverse with --track-renames")
|
2017-01-25 19:35:14 +00:00
|
|
|
s.noTraverse = false
|
|
|
|
}
|
2017-01-03 17:35:12 +00:00
|
|
|
}
|
2017-01-10 21:47:03 +00:00
|
|
|
// Make Fs for --backup-dir if required
|
|
|
|
if Config.BackupDir != "" {
|
|
|
|
var err error
|
|
|
|
s.backupDir, err = NewFs(Config.BackupDir)
|
|
|
|
if err != nil {
|
|
|
|
return nil, FatalError(errors.Errorf("Failed to make fs for --backup-dir %q: %v", Config.BackupDir, err))
|
|
|
|
}
|
|
|
|
if !CanServerSideMove(s.backupDir) {
|
|
|
|
return nil, FatalError(errors.New("can't use --backup-dir on a remote which doesn't support server side move or copy"))
|
|
|
|
}
|
|
|
|
if !SameConfig(fdst, s.backupDir) {
|
|
|
|
return nil, FatalError(errors.New("parameter to --backup-dir has to be on the same remote as destination"))
|
|
|
|
}
|
|
|
|
if Overlapping(fdst, s.backupDir) {
|
|
|
|
return nil, FatalError(errors.New("destination and parameter to --backup-dir mustn't overlap"))
|
|
|
|
}
|
|
|
|
if Overlapping(fsrc, s.backupDir) {
|
|
|
|
return nil, FatalError(errors.New("source and parameter to --backup-dir mustn't overlap"))
|
|
|
|
}
|
2017-01-19 17:26:29 +00:00
|
|
|
s.suffix = Config.Suffix
|
2017-01-10 21:47:03 +00:00
|
|
|
}
|
2017-06-06 22:04:01 +00:00
|
|
|
s.srcListDir = s.makeListDir(fsrc, false)
|
|
|
|
s.dstListDir = s.makeListDir(fdst, Config.Filter.DeleteExcluded)
|
2017-01-10 21:47:03 +00:00
|
|
|
return s, nil
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
|
|
|
|
2017-06-06 22:04:01 +00:00
|
|
|
// list a directory into entries, err
|
|
|
|
type listDirFn func(dir string) (entries DirEntries, err error)
|
|
|
|
|
|
|
|
// makeListDir makes a listing function for the given fs and includeAll flags
|
|
|
|
func (s *syncCopyMove) makeListDir(f Fs, includeAll bool) listDirFn {
|
|
|
|
if !Config.UseListR || f.Features().ListR == nil {
|
|
|
|
return func(dir string) (entries DirEntries, err error) {
|
|
|
|
return ListDirSorted(f, includeAll, dir)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
var (
|
|
|
|
mu sync.Mutex
|
|
|
|
started bool
|
|
|
|
dirs DirTree
|
|
|
|
dirsErr error
|
|
|
|
)
|
|
|
|
return func(dir string) (entries DirEntries, err error) {
|
|
|
|
mu.Lock()
|
|
|
|
defer mu.Unlock()
|
|
|
|
if !started {
|
|
|
|
dirs, dirsErr = NewDirTree(f, s.dir, includeAll, Config.MaxDepth)
|
|
|
|
started = true
|
|
|
|
}
|
|
|
|
if dirsErr != nil {
|
|
|
|
return nil, dirsErr
|
|
|
|
}
|
|
|
|
entries, ok := dirs[dir]
|
|
|
|
if !ok {
|
|
|
|
err = ErrorDirNotFound
|
|
|
|
} else {
|
|
|
|
delete(dirs, dir)
|
|
|
|
}
|
|
|
|
return entries, err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-07-04 12:12:33 +00:00
|
|
|
// Check to see if have set the abort flag
|
|
|
|
func (s *syncCopyMove) aborting() bool {
|
|
|
|
select {
|
|
|
|
case <-s.abort:
|
|
|
|
return true
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
2017-01-03 17:35:12 +00:00
|
|
|
// This reads the map and pumps it into the channel passed in, closing
|
|
|
|
// the channel at the end
|
|
|
|
func (s *syncCopyMove) pumpMapToChan(files map[string]Object, out chan<- Object) {
|
2016-07-04 12:12:33 +00:00
|
|
|
outer:
|
2017-01-03 17:35:12 +00:00
|
|
|
for _, o := range files {
|
2016-07-04 12:12:33 +00:00
|
|
|
if s.aborting() {
|
|
|
|
break outer
|
|
|
|
}
|
|
|
|
select {
|
2017-01-03 17:35:12 +00:00
|
|
|
case out <- o:
|
2016-07-04 12:12:33 +00:00
|
|
|
case <-s.abort:
|
|
|
|
break outer
|
|
|
|
}
|
|
|
|
}
|
2017-01-03 17:35:12 +00:00
|
|
|
close(out)
|
|
|
|
s.srcFilesResult <- nil
|
|
|
|
}
|
|
|
|
|
2016-10-23 16:34:17 +00:00
|
|
|
// NeedTransfer checks to see if src needs to be copied to dst using
|
|
|
|
// the current config.
|
2016-10-03 18:58:44 +00:00
|
|
|
//
|
2016-10-23 16:34:17 +00:00
|
|
|
// Returns a flag which indicates whether the file needs to be
|
|
|
|
// transferred or not.
|
|
|
|
func NeedTransfer(dst, src Object) bool {
|
2016-07-04 12:12:33 +00:00
|
|
|
if dst == nil {
|
2017-02-09 11:01:20 +00:00
|
|
|
Debugf(src, "Couldn't find file - need to transfer")
|
2016-10-03 18:58:44 +00:00
|
|
|
return true
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
|
|
|
// If we should ignore existing files, don't transfer
|
|
|
|
if Config.IgnoreExisting {
|
2017-02-09 11:01:20 +00:00
|
|
|
Debugf(src, "Destination exists, skipping")
|
2016-10-03 18:58:44 +00:00
|
|
|
return false
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
|
|
|
// If we should upload unconditionally
|
|
|
|
if Config.IgnoreTimes {
|
2017-02-09 11:01:20 +00:00
|
|
|
Debugf(src, "Transferring unconditionally as --ignore-times is in use")
|
2016-10-03 18:58:44 +00:00
|
|
|
return true
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
|
|
|
// If UpdateOlder is in effect, skip if dst is newer than src
|
|
|
|
if Config.UpdateOlder {
|
|
|
|
srcModTime := src.ModTime()
|
|
|
|
dstModTime := dst.ModTime()
|
|
|
|
dt := dstModTime.Sub(srcModTime)
|
|
|
|
// If have a mutually agreed precision then use that
|
|
|
|
modifyWindow := Config.ModifyWindow
|
|
|
|
if modifyWindow == ModTimeNotSupported {
|
|
|
|
// Otherwise use 1 second as a safe default as
|
|
|
|
// the resolution of the time a file was
|
|
|
|
// uploaded.
|
|
|
|
modifyWindow = time.Second
|
|
|
|
}
|
|
|
|
switch {
|
|
|
|
case dt >= modifyWindow:
|
2017-02-09 11:01:20 +00:00
|
|
|
Debugf(src, "Destination is newer than source, skipping")
|
2016-10-03 18:58:44 +00:00
|
|
|
return false
|
2016-07-04 12:12:33 +00:00
|
|
|
case dt <= -modifyWindow:
|
2017-02-09 11:01:20 +00:00
|
|
|
Debugf(src, "Destination is older than source, transferring")
|
2016-07-04 12:12:33 +00:00
|
|
|
default:
|
|
|
|
if src.Size() == dst.Size() {
|
2017-02-09 11:01:20 +00:00
|
|
|
Debugf(src, "Destination mod time is within %v of source and sizes identical, skipping", modifyWindow)
|
2016-10-03 18:58:44 +00:00
|
|
|
return false
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
2017-02-09 11:01:20 +00:00
|
|
|
Debugf(src, "Destination mod time is within %v of source but sizes differ, transferring", modifyWindow)
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// Check to see if changed or not
|
|
|
|
if Equal(src, dst) {
|
2017-02-09 11:01:20 +00:00
|
|
|
Debugf(src, "Unchanged skipping")
|
2016-10-03 18:58:44 +00:00
|
|
|
return false
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
|
|
|
}
|
2016-10-03 18:58:44 +00:00
|
|
|
return true
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// This checks the types of errors returned while copying files
|
|
|
|
func (s *syncCopyMove) processError(err error) {
|
|
|
|
if err == nil {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
s.errorMu.Lock()
|
|
|
|
defer s.errorMu.Unlock()
|
|
|
|
switch {
|
|
|
|
case IsFatalError(err):
|
2016-08-18 17:56:57 +00:00
|
|
|
if !s.aborting() {
|
|
|
|
close(s.abort)
|
|
|
|
}
|
2016-07-04 12:12:33 +00:00
|
|
|
s.fatalErr = err
|
|
|
|
case IsNoRetryError(err):
|
|
|
|
s.noRetryErr = err
|
|
|
|
default:
|
|
|
|
s.err = err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-01-03 17:35:12 +00:00
|
|
|
// Returns the current error (if any) in the order of prececedence
|
|
|
|
// fatalErr
|
|
|
|
// normal error
|
|
|
|
// noRetryErr
|
|
|
|
func (s *syncCopyMove) currentError() error {
|
|
|
|
s.errorMu.Lock()
|
|
|
|
defer s.errorMu.Unlock()
|
|
|
|
if s.fatalErr != nil {
|
|
|
|
return s.fatalErr
|
|
|
|
}
|
|
|
|
if s.err != nil {
|
|
|
|
return s.err
|
|
|
|
}
|
|
|
|
return s.noRetryErr
|
|
|
|
}
|
|
|
|
|
2016-07-04 12:12:33 +00:00
|
|
|
// pairChecker reads Objects~s on in send to out if they need transferring.
|
|
|
|
//
|
|
|
|
// FIXME potentially doing lots of hashes at once
|
|
|
|
func (s *syncCopyMove) pairChecker(in ObjectPairChan, out ObjectPairChan, wg *sync.WaitGroup) {
|
|
|
|
defer wg.Done()
|
|
|
|
for {
|
|
|
|
if s.aborting() {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
select {
|
|
|
|
case pair, ok := <-in:
|
|
|
|
if !ok {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
src := pair.src
|
|
|
|
Stats.Checking(src.Remote())
|
2016-10-03 18:58:44 +00:00
|
|
|
// Check to see if can store this
|
|
|
|
if src.Storable() {
|
2016-10-23 16:34:17 +00:00
|
|
|
if NeedTransfer(pair.dst, pair.src) {
|
2017-01-10 21:47:03 +00:00
|
|
|
// If destination already exists, then we must move it into --backup-dir if required
|
|
|
|
if pair.dst != nil && s.backupDir != nil {
|
2017-01-19 17:26:29 +00:00
|
|
|
remoteWithSuffix := pair.dst.Remote() + s.suffix
|
|
|
|
overwritten, _ := s.backupDir.NewObject(remoteWithSuffix)
|
|
|
|
err := Move(s.backupDir, overwritten, remoteWithSuffix, pair.dst)
|
2017-01-10 21:47:03 +00:00
|
|
|
if err != nil {
|
|
|
|
s.processError(err)
|
|
|
|
} else {
|
|
|
|
// If successful zero out the dst as it is no longer there and copy the file
|
|
|
|
pair.dst = nil
|
|
|
|
out <- pair
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
out <- pair
|
|
|
|
}
|
2016-10-03 18:58:44 +00:00
|
|
|
} else {
|
|
|
|
// If moving need to delete the files we don't need to copy
|
|
|
|
if s.DoMove {
|
|
|
|
// Delete src if no error on copy
|
2016-10-03 19:16:41 +00:00
|
|
|
s.processError(DeleteFile(src))
|
2016-10-03 18:58:44 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2016-07-04 12:12:33 +00:00
|
|
|
Stats.DoneChecking(src.Remote())
|
|
|
|
case <-s.abort:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-01-03 17:35:12 +00:00
|
|
|
// pairRenamer reads Objects~s on in and attempts to rename them,
|
|
|
|
// otherwise it sends them out if they need transferring.
|
|
|
|
func (s *syncCopyMove) pairRenamer(in ObjectPairChan, out ObjectPairChan, wg *sync.WaitGroup) {
|
|
|
|
defer wg.Done()
|
|
|
|
for {
|
|
|
|
if s.aborting() {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
select {
|
|
|
|
case pair, ok := <-in:
|
|
|
|
if !ok {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
src := pair.src
|
|
|
|
if !s.tryRename(src) {
|
|
|
|
// pass on if not renamed
|
|
|
|
out <- pair
|
|
|
|
}
|
|
|
|
case <-s.abort:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-10-22 16:53:10 +00:00
|
|
|
// pairCopyOrMove reads Objects on in and moves or copies them.
|
|
|
|
func (s *syncCopyMove) pairCopyOrMove(in ObjectPairChan, fdst Fs, wg *sync.WaitGroup) {
|
2016-07-04 12:12:33 +00:00
|
|
|
defer wg.Done()
|
2016-10-22 16:53:10 +00:00
|
|
|
var err error
|
2016-07-04 12:12:33 +00:00
|
|
|
for {
|
|
|
|
if s.aborting() {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
select {
|
|
|
|
case pair, ok := <-in:
|
|
|
|
if !ok {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
src := pair.src
|
|
|
|
Stats.Transferring(src.Remote())
|
2016-10-22 16:53:10 +00:00
|
|
|
if s.DoMove {
|
|
|
|
err = Move(fdst, pair.dst, src.Remote(), src)
|
2016-07-04 12:12:33 +00:00
|
|
|
} else {
|
2016-10-22 16:53:10 +00:00
|
|
|
err = Copy(fdst, pair.dst, src.Remote(), src)
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
2016-10-22 16:53:10 +00:00
|
|
|
s.processError(err)
|
2016-09-12 17:15:58 +00:00
|
|
|
Stats.DoneTransferring(src.Remote(), err == nil)
|
2016-07-04 12:12:33 +00:00
|
|
|
case <-s.abort:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// This starts the background checkers.
|
|
|
|
func (s *syncCopyMove) startCheckers() {
|
|
|
|
s.checkerWg.Add(Config.Checkers)
|
|
|
|
for i := 0; i < Config.Checkers; i++ {
|
|
|
|
go s.pairChecker(s.toBeChecked, s.toBeUploaded, &s.checkerWg)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// This stops the background checkers
|
|
|
|
func (s *syncCopyMove) stopCheckers() {
|
|
|
|
close(s.toBeChecked)
|
2017-02-09 17:08:51 +00:00
|
|
|
Infof(s.fdst, "Waiting for checks to finish")
|
2016-07-04 12:12:33 +00:00
|
|
|
s.checkerWg.Wait()
|
|
|
|
}
|
|
|
|
|
|
|
|
// This starts the background transfers
|
|
|
|
func (s *syncCopyMove) startTransfers() {
|
2016-10-22 16:53:10 +00:00
|
|
|
s.transfersWg.Add(Config.Transfers)
|
2016-07-04 12:12:33 +00:00
|
|
|
for i := 0; i < Config.Transfers; i++ {
|
2016-10-22 16:53:10 +00:00
|
|
|
go s.pairCopyOrMove(s.toBeUploaded, s.fdst, &s.transfersWg)
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// This stops the background transfers
|
|
|
|
func (s *syncCopyMove) stopTransfers() {
|
|
|
|
close(s.toBeUploaded)
|
2017-02-09 17:08:51 +00:00
|
|
|
Infof(s.fdst, "Waiting for transfers to finish")
|
2016-10-22 16:53:10 +00:00
|
|
|
s.transfersWg.Wait()
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
|
|
|
|
2017-01-03 17:35:12 +00:00
|
|
|
// This starts the background renamers.
|
|
|
|
func (s *syncCopyMove) startRenamers() {
|
|
|
|
if !s.trackRenames {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
s.renamerWg.Add(Config.Checkers)
|
|
|
|
for i := 0; i < Config.Checkers; i++ {
|
|
|
|
go s.pairRenamer(s.toBeRenamed, s.toBeUploaded, &s.renamerWg)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// This stops the background renamers
|
|
|
|
func (s *syncCopyMove) stopRenamers() {
|
|
|
|
if !s.trackRenames {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
close(s.toBeRenamed)
|
2017-02-09 17:08:51 +00:00
|
|
|
Infof(s.fdst, "Waiting for renames to finish")
|
2017-01-03 17:35:12 +00:00
|
|
|
s.renamerWg.Wait()
|
|
|
|
}
|
|
|
|
|
2017-01-24 11:04:09 +00:00
|
|
|
// This starts the collection of possible renames
|
|
|
|
func (s *syncCopyMove) startTrackRenames() {
|
|
|
|
if !s.trackRenames {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
s.trackRenamesWg.Add(1)
|
|
|
|
go func() {
|
|
|
|
defer s.trackRenamesWg.Done()
|
|
|
|
for o := range s.trackRenamesCh {
|
|
|
|
s.renameCheck = append(s.renameCheck, o)
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
|
|
|
|
// This stops the background rename collection
|
|
|
|
func (s *syncCopyMove) stopTrackRenames() {
|
|
|
|
if !s.trackRenames {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
close(s.trackRenamesCh)
|
|
|
|
s.trackRenamesWg.Wait()
|
|
|
|
}
|
|
|
|
|
2017-01-25 19:59:53 +00:00
|
|
|
// This starts the background deletion of files for --delete-during
|
|
|
|
func (s *syncCopyMove) startDeleters() {
|
|
|
|
if s.deleteMode != DeleteModeDuring && s.deleteMode != DeleteModeOnly {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
s.deletersWg.Add(1)
|
|
|
|
go func() {
|
|
|
|
defer s.deletersWg.Done()
|
|
|
|
err := deleteFilesWithBackupDir(s.deleteFilesCh, s.backupDir)
|
|
|
|
s.processError(err)
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
|
|
|
|
// This stops the background deleters
|
|
|
|
func (s *syncCopyMove) stopDeleters() {
|
|
|
|
if s.deleteMode != DeleteModeDuring && s.deleteMode != DeleteModeOnly {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
close(s.deleteFilesCh)
|
|
|
|
s.deletersWg.Wait()
|
|
|
|
}
|
|
|
|
|
2016-07-04 12:12:33 +00:00
|
|
|
// This deletes the files in the dstFiles map. If checkSrcMap is set
|
|
|
|
// then it checks to see if they exist first in srcFiles the source
|
|
|
|
// file map, otherwise it unconditionally deletes them. If
|
|
|
|
// checkSrcMap is clear then it assumes that the any source files that
|
|
|
|
// have been found have been removed from dstFiles already.
|
|
|
|
func (s *syncCopyMove) deleteFiles(checkSrcMap bool) error {
|
|
|
|
if Stats.Errored() {
|
2017-02-09 11:01:20 +00:00
|
|
|
Errorf(s.fdst, "%v", ErrorNotDeleting)
|
2016-07-04 12:12:33 +00:00
|
|
|
return ErrorNotDeleting
|
|
|
|
}
|
|
|
|
|
|
|
|
// Delete the spare files
|
|
|
|
toDelete := make(ObjectsChan, Config.Transfers)
|
|
|
|
go func() {
|
|
|
|
for remote, o := range s.dstFiles {
|
|
|
|
if checkSrcMap {
|
|
|
|
_, exists := s.srcFiles[remote]
|
|
|
|
if exists {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if s.aborting() {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
toDelete <- o
|
|
|
|
}
|
|
|
|
close(toDelete)
|
|
|
|
}()
|
2017-01-10 21:47:03 +00:00
|
|
|
return deleteFilesWithBackupDir(toDelete, s.backupDir)
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
|
|
|
|
2017-08-09 20:06:39 +00:00
|
|
|
// This deletes the empty directories in the slice passed in. It
|
|
|
|
// ignores any errors deleting directories
|
|
|
|
func deleteEmptyDirectories(f Fs, entries DirEntries) error {
|
|
|
|
if len(entries) == 0 {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
if Stats.Errored() {
|
|
|
|
Errorf(f, "%v", ErrorNotDeletingDirs)
|
|
|
|
return ErrorNotDeletingDirs
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now delete the empty directories starting from the longest path
|
|
|
|
sort.Sort(entries)
|
|
|
|
var errorCount int
|
|
|
|
var okCount int
|
|
|
|
for i := len(entries) - 1; i >= 0; i-- {
|
|
|
|
entry := entries[i]
|
|
|
|
dir, ok := entry.(Directory)
|
|
|
|
if ok {
|
|
|
|
// TryRmdir only deletes empty directories
|
|
|
|
err := TryRmdir(f, dir.Remote())
|
|
|
|
if err != nil {
|
|
|
|
Debugf(logDirName(f, dir.Remote()), "Failed to Rmdir: %v", err)
|
|
|
|
errorCount++
|
|
|
|
} else {
|
|
|
|
okCount++
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
Errorf(f, "Not a directory: %v", entry)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if errorCount > 0 {
|
|
|
|
Debugf(f, "failed to delete %d directories", errorCount)
|
|
|
|
}
|
|
|
|
if okCount > 0 {
|
|
|
|
Debugf(f, "deleted %d directories", okCount)
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2017-01-03 17:35:12 +00:00
|
|
|
// renameHash makes a string with the size and the hash for rename detection
|
|
|
|
//
|
|
|
|
// it may return an empty string in which case no hash could be made
|
2017-01-03 23:03:20 +00:00
|
|
|
func (s *syncCopyMove) renameHash(obj Object) (hash string) {
|
|
|
|
var err error
|
2017-01-03 17:35:12 +00:00
|
|
|
hash, err = obj.Hash(s.commonHash)
|
|
|
|
if err != nil {
|
2017-02-09 11:01:20 +00:00
|
|
|
Debugf(obj, "Hash failed: %v", err)
|
2017-01-03 23:03:20 +00:00
|
|
|
return ""
|
2016-12-18 10:03:56 +00:00
|
|
|
}
|
2017-01-03 17:35:12 +00:00
|
|
|
if hash == "" {
|
2017-01-03 23:03:20 +00:00
|
|
|
return ""
|
2017-01-03 17:35:12 +00:00
|
|
|
}
|
2017-01-03 23:03:20 +00:00
|
|
|
return fmt.Sprintf("%d,%s", obj.Size(), hash)
|
|
|
|
}
|
|
|
|
|
|
|
|
// pushRenameMap adds the object with hash to the rename map
|
|
|
|
func (s *syncCopyMove) pushRenameMap(hash string, obj Object) {
|
|
|
|
s.renameMapMu.Lock()
|
|
|
|
s.renameMap[hash] = append(s.renameMap[hash], obj)
|
|
|
|
s.renameMapMu.Unlock()
|
2017-01-03 17:35:12 +00:00
|
|
|
}
|
2016-12-18 10:03:56 +00:00
|
|
|
|
2017-01-03 23:03:20 +00:00
|
|
|
// popRenameMap finds the object with hash and pop the first match from
|
|
|
|
// renameMap or returns nil if not found.
|
|
|
|
func (s *syncCopyMove) popRenameMap(hash string) (dst Object) {
|
|
|
|
s.renameMapMu.Lock()
|
|
|
|
dsts, ok := s.renameMap[hash]
|
|
|
|
if ok && len(dsts) > 0 {
|
|
|
|
dst, dsts = dsts[0], dsts[1:]
|
|
|
|
if len(dsts) > 0 {
|
|
|
|
s.renameMap[hash] = dsts
|
|
|
|
} else {
|
|
|
|
delete(s.renameMap, hash)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
s.renameMapMu.Unlock()
|
|
|
|
return dst
|
|
|
|
}
|
|
|
|
|
|
|
|
// makeRenameMap builds a map of the destination files by hash that
|
2017-01-24 11:04:09 +00:00
|
|
|
// match sizes in the slice of objects in s.renameCheck
|
|
|
|
func (s *syncCopyMove) makeRenameMap() {
|
2017-02-09 17:08:51 +00:00
|
|
|
Infof(s.fdst, "Making map for --track-renames")
|
2016-12-18 10:03:56 +00:00
|
|
|
|
2017-01-03 23:03:20 +00:00
|
|
|
// first make a map of possible sizes we need to check
|
|
|
|
possibleSizes := map[int64]struct{}{}
|
2017-01-24 11:04:09 +00:00
|
|
|
for _, obj := range s.renameCheck {
|
2017-01-03 23:03:20 +00:00
|
|
|
possibleSizes[obj.Size()] = struct{}{}
|
|
|
|
}
|
|
|
|
|
|
|
|
// pump all the dstFiles into in
|
2017-01-03 17:35:12 +00:00
|
|
|
in := make(chan Object, Config.Checkers)
|
|
|
|
go s.pumpMapToChan(s.dstFiles, in)
|
2016-12-18 10:03:56 +00:00
|
|
|
|
2017-01-03 23:03:20 +00:00
|
|
|
// now make a map of size,hash for all dstFiles
|
|
|
|
s.renameMap = make(map[string][]Object)
|
2017-01-03 17:35:12 +00:00
|
|
|
var wg sync.WaitGroup
|
2016-12-18 10:03:56 +00:00
|
|
|
wg.Add(Config.Transfers)
|
|
|
|
for i := 0; i < Config.Transfers; i++ {
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
2017-01-03 23:03:20 +00:00
|
|
|
for obj := range in {
|
|
|
|
// only create hash for dst Object if its size could match
|
|
|
|
if _, found := possibleSizes[obj.Size()]; found {
|
2017-01-03 17:35:12 +00:00
|
|
|
Stats.Checking(obj.Remote())
|
2017-01-03 23:03:20 +00:00
|
|
|
hash := s.renameHash(obj)
|
|
|
|
if hash != "" {
|
|
|
|
s.pushRenameMap(hash, obj)
|
2017-01-03 17:35:12 +00:00
|
|
|
}
|
2017-01-03 23:03:20 +00:00
|
|
|
Stats.DoneChecking(obj.Remote())
|
2016-12-18 10:03:56 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
wg.Wait()
|
2017-02-09 17:08:51 +00:00
|
|
|
Infof(s.fdst, "Finished making map for --track-renames")
|
2017-01-03 17:35:12 +00:00
|
|
|
}
|
|
|
|
|
2017-01-03 23:03:20 +00:00
|
|
|
// tryRename renames a src object when doing track renames if
|
|
|
|
// possible, it returns true if the object was renamed.
|
|
|
|
func (s *syncCopyMove) tryRename(src Object) bool {
|
|
|
|
Stats.Checking(src.Remote())
|
|
|
|
defer Stats.DoneChecking(src.Remote())
|
2016-12-18 10:03:56 +00:00
|
|
|
|
2017-02-22 19:28:22 +00:00
|
|
|
// Calculate the hash of the src object
|
2017-01-03 23:03:20 +00:00
|
|
|
hash := s.renameHash(src)
|
2017-01-03 17:35:12 +00:00
|
|
|
if hash == "" {
|
2017-01-03 23:03:20 +00:00
|
|
|
return false
|
2017-01-03 17:35:12 +00:00
|
|
|
}
|
2017-01-03 23:03:20 +00:00
|
|
|
|
2017-02-22 19:28:22 +00:00
|
|
|
// Get a match on fdst
|
2017-01-03 23:03:20 +00:00
|
|
|
dst := s.popRenameMap(hash)
|
|
|
|
if dst == nil {
|
|
|
|
return false
|
2017-01-03 17:35:12 +00:00
|
|
|
}
|
2017-01-03 23:03:20 +00:00
|
|
|
|
2017-02-22 19:28:22 +00:00
|
|
|
// Find dst object we are about to overwrite if it exists
|
|
|
|
dstOverwritten, _ := s.fdst.NewObject(src.Remote())
|
|
|
|
|
|
|
|
// Rename dst to have name src.Remote()
|
|
|
|
err := Move(s.fdst, dstOverwritten, src.Remote(), dst)
|
2017-01-03 23:03:20 +00:00
|
|
|
if err != nil {
|
2017-02-09 11:01:20 +00:00
|
|
|
Debugf(src, "Failed to rename to %q: %v", dst.Remote(), err)
|
2017-01-03 23:03:20 +00:00
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// remove file from dstFiles if present
|
|
|
|
s.dstFilesMu.Lock()
|
|
|
|
delete(s.dstFiles, dst.Remote())
|
|
|
|
s.dstFilesMu.Unlock()
|
|
|
|
|
2017-02-09 17:08:51 +00:00
|
|
|
Infof(src, "Renamed from %q", dst.Remote())
|
2017-01-03 23:03:20 +00:00
|
|
|
return true
|
2016-12-18 10:03:56 +00:00
|
|
|
}
|
|
|
|
|
2017-01-25 19:59:53 +00:00
|
|
|
// listDirJob describe a directory listing that needs to be done
|
|
|
|
type listDirJob struct {
|
|
|
|
remote string
|
|
|
|
srcDepth int
|
|
|
|
dstDepth int
|
|
|
|
noSrc bool
|
|
|
|
noDst bool
|
|
|
|
}
|
|
|
|
|
|
|
|
// Syncs fsrc into fdst
|
|
|
|
//
|
|
|
|
// If Delete is true then it deletes any files in fdst that aren't in fsrc
|
|
|
|
//
|
|
|
|
// If DoMove is true then files will be moved instead of copied
|
|
|
|
//
|
|
|
|
// dir is the start directory, "" for root
|
2017-06-13 13:35:51 +00:00
|
|
|
func (s *syncCopyMove) run() error {
|
2017-01-25 19:59:53 +00:00
|
|
|
srcDepth := Config.MaxDepth
|
|
|
|
if srcDepth < 0 {
|
|
|
|
srcDepth = MaxLevel
|
|
|
|
}
|
|
|
|
dstDepth := srcDepth
|
|
|
|
if Config.Filter.DeleteExcluded {
|
|
|
|
dstDepth = MaxLevel
|
|
|
|
}
|
|
|
|
|
|
|
|
if Same(s.fdst, s.fsrc) {
|
2017-02-09 11:01:20 +00:00
|
|
|
Errorf(s.fdst, "Nothing to do as source and destination are the same")
|
2017-01-25 19:59:53 +00:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Start background checking and transferring pipeline
|
|
|
|
s.startCheckers()
|
|
|
|
s.startRenamers()
|
|
|
|
s.startTransfers()
|
|
|
|
s.startDeleters()
|
|
|
|
s.dstFiles = make(map[string]Object)
|
|
|
|
|
|
|
|
// Start some directory listing go routines
|
|
|
|
var wg sync.WaitGroup // sync closing of go routines
|
|
|
|
var traversing sync.WaitGroup // running directory traversals
|
|
|
|
in := make(chan listDirJob, Config.Checkers)
|
|
|
|
s.startTrackRenames()
|
|
|
|
for i := 0; i < Config.Checkers; i++ {
|
|
|
|
wg.Add(1)
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
|
|
|
for {
|
|
|
|
if s.aborting() {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
select {
|
|
|
|
case job, ok := <-in:
|
|
|
|
if !ok {
|
|
|
|
return
|
|
|
|
}
|
2017-08-09 20:06:39 +00:00
|
|
|
jobs := s.processJob(job)
|
2017-01-25 19:59:53 +00:00
|
|
|
if len(jobs) > 0 {
|
|
|
|
traversing.Add(len(jobs))
|
|
|
|
go func() {
|
|
|
|
// Now we have traversed this directory, send these
|
|
|
|
// jobs off for traversal in the background
|
|
|
|
for _, newJob := range jobs {
|
|
|
|
in <- newJob
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
traversing.Done()
|
|
|
|
case <-s.abort:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
|
|
|
|
// Start the process
|
|
|
|
traversing.Add(1)
|
|
|
|
in <- listDirJob{
|
|
|
|
remote: s.dir,
|
|
|
|
srcDepth: srcDepth - 1,
|
|
|
|
dstDepth: dstDepth - 1,
|
|
|
|
}
|
|
|
|
traversing.Wait()
|
|
|
|
close(in)
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
s.stopTrackRenames()
|
|
|
|
if s.trackRenames {
|
|
|
|
// Build the map of the remaining dstFiles by hash
|
|
|
|
s.makeRenameMap()
|
|
|
|
// Attempt renames for all the files which don't have a matching dst
|
|
|
|
for _, src := range s.renameCheck {
|
|
|
|
s.toBeRenamed <- ObjectPair{src, nil}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Stop background checking and transferring pipeline
|
|
|
|
s.stopCheckers()
|
|
|
|
s.stopRenamers()
|
|
|
|
s.stopTransfers()
|
|
|
|
s.stopDeleters()
|
|
|
|
|
|
|
|
// Delete files after
|
|
|
|
if s.deleteMode == DeleteModeAfter {
|
|
|
|
if s.currentError() != nil {
|
2017-02-09 11:01:20 +00:00
|
|
|
Errorf(s.fdst, "%v", ErrorNotDeleting)
|
2017-01-25 19:59:53 +00:00
|
|
|
} else {
|
|
|
|
s.processError(s.deleteFiles(false))
|
|
|
|
}
|
|
|
|
}
|
2017-08-09 20:06:39 +00:00
|
|
|
|
|
|
|
// Prune empty directories
|
|
|
|
if s.deleteMode != DeleteModeOff {
|
|
|
|
if s.currentError() != nil {
|
|
|
|
Errorf(s.fdst, "%v", ErrorNotDeletingDirs)
|
|
|
|
} else {
|
|
|
|
s.processError(deleteEmptyDirectories(s.fdst, s.dstEmptyDirs))
|
|
|
|
}
|
|
|
|
}
|
2017-01-25 19:59:53 +00:00
|
|
|
return s.currentError()
|
|
|
|
}
|
|
|
|
|
|
|
|
// Have an object which is in the destination only
|
2017-06-30 09:54:14 +00:00
|
|
|
func (s *syncCopyMove) dstOnly(dst DirEntry, job listDirJob, jobs *[]listDirJob) {
|
2017-01-25 19:59:53 +00:00
|
|
|
if s.deleteMode == DeleteModeOff {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
switch x := dst.(type) {
|
|
|
|
case Object:
|
|
|
|
switch s.deleteMode {
|
|
|
|
case DeleteModeAfter:
|
|
|
|
// record object as needs deleting
|
|
|
|
s.dstFilesMu.Lock()
|
|
|
|
s.dstFiles[x.Remote()] = x
|
|
|
|
s.dstFilesMu.Unlock()
|
|
|
|
case DeleteModeDuring, DeleteModeOnly:
|
|
|
|
s.deleteFilesCh <- x
|
|
|
|
default:
|
|
|
|
panic(fmt.Sprintf("unexpected delete mode %d", s.deleteMode))
|
|
|
|
}
|
2017-06-30 12:37:29 +00:00
|
|
|
case Directory:
|
2017-01-25 19:59:53 +00:00
|
|
|
// Do the same thing to the entire contents of the directory
|
|
|
|
if job.dstDepth > 0 {
|
|
|
|
*jobs = append(*jobs, listDirJob{
|
|
|
|
remote: dst.Remote(),
|
|
|
|
dstDepth: job.dstDepth - 1,
|
|
|
|
noSrc: true,
|
|
|
|
})
|
|
|
|
}
|
2017-08-09 20:06:39 +00:00
|
|
|
// Record directory as it is potentially empty and needs deleting
|
|
|
|
if s.fdst.Features().CanHaveEmptyDirectories {
|
|
|
|
s.dstEmptyDirsMu.Lock()
|
|
|
|
s.dstEmptyDirs = append(s.dstEmptyDirs, dst)
|
|
|
|
s.dstEmptyDirsMu.Unlock()
|
|
|
|
}
|
2017-01-25 19:59:53 +00:00
|
|
|
default:
|
|
|
|
panic("Bad object in DirEntries")
|
|
|
|
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Have an object which is in the source only
|
2017-06-30 09:54:14 +00:00
|
|
|
func (s *syncCopyMove) srcOnly(src DirEntry, job listDirJob, jobs *[]listDirJob) {
|
2017-01-25 19:59:53 +00:00
|
|
|
if s.deleteMode == DeleteModeOnly {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
switch x := src.(type) {
|
|
|
|
case Object:
|
|
|
|
if s.trackRenames {
|
|
|
|
// Save object to check for a rename later
|
|
|
|
s.trackRenamesCh <- x
|
|
|
|
} else {
|
|
|
|
// No need to check since doesn't exist
|
|
|
|
s.toBeUploaded <- ObjectPair{x, nil}
|
|
|
|
}
|
2017-06-30 12:37:29 +00:00
|
|
|
case Directory:
|
2017-01-25 19:59:53 +00:00
|
|
|
// Do the same thing to the entire contents of the directory
|
|
|
|
if job.srcDepth > 0 {
|
|
|
|
*jobs = append(*jobs, listDirJob{
|
|
|
|
remote: src.Remote(),
|
|
|
|
srcDepth: job.srcDepth - 1,
|
|
|
|
noDst: true,
|
|
|
|
})
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
panic("Bad object in DirEntries")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Given a src and a dst, transfer the src to dst
|
2017-06-30 09:54:14 +00:00
|
|
|
func (s *syncCopyMove) transfer(dst, src DirEntry, job listDirJob, jobs *[]listDirJob) {
|
2017-01-25 19:59:53 +00:00
|
|
|
switch srcX := src.(type) {
|
|
|
|
case Object:
|
|
|
|
if s.deleteMode == DeleteModeOnly {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
dstX, ok := dst.(Object)
|
|
|
|
if ok {
|
|
|
|
s.toBeChecked <- ObjectPair{srcX, dstX}
|
|
|
|
} else {
|
|
|
|
// FIXME src is file, dst is directory
|
|
|
|
err := errors.New("can't overwrite directory with file")
|
2017-02-13 09:23:21 +00:00
|
|
|
Errorf(dst, "%v", err)
|
2017-01-25 19:59:53 +00:00
|
|
|
s.processError(err)
|
|
|
|
}
|
2017-06-30 12:37:29 +00:00
|
|
|
case Directory:
|
2017-01-25 19:59:53 +00:00
|
|
|
// Do the same thing to the entire contents of the directory
|
2017-06-30 12:37:29 +00:00
|
|
|
_, ok := dst.(Directory)
|
2017-01-25 19:59:53 +00:00
|
|
|
if ok {
|
|
|
|
if job.srcDepth > 0 && job.dstDepth > 0 {
|
|
|
|
*jobs = append(*jobs, listDirJob{
|
|
|
|
remote: src.Remote(),
|
|
|
|
srcDepth: job.srcDepth - 1,
|
|
|
|
dstDepth: job.dstDepth - 1,
|
|
|
|
})
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// FIXME src is dir, dst is file
|
|
|
|
err := errors.New("can't overwrite file with directory")
|
2017-02-13 09:23:21 +00:00
|
|
|
Errorf(dst, "%v", err)
|
2017-01-25 19:59:53 +00:00
|
|
|
s.processError(err)
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
panic("Bad object in DirEntries")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-06-30 20:24:13 +00:00
|
|
|
type matchPair struct {
|
|
|
|
src, dst DirEntry
|
|
|
|
}
|
|
|
|
|
|
|
|
// Process the two sorted listings, matching up the items in the two
|
|
|
|
// sorted slices
|
|
|
|
//
|
|
|
|
// Into srcOnly go Entries which only exist in the srcList
|
|
|
|
// Into dstOnly go Entries which only exist in the dstList
|
|
|
|
// Into matches go matchPair's of src and dst which have the same name
|
|
|
|
//
|
|
|
|
// This checks for duplicates and checks the list is sorted.
|
|
|
|
func matchListings(srcList, dstList DirEntries) (srcOnly DirEntries, dstOnly DirEntries, matches []matchPair) {
|
|
|
|
for iSrc, iDst := 0, 0; ; iSrc, iDst = iSrc+1, iDst+1 {
|
|
|
|
var src, dst DirEntry
|
|
|
|
var srcRemote, dstRemote string
|
|
|
|
if iSrc < len(srcList) {
|
|
|
|
src = srcList[iSrc]
|
|
|
|
srcRemote = src.Remote()
|
|
|
|
}
|
|
|
|
if iDst < len(dstList) {
|
|
|
|
dst = dstList[iDst]
|
|
|
|
dstRemote = dst.Remote()
|
|
|
|
}
|
|
|
|
if src == nil && dst == nil {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
if src != nil && iSrc > 0 {
|
|
|
|
prev := srcList[iSrc-1].Remote()
|
|
|
|
if srcRemote == prev {
|
2017-07-08 14:42:18 +00:00
|
|
|
Logf(src, "Duplicate %s found in source - ignoring", DirEntryType(src))
|
2017-06-30 20:24:13 +00:00
|
|
|
src = nil // ignore the src
|
|
|
|
} else if srcRemote < prev {
|
|
|
|
Errorf(src, "Out of order listing in source")
|
|
|
|
src = nil // ignore the src
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if dst != nil && iDst > 0 {
|
|
|
|
prev := dstList[iDst-1].Remote()
|
|
|
|
if dstRemote == prev {
|
2017-07-08 14:42:18 +00:00
|
|
|
Logf(dst, "Duplicate %s found in destination - ignoring", DirEntryType(dst))
|
2017-06-30 20:24:13 +00:00
|
|
|
dst = nil // ignore the dst
|
|
|
|
} else if dstRemote < prev {
|
|
|
|
Errorf(dst, "Out of order listing in destination")
|
|
|
|
dst = nil // ignore the dst
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if src != nil && dst != nil {
|
|
|
|
if srcRemote < dstRemote {
|
|
|
|
dst = nil
|
|
|
|
iDst-- // retry the dst
|
|
|
|
} else if srcRemote > dstRemote {
|
|
|
|
src = nil
|
|
|
|
iSrc-- // retry the src
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Debugf(nil, "src = %v, dst = %v", src, dst)
|
|
|
|
switch {
|
|
|
|
case src == nil && dst == nil:
|
|
|
|
// do nothing
|
|
|
|
case src == nil:
|
|
|
|
dstOnly = append(dstOnly, dst)
|
|
|
|
case dst == nil:
|
|
|
|
srcOnly = append(srcOnly, src)
|
|
|
|
default:
|
|
|
|
matches = append(matches, matchPair{src: src, dst: dst})
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2017-08-09 20:06:39 +00:00
|
|
|
// processJob processes a listDirJob listing the source and
|
|
|
|
// destination directories, comparing them and returning a slice of
|
|
|
|
// more jobs
|
|
|
|
//
|
2017-01-25 19:59:53 +00:00
|
|
|
// returns errors using processError
|
2017-08-09 20:06:39 +00:00
|
|
|
func (s *syncCopyMove) processJob(job listDirJob) (jobs []listDirJob) {
|
2017-01-25 19:59:53 +00:00
|
|
|
var (
|
|
|
|
srcList, dstList DirEntries
|
|
|
|
srcListErr, dstListErr error
|
|
|
|
wg sync.WaitGroup
|
|
|
|
)
|
|
|
|
|
|
|
|
// List the src and dst directories
|
|
|
|
if !job.noSrc {
|
|
|
|
wg.Add(1)
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
2017-06-06 22:04:01 +00:00
|
|
|
srcList, srcListErr = s.srcListDir(job.remote)
|
2017-01-25 19:59:53 +00:00
|
|
|
}()
|
|
|
|
}
|
|
|
|
if !job.noDst {
|
|
|
|
wg.Add(1)
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
2017-06-06 22:04:01 +00:00
|
|
|
dstList, dstListErr = s.dstListDir(job.remote)
|
2017-01-25 19:59:53 +00:00
|
|
|
}()
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wait for listings to complete and report errors
|
|
|
|
wg.Wait()
|
|
|
|
if srcListErr != nil {
|
|
|
|
s.processError(errors.Wrapf(srcListErr, "error reading source directory %q", job.remote))
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
if dstListErr == ErrorDirNotFound {
|
|
|
|
// Copy the stuff anyway
|
|
|
|
} else if dstListErr != nil {
|
|
|
|
s.processError(errors.Wrapf(srcListErr, "error reading destination directory %q", job.remote))
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2017-06-30 20:24:13 +00:00
|
|
|
// Work out what to do and do it
|
|
|
|
srcOnly, dstOnly, matches := matchListings(srcList, dstList)
|
|
|
|
for _, src := range srcOnly {
|
2017-01-25 19:59:53 +00:00
|
|
|
if s.aborting() {
|
|
|
|
return nil
|
|
|
|
}
|
2017-06-30 20:24:13 +00:00
|
|
|
s.srcOnly(src, job, &jobs)
|
|
|
|
}
|
|
|
|
for _, dst := range dstOnly {
|
|
|
|
if s.aborting() {
|
|
|
|
return nil
|
2017-01-25 19:59:53 +00:00
|
|
|
}
|
2017-06-30 20:24:13 +00:00
|
|
|
s.dstOnly(dst, job, &jobs)
|
|
|
|
}
|
|
|
|
for _, match := range matches {
|
|
|
|
if s.aborting() {
|
|
|
|
return nil
|
2017-01-25 19:59:53 +00:00
|
|
|
}
|
2017-06-30 20:24:13 +00:00
|
|
|
s.transfer(match.dst, match.src, job, &jobs)
|
2017-01-25 19:59:53 +00:00
|
|
|
}
|
|
|
|
return jobs
|
|
|
|
}
|
|
|
|
|
2017-01-25 19:35:14 +00:00
|
|
|
// Syncs fsrc into fdst
|
|
|
|
//
|
|
|
|
// If Delete is true then it deletes any files in fdst that aren't in fsrc
|
|
|
|
//
|
|
|
|
// If DoMove is true then files will be moved instead of copied
|
|
|
|
//
|
|
|
|
// dir is the start directory, "" for root
|
|
|
|
func runSyncCopyMove(fdst, fsrc Fs, deleteMode DeleteMode, DoMove bool) error {
|
2017-06-13 13:35:51 +00:00
|
|
|
if *oldSyncMethod {
|
|
|
|
return FatalError(errors.New("--old-sync-method is deprecated use --fast-list instead"))
|
|
|
|
}
|
2017-01-25 19:35:14 +00:00
|
|
|
if deleteMode != DeleteModeOff && DoMove {
|
2017-06-13 13:35:51 +00:00
|
|
|
return FatalError(errors.New("can't delete and move at the same time"))
|
2017-01-25 19:35:14 +00:00
|
|
|
}
|
2017-01-25 19:59:53 +00:00
|
|
|
// Run an extra pass to delete only
|
2017-06-13 13:35:51 +00:00
|
|
|
if deleteMode == DeleteModeBefore {
|
2017-01-25 19:59:53 +00:00
|
|
|
if Config.TrackRenames {
|
2017-06-13 13:35:51 +00:00
|
|
|
return FatalError(errors.New("can't use --delete-before with --track-renames"))
|
2017-01-25 19:59:53 +00:00
|
|
|
}
|
|
|
|
// only delete stuff during in this pass
|
|
|
|
do, err := newSyncCopyMove(fdst, fsrc, DeleteModeOnly, false)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2017-06-13 13:35:51 +00:00
|
|
|
err = do.run()
|
2017-01-25 19:59:53 +00:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
// Next pass does a copy only
|
|
|
|
deleteMode = DeleteModeOff
|
|
|
|
}
|
2017-01-25 19:35:14 +00:00
|
|
|
do, err := newSyncCopyMove(fdst, fsrc, deleteMode, DoMove)
|
2017-01-10 21:47:03 +00:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2017-06-13 13:35:51 +00:00
|
|
|
return do.run()
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
|
|
|
|
2017-01-25 19:35:14 +00:00
|
|
|
// Sync fsrc into fdst
|
|
|
|
func Sync(fdst, fsrc Fs) error {
|
|
|
|
return runSyncCopyMove(fdst, fsrc, Config.DeleteMode, false)
|
|
|
|
}
|
|
|
|
|
2016-07-04 12:12:33 +00:00
|
|
|
// CopyDir copies fsrc into fdst
|
|
|
|
func CopyDir(fdst, fsrc Fs) error {
|
2017-01-25 19:35:14 +00:00
|
|
|
return runSyncCopyMove(fdst, fsrc, DeleteModeOff, false)
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// moveDir moves fsrc into fdst
|
|
|
|
func moveDir(fdst, fsrc Fs) error {
|
2017-01-25 19:35:14 +00:00
|
|
|
return runSyncCopyMove(fdst, fsrc, DeleteModeOff, true)
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// MoveDir moves fsrc into fdst
|
|
|
|
func MoveDir(fdst, fsrc Fs) error {
|
|
|
|
if Same(fdst, fsrc) {
|
2017-02-09 11:01:20 +00:00
|
|
|
Errorf(fdst, "Nothing to do as source and destination are the same")
|
2016-07-04 12:12:33 +00:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// First attempt to use DirMover if exists, same Fs and no filters are active
|
2017-01-13 17:21:47 +00:00
|
|
|
if fdstDirMove := fdst.Features().DirMove; fdstDirMove != nil && SameConfig(fsrc, fdst) && Config.Filter.InActive() {
|
2016-07-11 10:36:46 +00:00
|
|
|
if Config.DryRun {
|
2017-02-09 11:01:20 +00:00
|
|
|
Logf(fdst, "Not doing server side directory move as --dry-run")
|
2016-07-11 10:36:46 +00:00
|
|
|
return nil
|
|
|
|
}
|
2017-02-09 11:01:20 +00:00
|
|
|
Debugf(fdst, "Using server side directory move")
|
2017-02-05 21:20:56 +00:00
|
|
|
err := fdstDirMove(fsrc, "", "")
|
2016-07-04 12:12:33 +00:00
|
|
|
switch err {
|
|
|
|
case ErrorCantDirMove, ErrorDirExists:
|
2017-02-09 17:08:51 +00:00
|
|
|
Infof(fdst, "Server side directory move failed - fallback to file moves: %v", err)
|
2016-07-04 12:12:33 +00:00
|
|
|
case nil:
|
2017-02-09 17:08:51 +00:00
|
|
|
Infof(fdst, "Server side directory move succeeded")
|
2016-07-04 12:12:33 +00:00
|
|
|
return nil
|
|
|
|
default:
|
|
|
|
Stats.Error()
|
2017-02-09 11:01:20 +00:00
|
|
|
Errorf(fdst, "Server side directory move failed: %v", err)
|
2016-07-04 12:12:33 +00:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-10-22 16:53:52 +00:00
|
|
|
// The two remotes mustn't overlap if we didn't do server side move
|
|
|
|
if Overlapping(fdst, fsrc) {
|
|
|
|
err := ErrorCantMoveOverlapping
|
2017-02-09 11:01:20 +00:00
|
|
|
Errorf(fdst, "%v", err)
|
2016-10-22 16:53:52 +00:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2016-07-11 10:36:46 +00:00
|
|
|
// Otherwise move the files one by one
|
|
|
|
return moveDir(fdst, fsrc)
|
2016-07-04 12:12:33 +00:00
|
|
|
}
|