2021-05-16 16:39:33 +00:00
|
|
|
// Package bisync implements bisync
|
|
|
|
// Copyright (c) 2017-2020 Chris Nelson
|
|
|
|
package bisync
|
|
|
|
|
|
|
|
import (
|
|
|
|
"bufio"
|
|
|
|
"context"
|
2023-11-09 10:04:33 +00:00
|
|
|
"errors"
|
2021-05-16 16:39:33 +00:00
|
|
|
"fmt"
|
|
|
|
"io"
|
|
|
|
"os"
|
|
|
|
"regexp"
|
2023-10-01 13:36:19 +00:00
|
|
|
"sort"
|
2021-05-16 16:39:33 +00:00
|
|
|
"strconv"
|
|
|
|
"strings"
|
|
|
|
"time"
|
|
|
|
|
2023-10-06 20:38:47 +00:00
|
|
|
"github.com/rclone/rclone/cmd/bisync/bilib"
|
2021-05-16 16:39:33 +00:00
|
|
|
"github.com/rclone/rclone/fs"
|
2023-10-01 13:36:19 +00:00
|
|
|
"github.com/rclone/rclone/fs/filter"
|
2021-05-16 16:39:33 +00:00
|
|
|
"github.com/rclone/rclone/fs/hash"
|
2023-10-01 13:36:19 +00:00
|
|
|
"github.com/rclone/rclone/fs/operations"
|
|
|
|
"golang.org/x/exp/slices"
|
2021-05-16 16:39:33 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
// ListingHeader defines first line of a listing
|
|
|
|
const ListingHeader = "# bisync listing v1 from"
|
|
|
|
|
|
|
|
// lineRegex and lineFormat define listing line format
|
|
|
|
//
|
2022-08-05 15:35:41 +00:00
|
|
|
// flags <- size -> <- hash -> id <------------ modtime -----------> "<----- remote"
|
|
|
|
// - 3009805 md5:xxxxxx - 2006-01-02T15:04:05.000000000-0700 "12 - Wait.mp3"
|
2021-05-16 16:39:33 +00:00
|
|
|
//
|
|
|
|
// flags: "-" for a file and "d" for a directory (reserved)
|
|
|
|
// hash: "type:value" or "-" (example: "md5:378840336ab14afa9c6b8d887e68a340")
|
|
|
|
// id: "-" (reserved)
|
|
|
|
const lineFormat = "%s %8d %s %s %s %q\n"
|
|
|
|
|
|
|
|
var lineRegex = regexp.MustCompile(`^(\S) +(\d+) (\S+) (\S+) (\d{4}-\d\d-\d\dT\d\d:\d\d:\d\d\.\d{9}[+-]\d{4}) (".+")$`)
|
|
|
|
|
|
|
|
// timeFormat defines time format used in listings
|
|
|
|
const timeFormat = "2006-01-02T15:04:05.000000000-0700"
|
|
|
|
|
|
|
|
// TZ defines time zone used in listings
|
|
|
|
var TZ = time.UTC
|
|
|
|
var tzLocal = false
|
|
|
|
|
|
|
|
// fileInfo describes a file
|
|
|
|
type fileInfo struct {
|
2023-07-11 11:09:06 +00:00
|
|
|
size int64
|
|
|
|
time time.Time
|
|
|
|
hash string
|
|
|
|
id string
|
|
|
|
flags string
|
2021-05-16 16:39:33 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// fileList represents a listing
|
|
|
|
type fileList struct {
|
|
|
|
list []string
|
|
|
|
info map[string]*fileInfo
|
|
|
|
hash hash.Type
|
|
|
|
}
|
|
|
|
|
|
|
|
func newFileList() *fileList {
|
|
|
|
return &fileList{
|
|
|
|
info: map[string]*fileInfo{},
|
|
|
|
list: []string{},
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (ls *fileList) empty() bool {
|
2023-10-07 10:33:43 +00:00
|
|
|
if ls == nil {
|
|
|
|
return true
|
|
|
|
}
|
2021-05-16 16:39:33 +00:00
|
|
|
return len(ls.list) == 0
|
|
|
|
}
|
|
|
|
|
|
|
|
func (ls *fileList) has(file string) bool {
|
|
|
|
_, found := ls.info[file]
|
2023-10-01 13:36:19 +00:00
|
|
|
if !found {
|
|
|
|
//try unquoting
|
|
|
|
file, _ = strconv.Unquote(`"` + file + `"`)
|
|
|
|
_, found = ls.info[file]
|
|
|
|
}
|
2021-05-16 16:39:33 +00:00
|
|
|
return found
|
|
|
|
}
|
|
|
|
|
|
|
|
func (ls *fileList) get(file string) *fileInfo {
|
2023-10-01 13:36:19 +00:00
|
|
|
info, found := ls.info[file]
|
|
|
|
if !found {
|
|
|
|
//try unquoting
|
|
|
|
file, _ = strconv.Unquote(`"` + file + `"`)
|
|
|
|
info = ls.info[fmt.Sprint(file)]
|
|
|
|
}
|
|
|
|
return info
|
|
|
|
}
|
|
|
|
|
2023-10-06 20:38:47 +00:00
|
|
|
// copy file from ls to dest
|
|
|
|
func (ls *fileList) getPut(file string, dest *fileList) {
|
|
|
|
f := ls.get(file)
|
|
|
|
dest.put(file, f.size, f.time, f.hash, f.id, f.flags)
|
|
|
|
}
|
|
|
|
|
2023-10-07 10:33:43 +00:00
|
|
|
func (ls *fileList) getPutAll(dest *fileList) {
|
|
|
|
for file, f := range ls.info {
|
|
|
|
dest.put(file, f.size, f.time, f.hash, f.id, f.flags)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-10-01 13:36:19 +00:00
|
|
|
func (ls *fileList) remove(file string) {
|
|
|
|
if ls.has(file) {
|
|
|
|
ls.list = slices.Delete(ls.list, slices.Index(ls.list, file), slices.Index(ls.list, file)+1)
|
|
|
|
delete(ls.info, file)
|
|
|
|
}
|
2021-05-16 16:39:33 +00:00
|
|
|
}
|
|
|
|
|
2023-11-06 15:34:47 +00:00
|
|
|
func (ls *fileList) put(file string, size int64, modtime time.Time, hash, id string, flags string) {
|
2021-05-16 16:39:33 +00:00
|
|
|
fi := ls.get(file)
|
|
|
|
if fi != nil {
|
|
|
|
fi.size = size
|
2023-11-06 15:34:47 +00:00
|
|
|
// if already have higher precision of same time, avoid overwriting it
|
|
|
|
if fi.time != modtime {
|
|
|
|
if modtime.Before(fi.time) && fi.time.Sub(modtime) < time.Second {
|
|
|
|
modtime = fi.time
|
|
|
|
}
|
|
|
|
}
|
|
|
|
fi.time = modtime
|
2023-10-01 13:36:19 +00:00
|
|
|
fi.hash = hash
|
|
|
|
fi.id = id
|
|
|
|
fi.flags = flags
|
2021-05-16 16:39:33 +00:00
|
|
|
} else {
|
|
|
|
fi = &fileInfo{
|
2023-07-11 11:09:06 +00:00
|
|
|
size: size,
|
2023-11-06 15:34:47 +00:00
|
|
|
time: modtime,
|
2023-07-11 11:09:06 +00:00
|
|
|
hash: hash,
|
|
|
|
id: id,
|
|
|
|
flags: flags,
|
2021-05-16 16:39:33 +00:00
|
|
|
}
|
|
|
|
ls.info[file] = fi
|
|
|
|
ls.list = append(ls.list, file)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (ls *fileList) getTime(file string) time.Time {
|
|
|
|
fi := ls.get(file)
|
|
|
|
if fi == nil {
|
|
|
|
return time.Time{}
|
|
|
|
}
|
|
|
|
return fi.time
|
|
|
|
}
|
|
|
|
|
2023-10-06 20:36:00 +00:00
|
|
|
// also returns false if not found
|
|
|
|
func (ls *fileList) isDir(file string) bool {
|
|
|
|
fi := ls.get(file)
|
|
|
|
if fi != nil {
|
|
|
|
if fi.flags == "d" {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
2021-05-16 16:39:33 +00:00
|
|
|
func (ls *fileList) beforeOther(other *fileList, file string) bool {
|
|
|
|
thisTime := ls.getTime(file)
|
|
|
|
thatTime := other.getTime(file)
|
|
|
|
if thisTime.IsZero() || thatTime.IsZero() {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
return thisTime.Before(thatTime)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (ls *fileList) afterTime(file string, time time.Time) bool {
|
|
|
|
fi := ls.get(file)
|
|
|
|
if fi == nil {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
return fi.time.After(time)
|
|
|
|
}
|
|
|
|
|
2023-10-01 13:36:19 +00:00
|
|
|
// sort by path name
|
|
|
|
func (ls *fileList) sort() {
|
|
|
|
sort.SliceStable(ls.list, func(i, j int) bool {
|
|
|
|
return ls.list[i] < ls.list[j]
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2021-05-16 16:39:33 +00:00
|
|
|
// save will save listing to a file.
|
|
|
|
func (ls *fileList) save(ctx context.Context, listing string) error {
|
|
|
|
file, err := os.Create(listing)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2023-10-01 13:36:19 +00:00
|
|
|
ls.sort()
|
2021-05-16 16:39:33 +00:00
|
|
|
|
|
|
|
hashName := ""
|
|
|
|
if ls.hash != hash.None {
|
|
|
|
hashName = ls.hash.String()
|
|
|
|
}
|
|
|
|
|
|
|
|
_, err = fmt.Fprintf(file, "%s %s\n", ListingHeader, time.Now().In(TZ).Format(timeFormat))
|
|
|
|
if err != nil {
|
|
|
|
_ = file.Close()
|
|
|
|
_ = os.Remove(listing)
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, remote := range ls.list {
|
|
|
|
fi := ls.get(remote)
|
|
|
|
|
|
|
|
time := fi.time.In(TZ).Format(timeFormat)
|
|
|
|
|
|
|
|
hash := "-"
|
|
|
|
if hashName != "" && fi.hash != "" {
|
|
|
|
hash = hashName + ":" + fi.hash
|
|
|
|
}
|
|
|
|
|
|
|
|
id := fi.id
|
|
|
|
if id == "" {
|
|
|
|
id = "-"
|
|
|
|
}
|
|
|
|
|
2023-07-11 11:09:06 +00:00
|
|
|
flags := fi.flags
|
|
|
|
if flags == "" {
|
|
|
|
flags = "-"
|
|
|
|
}
|
|
|
|
|
2021-05-16 16:39:33 +00:00
|
|
|
_, err = fmt.Fprintf(file, lineFormat, flags, fi.size, hash, id, time, remote)
|
|
|
|
if err != nil {
|
|
|
|
_ = file.Close()
|
|
|
|
_ = os.Remove(listing)
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return file.Close()
|
|
|
|
}
|
|
|
|
|
|
|
|
// loadListing will load listing from a file.
|
|
|
|
// The key is the path to the file relative to the Path1/Path2 base.
|
|
|
|
// File size of -1, as for Google Docs, prints a warning and won't be loaded.
|
|
|
|
func (b *bisyncRun) loadListing(listing string) (*fileList, error) {
|
|
|
|
file, err := os.Open(listing)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
defer func() {
|
|
|
|
_ = file.Close()
|
|
|
|
}()
|
|
|
|
|
|
|
|
reader := bufio.NewReader(file)
|
|
|
|
ls := newFileList()
|
|
|
|
lastHashName := ""
|
|
|
|
|
|
|
|
for {
|
|
|
|
line, err := reader.ReadString('\n')
|
|
|
|
if err == io.EOF {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
line = strings.TrimSuffix(line, "\n")
|
|
|
|
if line == "" || line[0] == '#' {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
match := lineRegex.FindStringSubmatch(line)
|
|
|
|
if match == nil {
|
|
|
|
fs.Logf(listing, "Ignoring incorrect line: %q", line)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
flags, sizeStr, hashStr := match[1], match[2], match[3]
|
|
|
|
id, timeStr, nameStr := match[4], match[5], match[6]
|
|
|
|
|
|
|
|
sizeVal, sizeErr := strconv.ParseInt(sizeStr, 10, 64)
|
|
|
|
timeVal, timeErr := time.ParseInLocation(timeFormat, timeStr, TZ)
|
|
|
|
nameVal, nameErr := strconv.Unquote(nameStr)
|
|
|
|
|
|
|
|
hashName, hashVal, hashErr := parseHash(hashStr)
|
|
|
|
if hashErr == nil && hashName != "" {
|
|
|
|
if lastHashName == "" {
|
|
|
|
lastHashName = hashName
|
|
|
|
hashErr = ls.hash.Set(hashName)
|
|
|
|
} else if hashName != lastHashName {
|
|
|
|
fs.Logf(listing, "Inconsistent hash type in line: %q", line)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-07-11 11:09:06 +00:00
|
|
|
if (flags != "-" && flags != "d") || id != "-" || sizeErr != nil || timeErr != nil || hashErr != nil || nameErr != nil {
|
2021-05-16 16:39:33 +00:00
|
|
|
fs.Logf(listing, "Ignoring incorrect line: %q", line)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
if ls.has(nameVal) {
|
|
|
|
fs.Logf(listing, "Duplicate line (keeping latest): %q", line)
|
|
|
|
if ls.afterTime(nameVal, timeVal) {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-07-11 11:09:06 +00:00
|
|
|
ls.put(nameVal, sizeVal, timeVal.In(TZ), hashVal, id, flags)
|
2021-05-16 16:39:33 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return ls, nil
|
|
|
|
}
|
|
|
|
|
2023-10-07 10:33:43 +00:00
|
|
|
// saveOldListings saves the most recent successful listing, in case we need to rollback on error
|
2023-10-06 20:38:47 +00:00
|
|
|
func (b *bisyncRun) saveOldListings() {
|
2023-10-07 10:33:43 +00:00
|
|
|
b.handleErr(b.listing1, "error saving old Path1 listing", bilib.CopyFileIfExists(b.listing1, b.listing1+"-old"), true, true)
|
|
|
|
b.handleErr(b.listing2, "error saving old Path2 listing", bilib.CopyFileIfExists(b.listing2, b.listing2+"-old"), true, true)
|
|
|
|
}
|
|
|
|
|
|
|
|
// replaceCurrentListings saves both ".lst-new" listings as ".lst"
|
|
|
|
func (b *bisyncRun) replaceCurrentListings() {
|
|
|
|
b.handleErr(b.newListing1, "error replacing Path1 listing", bilib.CopyFileIfExists(b.newListing1, b.listing1), true, true)
|
|
|
|
b.handleErr(b.newListing2, "error replacing Path2 listing", bilib.CopyFileIfExists(b.newListing2, b.listing2), true, true)
|
2023-10-06 20:38:47 +00:00
|
|
|
}
|
|
|
|
|
2021-05-16 16:39:33 +00:00
|
|
|
func parseHash(str string) (string, string, error) {
|
|
|
|
if str == "-" {
|
|
|
|
return "", "", nil
|
|
|
|
}
|
|
|
|
if pos := strings.Index(str, ":"); pos > 0 {
|
|
|
|
name, val := str[:pos], str[pos+1:]
|
|
|
|
if name != "" && val != "" {
|
|
|
|
return name, val, nil
|
|
|
|
}
|
|
|
|
}
|
2021-11-04 10:12:57 +00:00
|
|
|
return "", "", fmt.Errorf("invalid hash %q", str)
|
2021-05-16 16:39:33 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// checkListing verifies that listing is not empty (unless resynching)
|
|
|
|
func (b *bisyncRun) checkListing(ls *fileList, listing, msg string) error {
|
|
|
|
if b.opt.Resync || !ls.empty() {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
fs.Errorf(nil, "Empty %s listing. Cannot sync to an empty directory: %s", msg, listing)
|
|
|
|
b.critical = true
|
2023-07-11 10:57:49 +00:00
|
|
|
b.retryable = true
|
2021-11-04 10:12:57 +00:00
|
|
|
return fmt.Errorf("empty %s listing: %s", msg, listing)
|
2021-05-16 16:39:33 +00:00
|
|
|
}
|
2023-07-11 11:09:06 +00:00
|
|
|
|
|
|
|
// listingNum should be 1 for path1 or 2 for path2
|
|
|
|
func (b *bisyncRun) loadListingNum(listingNum int) (*fileList, error) {
|
|
|
|
listingpath := b.basePath + ".path1.lst-new"
|
|
|
|
if listingNum == 2 {
|
|
|
|
listingpath = b.basePath + ".path2.lst-new"
|
|
|
|
}
|
|
|
|
|
|
|
|
if b.opt.DryRun {
|
|
|
|
listingpath = strings.Replace(listingpath, ".lst-", ".lst-dry-", 1)
|
|
|
|
}
|
|
|
|
|
|
|
|
fs.Debugf(nil, "loading listing for path %d at: %s", listingNum, listingpath)
|
|
|
|
return b.loadListing(listingpath)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (b *bisyncRun) listDirsOnly(listingNum int) (*fileList, error) {
|
|
|
|
var fulllisting *fileList
|
|
|
|
var dirsonly = newFileList()
|
|
|
|
var err error
|
|
|
|
|
|
|
|
if !b.opt.CreateEmptySrcDirs {
|
|
|
|
return dirsonly, err
|
|
|
|
}
|
|
|
|
|
|
|
|
fulllisting, err = b.loadListingNum(listingNum)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
b.critical = true
|
|
|
|
b.retryable = true
|
|
|
|
fs.Debugf(nil, "Error loading listing to generate dirsonly list: %v", err)
|
|
|
|
return dirsonly, err
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, obj := range fulllisting.list {
|
|
|
|
info := fulllisting.get(obj)
|
|
|
|
|
|
|
|
if info.flags == "d" {
|
|
|
|
fs.Debugf(nil, "found a dir: %s", obj)
|
|
|
|
dirsonly.put(obj, info.size, info.time, info.hash, info.id, info.flags)
|
|
|
|
} else {
|
|
|
|
fs.Debugf(nil, "not a dir: %s", obj)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return dirsonly, err
|
|
|
|
}
|
2023-10-01 13:36:19 +00:00
|
|
|
|
|
|
|
// ConvertPrecision returns the Modtime rounded to Dest's precision if lower, otherwise unchanged
|
|
|
|
// Need to use the other fs's precision (if lower) when copying
|
|
|
|
// Note: we need to use Truncate rather than Round so that After() is reliable.
|
|
|
|
// (2023-11-02 20:22:45.552679442 +0000 < UTC 2023-11-02 20:22:45.553 +0000 UTC)
|
|
|
|
func ConvertPrecision(Modtime time.Time, dst fs.Fs) time.Time {
|
|
|
|
DestPrecision := dst.Precision()
|
|
|
|
|
|
|
|
// In case it's wrapping an Fs with lower precision, try unwrapping and use the lowest.
|
|
|
|
if Modtime.Truncate(DestPrecision).After(Modtime.Truncate(fs.UnWrapFs(dst).Precision())) {
|
|
|
|
DestPrecision = fs.UnWrapFs(dst).Precision()
|
|
|
|
}
|
|
|
|
|
|
|
|
if Modtime.After(Modtime.Truncate(DestPrecision)) {
|
|
|
|
return Modtime.Truncate(DestPrecision)
|
|
|
|
}
|
|
|
|
return Modtime
|
|
|
|
}
|
|
|
|
|
|
|
|
// modifyListing will modify the listing based on the results of the sync
|
2023-10-06 20:38:47 +00:00
|
|
|
func (b *bisyncRun) modifyListing(ctx context.Context, src fs.Fs, dst fs.Fs, results []Results, queues queues, is1to2 bool) (err error) {
|
2023-10-01 13:36:19 +00:00
|
|
|
queue := queues.copy2to1
|
|
|
|
renames := queues.renamed2
|
|
|
|
direction := "2to1"
|
|
|
|
if is1to2 {
|
|
|
|
queue = queues.copy1to2
|
|
|
|
renames = queues.renamed1
|
|
|
|
direction = "1to2"
|
|
|
|
}
|
|
|
|
|
|
|
|
fs.Debugf(nil, "updating %s", direction)
|
|
|
|
fs.Debugf(nil, "RESULTS: %v", results)
|
|
|
|
fs.Debugf(nil, "QUEUE: %v", queue)
|
|
|
|
|
2023-10-06 20:38:47 +00:00
|
|
|
srcListing, dstListing := b.getListingNames(is1to2)
|
2023-10-01 13:36:19 +00:00
|
|
|
srcList, err := b.loadListing(srcListing)
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("cannot read prior listing: %w", err)
|
|
|
|
}
|
|
|
|
dstList, err := b.loadListing(dstListing)
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("cannot read prior listing: %w", err)
|
|
|
|
}
|
|
|
|
// set list hash type
|
|
|
|
if b.opt.Resync && !b.opt.IgnoreListingChecksum {
|
|
|
|
srcList.hash = src.Hashes().GetOne()
|
|
|
|
dstList.hash = dst.Hashes().GetOne()
|
|
|
|
}
|
|
|
|
|
|
|
|
srcWinners := newFileList()
|
|
|
|
dstWinners := newFileList()
|
|
|
|
errors := newFileList()
|
|
|
|
ctxRecheck, filterRecheck := filter.AddConfig(ctx)
|
|
|
|
|
|
|
|
for _, result := range results {
|
|
|
|
if result.Name == "" {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
2023-11-09 10:04:33 +00:00
|
|
|
if result.AltName != "" {
|
|
|
|
b.aliases.Add(result.Name, result.AltName)
|
|
|
|
}
|
|
|
|
|
2023-10-01 13:36:19 +00:00
|
|
|
if result.Flags == "d" && !b.opt.CreateEmptySrcDirs {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
// build src winners list
|
|
|
|
if result.IsSrc && result.Src != "" && (result.Winner.Err == nil || result.Flags == "d") {
|
|
|
|
srcWinners.put(result.Name, result.Size, ConvertPrecision(result.Modtime, src), result.Hash, "-", result.Flags)
|
|
|
|
fs.Debugf(nil, "winner: copy to src: %v", result)
|
|
|
|
}
|
|
|
|
|
|
|
|
// build dst winners list
|
|
|
|
if result.IsWinner && result.Winner.Side != "none" && (result.Winner.Err == nil || result.Flags == "d") {
|
|
|
|
dstWinners.put(result.Name, result.Size, ConvertPrecision(result.Modtime, dst), result.Hash, "-", result.Flags)
|
|
|
|
fs.Debugf(nil, "winner: copy to dst: %v", result)
|
|
|
|
}
|
|
|
|
|
|
|
|
// build errors list
|
|
|
|
if result.Err != nil || result.Winner.Err != nil {
|
|
|
|
errors.put(result.Name, result.Size, result.Modtime, result.Hash, "-", result.Flags)
|
|
|
|
if err := filterRecheck.AddFile(result.Name); err != nil {
|
|
|
|
fs.Debugf(result.Name, "error adding file to recheck filter: %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-11-09 10:04:33 +00:00
|
|
|
ci := fs.GetConfig(ctx)
|
2023-10-01 13:36:19 +00:00
|
|
|
updateLists := func(side string, winners, list *fileList) {
|
2023-10-09 03:16:23 +00:00
|
|
|
for _, queueFile := range queue.ToList() {
|
|
|
|
if !winners.has(queueFile) && list.has(queueFile) && !errors.has(queueFile) {
|
|
|
|
// removals from side
|
|
|
|
list.remove(queueFile)
|
|
|
|
fs.Debugf(nil, "decision: removed from %s: %v", side, queueFile)
|
|
|
|
} else if winners.has(queueFile) {
|
2023-10-01 13:36:19 +00:00
|
|
|
// copies to side
|
2023-10-09 03:16:23 +00:00
|
|
|
new := winners.get(queueFile)
|
|
|
|
|
2023-11-09 10:04:33 +00:00
|
|
|
// handle normalization
|
|
|
|
if side == "dst" {
|
|
|
|
alias := b.aliases.Alias(queueFile)
|
|
|
|
if alias != queueFile {
|
2023-10-09 03:16:23 +00:00
|
|
|
// use the (non-identical) existing name, unless --fix-case
|
|
|
|
if ci.FixCase {
|
2023-11-09 10:04:33 +00:00
|
|
|
fs.Debugf(direction, "removing %s and adding %s as --fix-case was specified", alias, queueFile)
|
|
|
|
list.remove(alias)
|
2023-10-09 03:16:23 +00:00
|
|
|
} else {
|
2023-11-09 10:04:33 +00:00
|
|
|
fs.Debugf(direction, "casing/unicode difference detected. using %s instead of %s", alias, queueFile)
|
|
|
|
queueFile = alias
|
2023-10-09 03:16:23 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
list.put(queueFile, new.size, new.time, new.hash, new.id, new.flags)
|
|
|
|
fs.Debugf(nil, "decision: copied to %s: %v", side, queueFile)
|
2023-10-01 13:36:19 +00:00
|
|
|
} else {
|
2023-10-09 03:16:23 +00:00
|
|
|
fs.Debugf(queueFile, "file in queue but missing from %s transfers", side)
|
|
|
|
if err := filterRecheck.AddFile(queueFile); err != nil {
|
|
|
|
fs.Debugf(queueFile, "error adding file to recheck filter: %v", err)
|
2023-10-01 13:36:19 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
updateLists("src", srcWinners, srcList)
|
|
|
|
updateLists("dst", dstWinners, dstList)
|
|
|
|
|
|
|
|
// account for "deltaOthers" we handled separately
|
|
|
|
if queues.deletedonboth.NotEmpty() {
|
|
|
|
for file := range queues.deletedonboth {
|
|
|
|
srcList.remove(file)
|
|
|
|
dstList.remove(file)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if renames.NotEmpty() && !b.opt.DryRun {
|
|
|
|
// renamed on src and copied to dst
|
|
|
|
renamesList := renames.ToList()
|
|
|
|
for _, file := range renamesList {
|
|
|
|
// we'll handle the other side when we go the other direction
|
|
|
|
newName := file + "..path2"
|
|
|
|
oppositeName := file + "..path1"
|
|
|
|
if is1to2 {
|
|
|
|
newName = file + "..path1"
|
|
|
|
oppositeName = file + "..path2"
|
|
|
|
}
|
|
|
|
var new *fileInfo
|
|
|
|
// we prefer to get the info from the ..path1 / ..path2 versions
|
|
|
|
// since they were actually copied as opposed to operations.MoveFile()'d.
|
|
|
|
// the size/time/hash info is therefore fresher on the renames
|
|
|
|
// but we'll settle for the original if we have to.
|
|
|
|
if srcList.has(newName) {
|
|
|
|
new = srcList.get(newName)
|
|
|
|
} else if srcList.has(oppositeName) {
|
|
|
|
new = srcList.get(oppositeName)
|
|
|
|
} else if srcList.has(file) {
|
|
|
|
new = srcList.get(file)
|
|
|
|
} else {
|
|
|
|
if err := filterRecheck.AddFile(file); err != nil {
|
|
|
|
fs.Debugf(file, "error adding file to recheck filter: %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
srcList.put(newName, new.size, new.time, new.hash, new.id, new.flags)
|
|
|
|
dstList.put(newName, new.size, ConvertPrecision(new.time, src), new.hash, new.id, new.flags)
|
|
|
|
srcList.remove(file)
|
|
|
|
dstList.remove(file)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// recheck the ones we skipped because they were equal
|
|
|
|
// we never got their info because they were never synced.
|
2023-10-07 10:33:43 +00:00
|
|
|
// TODO: add flag to skip this? (since it re-lists)
|
2023-10-01 13:36:19 +00:00
|
|
|
if queues.renameSkipped.NotEmpty() {
|
|
|
|
skippedList := queues.renameSkipped.ToList()
|
|
|
|
for _, file := range skippedList {
|
|
|
|
if err := filterRecheck.AddFile(file); err != nil {
|
|
|
|
fs.Debugf(file, "error adding file to recheck filter: %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2023-10-07 10:33:43 +00:00
|
|
|
// skipped dirs -- nothing to recheck, just add them
|
|
|
|
// (they are not necessarily there already, if they are new)
|
|
|
|
path1List := srcList
|
|
|
|
path2List := dstList
|
|
|
|
if !is1to2 {
|
|
|
|
path1List = dstList
|
|
|
|
path2List = srcList
|
|
|
|
}
|
|
|
|
if !queues.skippedDirs1.empty() {
|
|
|
|
queues.skippedDirs1.getPutAll(path1List)
|
|
|
|
}
|
|
|
|
if !queues.skippedDirs2.empty() {
|
|
|
|
queues.skippedDirs2.getPutAll(path2List)
|
|
|
|
}
|
2023-10-01 13:36:19 +00:00
|
|
|
|
|
|
|
if filterRecheck.HaveFilesFrom() {
|
2023-11-09 10:04:33 +00:00
|
|
|
// also include any aliases
|
|
|
|
recheckFiles := filterRecheck.Files()
|
|
|
|
for recheckFile := range recheckFiles {
|
|
|
|
alias := b.aliases.Alias(recheckFile)
|
|
|
|
if recheckFile != alias {
|
|
|
|
if err := filterRecheck.AddFile(alias); err != nil {
|
|
|
|
fs.Debugf(alias, "error adding file to recheck filter: %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2023-10-06 20:38:47 +00:00
|
|
|
b.recheck(ctxRecheck, src, dst, srcList, dstList, is1to2)
|
2023-10-01 13:36:19 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// update files
|
|
|
|
err = srcList.save(ctx, srcListing)
|
|
|
|
if err != nil {
|
|
|
|
b.abort = true
|
|
|
|
}
|
|
|
|
err = dstList.save(ctx, dstListing)
|
|
|
|
if err != nil {
|
|
|
|
b.abort = true
|
|
|
|
}
|
|
|
|
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// recheck the ones we're not sure about
|
2023-10-06 20:38:47 +00:00
|
|
|
func (b *bisyncRun) recheck(ctxRecheck context.Context, src, dst fs.Fs, srcList, dstList *fileList, is1to2 bool) {
|
2023-10-01 13:36:19 +00:00
|
|
|
var srcObjs []fs.Object
|
|
|
|
var dstObjs []fs.Object
|
2023-10-06 20:38:47 +00:00
|
|
|
var resolved []string
|
|
|
|
var toRollback []string
|
2023-10-01 13:36:19 +00:00
|
|
|
|
|
|
|
if err := operations.ListFn(ctxRecheck, src, func(obj fs.Object) {
|
|
|
|
srcObjs = append(srcObjs, obj)
|
|
|
|
}); err != nil {
|
|
|
|
fs.Debugf(src, "error recchecking src obj: %v", err)
|
|
|
|
}
|
|
|
|
if err := operations.ListFn(ctxRecheck, dst, func(obj fs.Object) {
|
|
|
|
dstObjs = append(dstObjs, obj)
|
|
|
|
}); err != nil {
|
|
|
|
fs.Debugf(dst, "error recchecking dst obj: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
putObj := func(obj fs.Object, f fs.Fs, list *fileList) {
|
|
|
|
hashVal := ""
|
|
|
|
if !b.opt.IgnoreListingChecksum {
|
|
|
|
hashType := f.Hashes().GetOne()
|
|
|
|
if hashType != hash.None {
|
|
|
|
hashVal, _ = obj.Hash(ctxRecheck, hashType)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
list.put(obj.Remote(), obj.Size(), obj.ModTime(ctxRecheck), hashVal, "-", "-")
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, srcObj := range srcObjs {
|
|
|
|
fs.Debugf(srcObj, "rechecking")
|
|
|
|
for _, dstObj := range dstObjs {
|
2023-11-09 10:04:33 +00:00
|
|
|
if srcObj.Remote() == dstObj.Remote() || srcObj.Remote() == b.aliases.Alias(dstObj.Remote()) {
|
2023-10-01 13:36:19 +00:00
|
|
|
if operations.Equal(ctxRecheck, srcObj, dstObj) || b.opt.DryRun {
|
|
|
|
putObj(srcObj, src, srcList)
|
|
|
|
putObj(dstObj, dst, dstList)
|
2023-10-06 20:38:47 +00:00
|
|
|
resolved = append(resolved, srcObj.Remote())
|
2023-10-01 13:36:19 +00:00
|
|
|
} else {
|
|
|
|
fs.Infof(srcObj, "files not equal on recheck: %v %v", srcObj, dstObj)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2023-10-06 20:38:47 +00:00
|
|
|
// if srcObj not resolved by now (either because no dstObj match or files not equal),
|
|
|
|
// roll it back to old version, so it gets retried next time.
|
2023-11-09 10:04:33 +00:00
|
|
|
// skip and error during --resync, as rollback is not possible
|
2023-10-06 20:38:47 +00:00
|
|
|
if !slices.Contains(resolved, srcObj.Remote()) && !b.opt.DryRun {
|
2023-11-09 10:04:33 +00:00
|
|
|
if b.opt.Resync {
|
|
|
|
b.handleErr(srcObj, "Unable to rollback during --resync", errors.New("no dstObj match or files not equal"), true, false)
|
|
|
|
} else {
|
|
|
|
toRollback = append(toRollback, srcObj.Remote())
|
|
|
|
}
|
2023-10-06 20:38:47 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
if len(toRollback) > 0 {
|
|
|
|
srcListing, dstListing := b.getListingNames(is1to2)
|
|
|
|
oldSrc, err := b.loadListing(srcListing + "-old")
|
2023-10-07 10:33:43 +00:00
|
|
|
b.handleErr(oldSrc, "error loading old src listing", err, true, true)
|
2023-10-06 20:38:47 +00:00
|
|
|
oldDst, err := b.loadListing(dstListing + "-old")
|
2023-10-07 10:33:43 +00:00
|
|
|
b.handleErr(oldDst, "error loading old dst listing", err, true, true)
|
2023-11-09 10:04:33 +00:00
|
|
|
if b.critical {
|
|
|
|
return
|
|
|
|
}
|
2023-10-06 20:38:47 +00:00
|
|
|
|
|
|
|
for _, item := range toRollback {
|
|
|
|
rollback(item, oldSrc, srcList)
|
|
|
|
rollback(item, oldDst, dstList)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (b *bisyncRun) getListingNames(is1to2 bool) (srcListing string, dstListing string) {
|
|
|
|
if is1to2 {
|
|
|
|
return b.listing1, b.listing2
|
|
|
|
}
|
|
|
|
return b.listing2, b.listing1
|
|
|
|
}
|
|
|
|
|
|
|
|
func rollback(item string, oldList, newList *fileList) {
|
|
|
|
if oldList.has(item) {
|
|
|
|
oldList.getPut(item, newList)
|
|
|
|
} else {
|
|
|
|
newList.remove(item)
|
2023-10-01 13:36:19 +00:00
|
|
|
}
|
|
|
|
}
|