forked from TrueCloudLab/restic
Merge pull request #2750 from metalsp0rk/min-packsize
Add `backup --file-read-concurrency` flag
This commit is contained in:
commit
e99ad39b34
7 changed files with 34 additions and 10 deletions
6
changelog/unreleased/pull-2750
Normal file
6
changelog/unreleased/pull-2750
Normal file
|
@ -0,0 +1,6 @@
|
||||||
|
Enhancement: Make backup file read concurrency configurable
|
||||||
|
|
||||||
|
The `backup` command now supports a `--read-concurrency` flag to allowing
|
||||||
|
tuning restic for very fast storage like NVME disks.
|
||||||
|
|
||||||
|
https://github.com/restic/restic/pull/2750
|
|
@ -11,6 +11,7 @@ import (
|
||||||
"path"
|
"path"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
"runtime"
|
"runtime"
|
||||||
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
@ -100,6 +101,7 @@ type BackupOptions struct {
|
||||||
IgnoreCtime bool
|
IgnoreCtime bool
|
||||||
UseFsSnapshot bool
|
UseFsSnapshot bool
|
||||||
DryRun bool
|
DryRun bool
|
||||||
|
ReadConcurrency uint
|
||||||
}
|
}
|
||||||
|
|
||||||
var backupOptions BackupOptions
|
var backupOptions BackupOptions
|
||||||
|
@ -124,7 +126,7 @@ func init() {
|
||||||
f.BoolVar(&backupOptions.Stdin, "stdin", false, "read backup from stdin")
|
f.BoolVar(&backupOptions.Stdin, "stdin", false, "read backup from stdin")
|
||||||
f.StringVar(&backupOptions.StdinFilename, "stdin-filename", "stdin", "`filename` to use when reading from stdin")
|
f.StringVar(&backupOptions.StdinFilename, "stdin-filename", "stdin", "`filename` to use when reading from stdin")
|
||||||
f.Var(&backupOptions.Tags, "tag", "add `tags` for the new snapshot in the format `tag[,tag,...]` (can be specified multiple times)")
|
f.Var(&backupOptions.Tags, "tag", "add `tags` for the new snapshot in the format `tag[,tag,...]` (can be specified multiple times)")
|
||||||
|
f.UintVar(&backupOptions.ReadConcurrency, "read-concurrency", 0, "read `n` file concurrently. (default: $RESTIC_READ_CONCURRENCY or 2)")
|
||||||
f.StringVarP(&backupOptions.Host, "host", "H", "", "set the `hostname` for the snapshot manually. To prevent an expensive rescan use the \"parent\" flag")
|
f.StringVarP(&backupOptions.Host, "host", "H", "", "set the `hostname` for the snapshot manually. To prevent an expensive rescan use the \"parent\" flag")
|
||||||
f.StringVar(&backupOptions.Host, "hostname", "", "set the `hostname` for the snapshot manually")
|
f.StringVar(&backupOptions.Host, "hostname", "", "set the `hostname` for the snapshot manually")
|
||||||
err := f.MarkDeprecated("hostname", "use --host")
|
err := f.MarkDeprecated("hostname", "use --host")
|
||||||
|
@ -132,7 +134,6 @@ func init() {
|
||||||
// MarkDeprecated only returns an error when the flag could not be found
|
// MarkDeprecated only returns an error when the flag could not be found
|
||||||
panic(err)
|
panic(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
f.StringArrayVar(&backupOptions.FilesFrom, "files-from", nil, "read the files to backup from `file` (can be combined with file args; can be specified multiple times)")
|
f.StringArrayVar(&backupOptions.FilesFrom, "files-from", nil, "read the files to backup from `file` (can be combined with file args; can be specified multiple times)")
|
||||||
f.StringArrayVar(&backupOptions.FilesFromVerbatim, "files-from-verbatim", nil, "read the files to backup from `file` (can be combined with file args; can be specified multiple times)")
|
f.StringArrayVar(&backupOptions.FilesFromVerbatim, "files-from-verbatim", nil, "read the files to backup from `file` (can be combined with file args; can be specified multiple times)")
|
||||||
f.StringArrayVar(&backupOptions.FilesFromRaw, "files-from-raw", nil, "read the files to backup from `file` (can be combined with file args; can be specified multiple times)")
|
f.StringArrayVar(&backupOptions.FilesFromRaw, "files-from-raw", nil, "read the files to backup from `file` (can be combined with file args; can be specified multiple times)")
|
||||||
|
@ -144,6 +145,10 @@ func init() {
|
||||||
if runtime.GOOS == "windows" {
|
if runtime.GOOS == "windows" {
|
||||||
f.BoolVar(&backupOptions.UseFsSnapshot, "use-fs-snapshot", false, "use filesystem snapshot where possible (currently only Windows VSS)")
|
f.BoolVar(&backupOptions.UseFsSnapshot, "use-fs-snapshot", false, "use filesystem snapshot where possible (currently only Windows VSS)")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// parse read concurrency from env, on error the default value will be used
|
||||||
|
readConcurrency, _ := strconv.ParseUint(os.Getenv("RESTIC_READ_CONCURRENCY"), 10, 32)
|
||||||
|
backupOptions.ReadConcurrency = uint(readConcurrency)
|
||||||
}
|
}
|
||||||
|
|
||||||
// filterExisting returns a slice of all existing items, or an error if no
|
// filterExisting returns a slice of all existing items, or an error if no
|
||||||
|
@ -685,7 +690,7 @@ func runBackup(opts BackupOptions, gopts GlobalOptions, term *termstatus.Termina
|
||||||
}
|
}
|
||||||
wg.Go(func() error { return sc.Scan(cancelCtx, targets) })
|
wg.Go(func() error { return sc.Scan(cancelCtx, targets) })
|
||||||
|
|
||||||
arch := archiver.New(repo, targetFS, archiver.Options{})
|
arch := archiver.New(repo, targetFS, archiver.Options{ReadConcurrency: backupOptions.ReadConcurrency})
|
||||||
arch.SelectByName = selectByNameFilter
|
arch.SelectByName = selectByNameFilter
|
||||||
arch.Select = selectFilter
|
arch.Select = selectFilter
|
||||||
arch.WithAtime = opts.WithAtime
|
arch.WithAtime = opts.WithAtime
|
||||||
|
|
|
@ -555,6 +555,7 @@ environment variables. The following lists these environment variables:
|
||||||
RESTIC_COMPRESSION Compression mode (only available for repository format version 2)
|
RESTIC_COMPRESSION Compression mode (only available for repository format version 2)
|
||||||
RESTIC_PROGRESS_FPS Frames per second by which the progress bar is updated
|
RESTIC_PROGRESS_FPS Frames per second by which the progress bar is updated
|
||||||
RESTIC_PACK_SIZE Target size for pack files
|
RESTIC_PACK_SIZE Target size for pack files
|
||||||
|
RESTIC_READ_CONCURRENCY Concurrency for file reads
|
||||||
|
|
||||||
TMPDIR Location for temporary files
|
TMPDIR Location for temporary files
|
||||||
|
|
||||||
|
|
|
@ -51,6 +51,16 @@ only applied for the single run of restic. The option can also be set via the en
|
||||||
variable ``RESTIC_COMPRESSION``.
|
variable ``RESTIC_COMPRESSION``.
|
||||||
|
|
||||||
|
|
||||||
|
File Read Concurrency
|
||||||
|
=====================
|
||||||
|
|
||||||
|
When backing up fast storage like NVME disks, it can be beneficial to increase the read
|
||||||
|
concurrency. This can increase the overall performance of the backup operation by reading
|
||||||
|
more files in parallel. You can specify the concurrency of file reads with the
|
||||||
|
``RESTIC_READ_CONCURRENCY`` environment variable or the ``--read-concurrency`` flag for
|
||||||
|
the ``backup`` command.
|
||||||
|
|
||||||
|
|
||||||
Pack Size
|
Pack Size
|
||||||
=========
|
=========
|
||||||
|
|
||||||
|
|
|
@ -112,6 +112,7 @@ command:
|
||||||
--ignore-inode ignore inode number changes when checking for modified files
|
--ignore-inode ignore inode number changes when checking for modified files
|
||||||
-x, --one-file-system exclude other file systems, don't cross filesystem boundaries and subvolumes
|
-x, --one-file-system exclude other file systems, don't cross filesystem boundaries and subvolumes
|
||||||
--parent snapshot use this parent snapshot (default: last snapshot in the repository that has the same target files/directories, and is not newer than the snapshot time)
|
--parent snapshot use this parent snapshot (default: last snapshot in the repository that has the same target files/directories, and is not newer than the snapshot time)
|
||||||
|
--read-concurrency n read n file concurrently. (default: $RESTIC_READ_CONCURRENCY or 2)
|
||||||
--stdin read backup from stdin
|
--stdin read backup from stdin
|
||||||
--stdin-filename filename filename to use when reading from stdin (default "stdin")
|
--stdin-filename filename filename to use when reading from stdin (default "stdin")
|
||||||
--tag tags add tags for the new snapshot in the format `tag[,tag,...]` (can be specified multiple times) (default [])
|
--tag tags add tags for the new snapshot in the format `tag[,tag,...]` (can be specified multiple times) (default [])
|
||||||
|
@ -442,3 +443,4 @@ time it is used, so by looking at the timestamps of the sub directories of the
|
||||||
cache directory it can decide which sub directories are old and probably not
|
cache directory it can decide which sub directories are old and probably not
|
||||||
needed any more. You can either remove these directories manually, or run a
|
needed any more. You can either remove these directories manually, or run a
|
||||||
restic command with the ``--cleanup-cache`` flag.
|
restic command with the ``--cleanup-cache`` flag.
|
||||||
|
|
||||||
|
|
|
@ -95,10 +95,10 @@ const (
|
||||||
|
|
||||||
// Options is used to configure the archiver.
|
// Options is used to configure the archiver.
|
||||||
type Options struct {
|
type Options struct {
|
||||||
// FileReadConcurrency sets how many files are read in concurrently. If
|
// ReadConcurrency sets how many files are read in concurrently. If
|
||||||
// it's set to zero, at most two files are read in concurrently (which
|
// it's set to zero, at most two files are read in concurrently (which
|
||||||
// turned out to be a good default for most situations).
|
// turned out to be a good default for most situations).
|
||||||
FileReadConcurrency uint
|
ReadConcurrency uint
|
||||||
|
|
||||||
// SaveBlobConcurrency sets how many blobs are hashed and saved
|
// SaveBlobConcurrency sets how many blobs are hashed and saved
|
||||||
// concurrently. If it's set to zero, the default is the number of CPUs
|
// concurrently. If it's set to zero, the default is the number of CPUs
|
||||||
|
@ -113,11 +113,11 @@ type Options struct {
|
||||||
// ApplyDefaults returns a copy of o with the default options set for all unset
|
// ApplyDefaults returns a copy of o with the default options set for all unset
|
||||||
// fields.
|
// fields.
|
||||||
func (o Options) ApplyDefaults() Options {
|
func (o Options) ApplyDefaults() Options {
|
||||||
if o.FileReadConcurrency == 0 {
|
if o.ReadConcurrency == 0 {
|
||||||
// two is a sweet spot for almost all situations. We've done some
|
// two is a sweet spot for almost all situations. We've done some
|
||||||
// experiments documented here:
|
// experiments documented here:
|
||||||
// https://github.com/borgbackup/borg/issues/3500
|
// https://github.com/borgbackup/borg/issues/3500
|
||||||
o.FileReadConcurrency = 2
|
o.ReadConcurrency = 2
|
||||||
}
|
}
|
||||||
|
|
||||||
if o.SaveBlobConcurrency == 0 {
|
if o.SaveBlobConcurrency == 0 {
|
||||||
|
@ -132,7 +132,7 @@ func (o Options) ApplyDefaults() Options {
|
||||||
// Also allow waiting for FileReadConcurrency files, this is the maximum of FutureFiles
|
// Also allow waiting for FileReadConcurrency files, this is the maximum of FutureFiles
|
||||||
// which currently can be in progress. The main backup loop blocks when trying to queue
|
// which currently can be in progress. The main backup loop blocks when trying to queue
|
||||||
// more files to read.
|
// more files to read.
|
||||||
o.SaveTreeConcurrency = uint(runtime.GOMAXPROCS(0)) + o.FileReadConcurrency
|
o.SaveTreeConcurrency = uint(runtime.GOMAXPROCS(0)) + o.ReadConcurrency
|
||||||
}
|
}
|
||||||
|
|
||||||
return o
|
return o
|
||||||
|
@ -782,7 +782,7 @@ func (arch *Archiver) runWorkers(ctx context.Context, wg *errgroup.Group) {
|
||||||
arch.fileSaver = NewFileSaver(ctx, wg,
|
arch.fileSaver = NewFileSaver(ctx, wg,
|
||||||
arch.blobSaver.Save,
|
arch.blobSaver.Save,
|
||||||
arch.Repo.Config().ChunkerPolynomial,
|
arch.Repo.Config().ChunkerPolynomial,
|
||||||
arch.Options.FileReadConcurrency, arch.Options.SaveBlobConcurrency)
|
arch.Options.ReadConcurrency, arch.Options.SaveBlobConcurrency)
|
||||||
arch.fileSaver.CompleteBlob = arch.CompleteBlob
|
arch.fileSaver.CompleteBlob = arch.CompleteBlob
|
||||||
arch.fileSaver.NodeFromFileInfo = arch.nodeFromFileInfo
|
arch.fileSaver.NodeFromFileInfo = arch.nodeFromFileInfo
|
||||||
|
|
||||||
|
|
|
@ -2040,7 +2040,7 @@ func TestArchiverAbortEarlyOnError(t *testing.T) {
|
||||||
|
|
||||||
// at most two files may be queued
|
// at most two files may be queued
|
||||||
arch := New(testRepo, testFS, Options{
|
arch := New(testRepo, testFS, Options{
|
||||||
FileReadConcurrency: 2,
|
ReadConcurrency: 2,
|
||||||
})
|
})
|
||||||
|
|
||||||
_, _, err := arch.Snapshot(ctx, []string{"."}, SnapshotOptions{Time: time.Now()})
|
_, _, err := arch.Snapshot(ctx, []string{"."}, SnapshotOptions{Time: time.Now()})
|
||||||
|
|
Loading…
Reference in a new issue