doc: fix typos throughout docs
This commit is contained in:
parent
bcbfad1482
commit
fb169a8b54
32 changed files with 66 additions and 66 deletions
|
@ -56,7 +56,7 @@ func init() {
|
||||||
Help: "QingStor Secret Access Key (password)\nLeave blank for anonymous access or runtime credentials.",
|
Help: "QingStor Secret Access Key (password)\nLeave blank for anonymous access or runtime credentials.",
|
||||||
}, {
|
}, {
|
||||||
Name: "endpoint",
|
Name: "endpoint",
|
||||||
Help: "Enter a endpoint URL to connection QingStor API.\nLeave blank will use the default value \"https://qingstor.com:443\"",
|
Help: "Enter an endpoint URL to connection QingStor API.\nLeave blank will use the default value \"https://qingstor.com:443\"",
|
||||||
}, {
|
}, {
|
||||||
Name: "zone",
|
Name: "zone",
|
||||||
Help: "Zone to connect to.\nDefault is \"pek3a\".",
|
Help: "Zone to connect to.\nDefault is \"pek3a\".",
|
||||||
|
@ -206,7 +206,7 @@ func (o *Object) split() (bucket, bucketPath string) {
|
||||||
// Split an URL into three parts: protocol host and port
|
// Split an URL into three parts: protocol host and port
|
||||||
func qsParseEndpoint(endpoint string) (protocol, host, port string, err error) {
|
func qsParseEndpoint(endpoint string) (protocol, host, port string, err error) {
|
||||||
/*
|
/*
|
||||||
Pattern to match a endpoint,
|
Pattern to match an endpoint,
|
||||||
eg: "http(s)://qingstor.com:443" --> "http(s)", "qingstor.com", 443
|
eg: "http(s)://qingstor.com:443" --> "http(s)", "qingstor.com", 443
|
||||||
"http(s)//qingstor.com" --> "http(s)", "qingstor.com", ""
|
"http(s)//qingstor.com" --> "http(s)", "qingstor.com", ""
|
||||||
"qingstor.com" --> "", "qingstor.com", ""
|
"qingstor.com" --> "", "qingstor.com", ""
|
||||||
|
|
|
@ -25,7 +25,7 @@ func init() {
|
||||||
|
|
||||||
var commandDefinition = &cobra.Command{
|
var commandDefinition = &cobra.Command{
|
||||||
Use: "hashsum <hash> remote:path",
|
Use: "hashsum <hash> remote:path",
|
||||||
Short: `Produces an hashsum file for all the objects in the path.`,
|
Short: `Produces a hashsum file for all the objects in the path.`,
|
||||||
Long: `
|
Long: `
|
||||||
Produces a hash file for all the objects in the path using the hash
|
Produces a hash file for all the objects in the path using the hash
|
||||||
named. The output is in the same format as the standard
|
named. The output is in the same format as the standard
|
||||||
|
|
|
@ -33,7 +33,7 @@ the remote does not support a server side directory move operation.
|
||||||
|
|
||||||
If no filters are in use and if possible this will server side move
|
If no filters are in use and if possible this will server side move
|
||||||
` + "`source:path`" + ` into ` + "`dest:path`" + `. After this ` + "`source:path`" + ` will no
|
` + "`source:path`" + ` into ` + "`dest:path`" + `. After this ` + "`source:path`" + ` will no
|
||||||
longer longer exist.
|
longer exist.
|
||||||
|
|
||||||
Otherwise for each file in ` + "`source:path`" + ` selected by the filters (if
|
Otherwise for each file in ` + "`source:path`" + ` selected by the filters (if
|
||||||
any) this will move it into ` + "`dest:path`" + `. If possible a server side
|
any) this will move it into ` + "`dest:path`" + `. If possible a server side
|
||||||
|
|
|
@ -103,7 +103,7 @@ https. You will need to supply the --cert and --key flags. If you
|
||||||
wish to do client side certificate validation then you will need to
|
wish to do client side certificate validation then you will need to
|
||||||
supply --client-ca also.
|
supply --client-ca also.
|
||||||
|
|
||||||
--cert should be a either a PEM encoded certificate or a concatenation
|
--cert should be either a PEM encoded certificate or a concatenation
|
||||||
of that with the CA certificate. --key should be the PEM encoded
|
of that with the CA certificate. --key should be the PEM encoded
|
||||||
private key and --client-ca should be the PEM encoded client
|
private key and --client-ca should be the PEM encoded client
|
||||||
certificate authority certificate.
|
certificate authority certificate.
|
||||||
|
|
|
@ -15,7 +15,7 @@ eg `remote:directory/subdirectory` or `/directory/subdirectory`.
|
||||||
During the initial setup with `rclone config` you will specify the target
|
During the initial setup with `rclone config` you will specify the target
|
||||||
remote. The target remote can either be a local path or another remote.
|
remote. The target remote can either be a local path or another remote.
|
||||||
|
|
||||||
Subfolders can be used in target remote. Assume a alias remote named `backup`
|
Subfolders can be used in target remote. Assume an alias remote named `backup`
|
||||||
with the target `mydrive:private/backup`. Invoking `rclone mkdir backup:desktop`
|
with the target `mydrive:private/backup`. Invoking `rclone mkdir backup:desktop`
|
||||||
is exactly the same as invoking `rclone mkdir mydrive:private/backup/desktop`.
|
is exactly the same as invoking `rclone mkdir mydrive:private/backup/desktop`.
|
||||||
|
|
||||||
|
@ -25,7 +25,7 @@ Invoking `rclone mkdir backup:../desktop` is exactly the same as invoking
|
||||||
The empty path is not allowed as a remote. To alias the current directory
|
The empty path is not allowed as a remote. To alias the current directory
|
||||||
use `.` instead.
|
use `.` instead.
|
||||||
|
|
||||||
Here is an example of how to make a alias called `remote` for local folder.
|
Here is an example of how to make an alias called `remote` for local folder.
|
||||||
First run:
|
First run:
|
||||||
|
|
||||||
rclone config
|
rclone config
|
||||||
|
|
|
@ -38,7 +38,7 @@ which pass through it.
|
||||||
|
|
||||||
Since rclone doesn't currently have its own Amazon Drive credentials
|
Since rclone doesn't currently have its own Amazon Drive credentials
|
||||||
so you will either need to have your own `client_id` and
|
so you will either need to have your own `client_id` and
|
||||||
`client_secret` with Amazon Drive, or use a a third party ouath proxy
|
`client_secret` with Amazon Drive, or use a third party oauth proxy
|
||||||
in which case you will need to enter `client_id`, `client_secret`,
|
in which case you will need to enter `client_id`, `client_secret`,
|
||||||
`auth_url` and `token_url`.
|
`auth_url` and `token_url`.
|
||||||
|
|
||||||
|
|
|
@ -301,7 +301,7 @@ MD5 sums are only uploaded with chunked files if the source has an MD5
|
||||||
sum. This will always be the case for a local to azure copy.
|
sum. This will always be the case for a local to azure copy.
|
||||||
|
|
||||||
### Azure Storage Emulator Support ###
|
### Azure Storage Emulator Support ###
|
||||||
You can test rlcone with storage emulator locally, to do this make sure azure storage emulator
|
You can test rclone with storage emulator locally, to do this make sure azure storage emulator
|
||||||
installed locally and set up a new remote with `rclone config` follow instructions described in
|
installed locally and set up a new remote with `rclone config` follow instructions described in
|
||||||
introduction, set `use_emulator` config as `true`, you do not need to provide default account name
|
introduction, set `use_emulator` config as `true`, you do not need to provide default account name
|
||||||
or key if using emulator.
|
or key if using emulator.
|
||||||
|
|
|
@ -185,10 +185,10 @@ Affected settings:
|
||||||
##### Certificate Validation #####
|
##### Certificate Validation #####
|
||||||
|
|
||||||
When the Plex server is configured to only accept secure connections, it is
|
When the Plex server is configured to only accept secure connections, it is
|
||||||
possible to use `.plex.direct` URL's to ensure certificate validation succeeds.
|
possible to use `.plex.direct` URLs to ensure certificate validation succeeds.
|
||||||
These URL's are used by Plex internally to connect to the Plex server securely.
|
These URLs are used by Plex internally to connect to the Plex server securely.
|
||||||
|
|
||||||
The format for this URL's is the following:
|
The format for these URLs is the following:
|
||||||
|
|
||||||
https://ip-with-dots-replaced.server-hash.plex.direct:32400/
|
https://ip-with-dots-replaced.server-hash.plex.direct:32400/
|
||||||
|
|
||||||
|
@ -241,7 +241,7 @@ there is a valid concern that the expiring cache listings can lead to cloud prov
|
||||||
throttles or bans due to repeated queries on it for very large mounts.
|
throttles or bans due to repeated queries on it for very large mounts.
|
||||||
|
|
||||||
Some recommendations:
|
Some recommendations:
|
||||||
- don't use a very small interval for entry informations (`--cache-info-age`)
|
- don't use a very small interval for entry information (`--cache-info-age`)
|
||||||
- while writes aren't yet optimised, you can still write through `cache` which gives you the advantage
|
- while writes aren't yet optimised, you can still write through `cache` which gives you the advantage
|
||||||
of adding the file in the cache at the same time if configured to do so.
|
of adding the file in the cache at the same time if configured to do so.
|
||||||
|
|
||||||
|
@ -405,7 +405,7 @@ The plex token for authentication - auto set normally
|
||||||
|
|
||||||
#### --cache-plex-insecure
|
#### --cache-plex-insecure
|
||||||
|
|
||||||
Skip all certificate verifications when connecting to the Plex server
|
Skip all certificate verification when connecting to the Plex server
|
||||||
|
|
||||||
- Config: plex_insecure
|
- Config: plex_insecure
|
||||||
- Env Var: RCLONE_CACHE_PLEX_INSECURE
|
- Env Var: RCLONE_CACHE_PLEX_INSECURE
|
||||||
|
|
|
@ -189,7 +189,7 @@ date: "2020-02-01"
|
||||||
* copyurl
|
* copyurl
|
||||||
* Add `--auto-filename` flag for using file name from URL in destination path (Denis)
|
* Add `--auto-filename` flag for using file name from URL in destination path (Denis)
|
||||||
* serve dlna:
|
* serve dlna:
|
||||||
* Many compatability improvements (Dan Walters)
|
* Many compatibility improvements (Dan Walters)
|
||||||
* Support for external srt subtitles (Dan Walters)
|
* Support for external srt subtitles (Dan Walters)
|
||||||
* rc
|
* rc
|
||||||
* Added command core/quit (Saksham Khanna)
|
* Added command core/quit (Saksham Khanna)
|
||||||
|
@ -220,7 +220,7 @@ date: "2020-02-01"
|
||||||
* filter: Prevent mixing options when `--files-from` is in use (Michele Caci)
|
* filter: Prevent mixing options when `--files-from` is in use (Michele Caci)
|
||||||
* serve sftp: Fix crash on unsupported operations (eg Readlink) (Nick Craig-Wood)
|
* serve sftp: Fix crash on unsupported operations (eg Readlink) (Nick Craig-Wood)
|
||||||
* Mount
|
* Mount
|
||||||
* Allow files of unkown size to be read properly (Nick Craig-Wood)
|
* Allow files of unknown size to be read properly (Nick Craig-Wood)
|
||||||
* Skip tests on <= 2 CPUs to avoid lockup (Nick Craig-Wood)
|
* Skip tests on <= 2 CPUs to avoid lockup (Nick Craig-Wood)
|
||||||
* Fix panic on File.Open (Nick Craig-Wood)
|
* Fix panic on File.Open (Nick Craig-Wood)
|
||||||
* Fix "mount_fusefs: -o timeout=: option not supported" on FreeBSD (Nick Craig-Wood)
|
* Fix "mount_fusefs: -o timeout=: option not supported" on FreeBSD (Nick Craig-Wood)
|
||||||
|
@ -259,7 +259,7 @@ date: "2020-02-01"
|
||||||
* Fix signature v2_auth headers (Anthony Rusdi)
|
* Fix signature v2_auth headers (Anthony Rusdi)
|
||||||
* Fix encoding for control characters (Nick Craig-Wood)
|
* Fix encoding for control characters (Nick Craig-Wood)
|
||||||
* Only ask for URL encoded directory listings if we need them on Ceph (Nick Craig-Wood)
|
* Only ask for URL encoded directory listings if we need them on Ceph (Nick Craig-Wood)
|
||||||
* Add option for multipart failiure behaviour (Aleksandar Jankovic)
|
* Add option for multipart failure behaviour (Aleksandar Jankovic)
|
||||||
* Support for multipart copy (庄天翼)
|
* Support for multipart copy (庄天翼)
|
||||||
* Fix nil pointer reference if no metadata returned for object (Nick Craig-Wood)
|
* Fix nil pointer reference if no metadata returned for object (Nick Craig-Wood)
|
||||||
* SFTP
|
* SFTP
|
||||||
|
@ -695,7 +695,7 @@ date: "2020-02-01"
|
||||||
* Jottacloud
|
* Jottacloud
|
||||||
* Resume and deduplication support (Oliver Heyme)
|
* Resume and deduplication support (Oliver Heyme)
|
||||||
* Use token auth for all API requests Don't store password anymore (Sebastian Bünger)
|
* Use token auth for all API requests Don't store password anymore (Sebastian Bünger)
|
||||||
* Add support for 2-factor authentification (Sebastian Bünger)
|
* Add support for 2-factor authentication (Sebastian Bünger)
|
||||||
* Mega
|
* Mega
|
||||||
* Implement v2 account login which fixes logins for newer Mega accounts (Nick Craig-Wood)
|
* Implement v2 account login which fixes logins for newer Mega accounts (Nick Craig-Wood)
|
||||||
* Return error if an unknown length file is attempted to be uploaded (Nick Craig-Wood)
|
* Return error if an unknown length file is attempted to be uploaded (Nick Craig-Wood)
|
||||||
|
@ -710,7 +710,7 @@ date: "2020-02-01"
|
||||||
* Default `--qingstor-upload-concurrency` to 1 to work around bug (Nick Craig-Wood)
|
* Default `--qingstor-upload-concurrency` to 1 to work around bug (Nick Craig-Wood)
|
||||||
* S3
|
* S3
|
||||||
* Implement `--s3-upload-cutoff` for single part uploads below this (Nick Craig-Wood)
|
* Implement `--s3-upload-cutoff` for single part uploads below this (Nick Craig-Wood)
|
||||||
* Change `--s3-upload-concurrency` default to 4 to increase perfomance (Nick Craig-Wood)
|
* Change `--s3-upload-concurrency` default to 4 to increase performance (Nick Craig-Wood)
|
||||||
* Add `--s3-bucket-acl` to control bucket ACL (Nick Craig-Wood)
|
* Add `--s3-bucket-acl` to control bucket ACL (Nick Craig-Wood)
|
||||||
* Auto detect region for buckets on operation failure (Nick Craig-Wood)
|
* Auto detect region for buckets on operation failure (Nick Craig-Wood)
|
||||||
* Add GLACIER storage class (William Cocker)
|
* Add GLACIER storage class (William Cocker)
|
||||||
|
@ -892,7 +892,7 @@ date: "2020-02-01"
|
||||||
* Add link sharing support (jackyzy823)
|
* Add link sharing support (jackyzy823)
|
||||||
* S3
|
* S3
|
||||||
* Use custom pacer, to retry operations when reasonable (Craig Miskell)
|
* Use custom pacer, to retry operations when reasonable (Craig Miskell)
|
||||||
* Use configured server-side-encryption and storace class options when calling CopyObject() (Paul Kohout)
|
* Use configured server-side-encryption and storage class options when calling CopyObject() (Paul Kohout)
|
||||||
* Make `--s3-v2-auth` flag (Nick Craig-Wood)
|
* Make `--s3-v2-auth` flag (Nick Craig-Wood)
|
||||||
* Fix v2 auth on files with spaces (Nick Craig-Wood)
|
* Fix v2 auth on files with spaces (Nick Craig-Wood)
|
||||||
* Union
|
* Union
|
||||||
|
@ -1018,8 +1018,8 @@ Point release to fix hubic and azureblob backends.
|
||||||
* Remove leading / from paths (Nick Craig-Wood)
|
* Remove leading / from paths (Nick Craig-Wood)
|
||||||
* Swift
|
* Swift
|
||||||
* Add `storage_policy` (Ruben Vandamme)
|
* Add `storage_policy` (Ruben Vandamme)
|
||||||
* Make it so just `storage_url` or `auth_token` can be overidden (Nick Craig-Wood)
|
* Make it so just `storage_url` or `auth_token` can be overridden (Nick Craig-Wood)
|
||||||
* Fix server side copy bug for unusal file names (Nick Craig-Wood)
|
* Fix server side copy bug for unusual file names (Nick Craig-Wood)
|
||||||
* Remove leading / from paths (Nick Craig-Wood)
|
* Remove leading / from paths (Nick Craig-Wood)
|
||||||
* WebDAV
|
* WebDAV
|
||||||
* Ensure we call MKCOL with a URL with a trailing / for QNAP interop (Nick Craig-Wood)
|
* Ensure we call MKCOL with a URL with a trailing / for QNAP interop (Nick Craig-Wood)
|
||||||
|
@ -1519,7 +1519,7 @@ Point release to fix hubic and azureblob backends.
|
||||||
* --old-sync-method deprecated - the remaining uses are covered by --fast-list
|
* --old-sync-method deprecated - the remaining uses are covered by --fast-list
|
||||||
* This involved a major re-write of all the listing code
|
* This involved a major re-write of all the listing code
|
||||||
* Add --tpslimit and --tpslimit-burst to limit transactions per second
|
* Add --tpslimit and --tpslimit-burst to limit transactions per second
|
||||||
* this is useful in conjuction with `rclone mount` to limit external apps
|
* this is useful in conjunction with `rclone mount` to limit external apps
|
||||||
* Add --stats-log-level so can see --stats without -v
|
* Add --stats-log-level so can see --stats without -v
|
||||||
* Print password prompts to stderr - Hraban Luyat
|
* Print password prompts to stderr - Hraban Luyat
|
||||||
* Warn about duplicate files when syncing
|
* Warn about duplicate files when syncing
|
||||||
|
@ -1853,7 +1853,7 @@ Point release to fix hubic and azureblob backends.
|
||||||
* Add ap-northeast-2 (Seoul) and ap-south-1 (Mumbai) regions.
|
* Add ap-northeast-2 (Seoul) and ap-south-1 (Mumbai) regions.
|
||||||
* Skip setting the modified time for objects > 5GB as it isn't possible.
|
* Skip setting the modified time for objects > 5GB as it isn't possible.
|
||||||
* Backblaze B2
|
* Backblaze B2
|
||||||
* Add --b2-versions flag so old versions can be listed and retreived.
|
* Add --b2-versions flag so old versions can be listed and retrieved.
|
||||||
* Treat 403 errors (eg cap exceeded) as fatal.
|
* Treat 403 errors (eg cap exceeded) as fatal.
|
||||||
* Implement cleanup command for deleting old file versions.
|
* Implement cleanup command for deleting old file versions.
|
||||||
* Make error handling compliant with B2 integrations notes.
|
* Make error handling compliant with B2 integrations notes.
|
||||||
|
@ -1979,7 +1979,7 @@ Point release to fix hubic and azureblob backends.
|
||||||
* Add `--delete-before`, `--delete-during`, `--delete-after` flags.
|
* Add `--delete-before`, `--delete-during`, `--delete-after` flags.
|
||||||
* Add `--memprofile` flag to debug memory use.
|
* Add `--memprofile` flag to debug memory use.
|
||||||
* Warn the user about files with same name but different case
|
* Warn the user about files with same name but different case
|
||||||
* Make `--include` rules add their implict exclude * at the end of the filter list
|
* Make `--include` rules add their implicit exclude * at the end of the filter list
|
||||||
* Deprecate compiling with go1.3
|
* Deprecate compiling with go1.3
|
||||||
* Amazon Drive
|
* Amazon Drive
|
||||||
* Fix download of files > 10 GB
|
* Fix download of files > 10 GB
|
||||||
|
|
|
@ -212,7 +212,7 @@ guarantee given hash for all files. If wrapped remote doesn't support it,
|
||||||
chunker will then add metadata to all files, even small. However, this can
|
chunker will then add metadata to all files, even small. However, this can
|
||||||
double the amount of small files in storage and incur additional service charges.
|
double the amount of small files in storage and incur additional service charges.
|
||||||
You can even use chunker to force md5/sha1 support in any other remote
|
You can even use chunker to force md5/sha1 support in any other remote
|
||||||
at expence of sidecar meta objects by setting eg. `chunk_type=sha1all`
|
at expense of sidecar meta objects by setting eg. `chunk_type=sha1all`
|
||||||
to force hashsums and `chunk_size=1P` to effectively disable chunking.
|
to force hashsums and `chunk_size=1P` to effectively disable chunking.
|
||||||
|
|
||||||
Normally, when a file is copied to chunker controlled remote, chunker
|
Normally, when a file is copied to chunker controlled remote, chunker
|
||||||
|
@ -293,7 +293,7 @@ Chunker will not automatically rename existing chunks when you run
|
||||||
Beware that in result of this some files which have been treated as chunks
|
Beware that in result of this some files which have been treated as chunks
|
||||||
before the change can pop up in directory listings as normal files
|
before the change can pop up in directory listings as normal files
|
||||||
and vice versa. The same warning holds for the chunk size.
|
and vice versa. The same warning holds for the chunk size.
|
||||||
If you desperately need to change critical chunking setings, you should
|
If you desperately need to change critical chunking settings, you should
|
||||||
run data migration as described above.
|
run data migration as described above.
|
||||||
|
|
||||||
If wrapped remote is case insensitive, the chunker overlay will inherit
|
If wrapped remote is case insensitive, the chunker overlay will inherit
|
||||||
|
|
|
@ -113,7 +113,7 @@ due to the different salt.
|
||||||
|
|
||||||
Note that rclone does not encrypt
|
Note that rclone does not encrypt
|
||||||
|
|
||||||
* file length - this can be calcuated within 16 bytes
|
* file length - this can be calculated within 16 bytes
|
||||||
* modification time - used for syncing
|
* modification time - used for syncing
|
||||||
|
|
||||||
## Specifying the remote ##
|
## Specifying the remote ##
|
||||||
|
@ -367,7 +367,7 @@ names, or for debugging purposes.
|
||||||
|
|
||||||
## Backing up a crypted remote ##
|
## Backing up a crypted remote ##
|
||||||
|
|
||||||
If you wish to backup a crypted remote, it it recommended that you use
|
If you wish to backup a crypted remote, it is recommended that you use
|
||||||
`rclone sync` on the encrypted files, and make sure the passwords are
|
`rclone sync` on the encrypted files, and make sure the passwords are
|
||||||
the same in the new encrypted remote.
|
the same in the new encrypted remote.
|
||||||
|
|
||||||
|
@ -451,7 +451,7 @@ files.
|
||||||
File names are encrypted segment by segment - the path is broken up
|
File names are encrypted segment by segment - the path is broken up
|
||||||
into `/` separated strings and these are encrypted individually.
|
into `/` separated strings and these are encrypted individually.
|
||||||
|
|
||||||
File segments are padded using using PKCS#7 to a multiple of 16 bytes
|
File segments are padded using PKCS#7 to a multiple of 16 bytes
|
||||||
before encryption.
|
before encryption.
|
||||||
|
|
||||||
They are then encrypted with EME using AES with 256 bit key. EME
|
They are then encrypted with EME using AES with 256 bit key. EME
|
||||||
|
|
|
@ -522,7 +522,7 @@ to disable server side move and server side copy use:
|
||||||
|
|
||||||
--disable move,copy
|
--disable move,copy
|
||||||
|
|
||||||
The features can be put in in any case.
|
The features can be put in any case.
|
||||||
|
|
||||||
To see a list of which features can be disabled use:
|
To see a list of which features can be disabled use:
|
||||||
|
|
||||||
|
@ -794,13 +794,13 @@ Rclone will exit with exit code 8 if the transfer limit is reached.
|
||||||
This modifies the behavior of `--max-transfer`
|
This modifies the behavior of `--max-transfer`
|
||||||
Defaults to `--cutoff-mode=hard`.
|
Defaults to `--cutoff-mode=hard`.
|
||||||
|
|
||||||
Specifiying `--cutoff-mode=hard` will stop transferring immediately
|
Specifying `--cutoff-mode=hard` will stop transferring immediately
|
||||||
when Rclone reaches the limit.
|
when Rclone reaches the limit.
|
||||||
|
|
||||||
Specifiying `--cutoff-mode=soft` will stop starting new transfers
|
Specifying `--cutoff-mode=soft` will stop starting new transfers
|
||||||
when Rclone reaches the limit.
|
when Rclone reaches the limit.
|
||||||
|
|
||||||
Specifiying `--cutoff-mode=cautious` will try to prevent Rclone
|
Specifying `--cutoff-mode=cautious` will try to prevent Rclone
|
||||||
from reaching the limit.
|
from reaching the limit.
|
||||||
|
|
||||||
### --modify-window=TIME ###
|
### --modify-window=TIME ###
|
||||||
|
@ -827,7 +827,7 @@ time) then each thread writes directly into the file at the correct
|
||||||
place. This means that rclone won't create fragmented or sparse files
|
place. This means that rclone won't create fragmented or sparse files
|
||||||
and there won't be any assembly time at the end of the transfer.
|
and there won't be any assembly time at the end of the transfer.
|
||||||
|
|
||||||
The number of threads used to dowload is controlled by
|
The number of threads used to download is controlled by
|
||||||
`--multi-thread-streams`.
|
`--multi-thread-streams`.
|
||||||
|
|
||||||
Use `-vv` if you wish to see info about the threads.
|
Use `-vv` if you wish to see info about the threads.
|
||||||
|
@ -1230,7 +1230,7 @@ Note also that `--track-renames` is incompatible with
|
||||||
### --track-renames-strategy (hash,modtime) ###
|
### --track-renames-strategy (hash,modtime) ###
|
||||||
|
|
||||||
This option changes the matching criteria for `--track-renames` to match
|
This option changes the matching criteria for `--track-renames` to match
|
||||||
by any combination of modtime, hash, size. Matchig by size is always enabled
|
by any combination of modtime, hash, size. Matching by size is always enabled
|
||||||
no matter what option is selected here. This also means
|
no matter what option is selected here. This also means
|
||||||
that it enables `--track-renames` support for encrypted destinations.
|
that it enables `--track-renames` support for encrypted destinations.
|
||||||
If nothing is specified, the default option is matching by hashes.
|
If nothing is specified, the default option is matching by hashes.
|
||||||
|
@ -1377,7 +1377,7 @@ Prints the version number
|
||||||
SSL/TLS options
|
SSL/TLS options
|
||||||
---------------
|
---------------
|
||||||
|
|
||||||
The outoing SSL/TLS connections rclone makes can be controlled with
|
The outgoing SSL/TLS connections rclone makes can be controlled with
|
||||||
these options. For example this can be very useful with the HTTP or
|
these options. For example this can be very useful with the HTTP or
|
||||||
WebDAV backends. Rclone HTTP servers have their own set of
|
WebDAV backends. Rclone HTTP servers have their own set of
|
||||||
configuration for SSL/TLS which you can find in their documentation.
|
configuration for SSL/TLS which you can find in their documentation.
|
||||||
|
|
|
@ -678,7 +678,7 @@ videos.
|
||||||
Setting this flag will cause Google photos and videos to return a
|
Setting this flag will cause Google photos and videos to return a
|
||||||
blank MD5 checksum.
|
blank MD5 checksum.
|
||||||
|
|
||||||
Google photos are identifed by being in the "photos" space.
|
Google photos are identified by being in the "photos" space.
|
||||||
|
|
||||||
Corrupted checksums are caused by Google modifying the image/video but
|
Corrupted checksums are caused by Google modifying the image/video but
|
||||||
not updating the checksum.
|
not updating the checksum.
|
||||||
|
@ -877,7 +877,7 @@ Keep new head revision of each file forever.
|
||||||
|
|
||||||
Show sizes as storage quota usage, not actual size.
|
Show sizes as storage quota usage, not actual size.
|
||||||
|
|
||||||
Show the size of a file as the the storage quota used. This is the
|
Show the size of a file as the storage quota used. This is the
|
||||||
current version plus any older versions that have been set to keep
|
current version plus any older versions that have been set to keep
|
||||||
forever.
|
forever.
|
||||||
|
|
||||||
|
@ -1028,7 +1028,7 @@ etc as rclone knows to ignore the size when doing the transfer.
|
||||||
However an unfortunate consequence of this is that you may not be able
|
However an unfortunate consequence of this is that you may not be able
|
||||||
to download Google docs using `rclone mount`. If it doesn't work you
|
to download Google docs using `rclone mount`. If it doesn't work you
|
||||||
will get a 0 sized file. If you try again the doc may gain its
|
will get a 0 sized file. If you try again the doc may gain its
|
||||||
correct size and be downloadable. Whther it will work on not depends
|
correct size and be downloadable. Whether it will work on not depends
|
||||||
on the application accessing the mount and the OS you are running -
|
on the application accessing the mount and the OS you are running -
|
||||||
experiment to find out if it does work for you!
|
experiment to find out if it does work for you!
|
||||||
|
|
||||||
|
|
|
@ -7,7 +7,7 @@ date: "2015-10-14"
|
||||||
<i class="fa fa-archive"></i> 1Fichier
|
<i class="fa fa-archive"></i> 1Fichier
|
||||||
-----------------------------------------
|
-----------------------------------------
|
||||||
|
|
||||||
This is a backend for the [1ficher](https://1fichier.com) cloud
|
This is a backend for the [1fichier](https://1fichier.com) cloud
|
||||||
storage service. Note that a Premium subscription is required to use
|
storage service. Note that a Premium subscription is required to use
|
||||||
the API.
|
the API.
|
||||||
|
|
||||||
|
|
|
@ -308,7 +308,7 @@ This reads a list of file names from the file passed in and **only**
|
||||||
these files are transferred. The **filtering rules are ignored**
|
these files are transferred. The **filtering rules are ignored**
|
||||||
completely if you use this option.
|
completely if you use this option.
|
||||||
|
|
||||||
`--files-from` expects a list of files as it's input. Leading / trailing
|
`--files-from` expects a list of files as its input. Leading / trailing
|
||||||
whitespace is stripped from the input lines and lines starting with `#`
|
whitespace is stripped from the input lines and lines starting with `#`
|
||||||
and `;` are ignored.
|
and `;` are ignored.
|
||||||
|
|
||||||
|
|
|
@ -344,7 +344,7 @@ and may be set in the config file.
|
||||||
--qingstor-chunk-size SizeSuffix Chunk size to use for uploading. (default 4M)
|
--qingstor-chunk-size SizeSuffix Chunk size to use for uploading. (default 4M)
|
||||||
--qingstor-connection-retries int Number of connection retries. (default 3)
|
--qingstor-connection-retries int Number of connection retries. (default 3)
|
||||||
--qingstor-encoding MultiEncoder This sets the encoding for the backend. (default Slash,Ctl,InvalidUtf8)
|
--qingstor-encoding MultiEncoder This sets the encoding for the backend. (default Slash,Ctl,InvalidUtf8)
|
||||||
--qingstor-endpoint string Enter a endpoint URL to connection QingStor API.
|
--qingstor-endpoint string Enter an endpoint URL to connection QingStor API.
|
||||||
--qingstor-env-auth Get QingStor credentials from runtime. Only applies if access_key_id and secret_access_key is blank.
|
--qingstor-env-auth Get QingStor credentials from runtime. Only applies if access_key_id and secret_access_key is blank.
|
||||||
--qingstor-secret-access-key string QingStor Secret Access Key (password)
|
--qingstor-secret-access-key string QingStor Secret Access Key (password)
|
||||||
--qingstor-upload-concurrency int Concurrency for multipart uploads. (default 1)
|
--qingstor-upload-concurrency int Concurrency for multipart uploads. (default 1)
|
||||||
|
|
|
@ -132,7 +132,7 @@ Note that not all FTP servers can have all characters in file names, for example
|
||||||
|
|
||||||
FTP supports implicit FTP over TLS servers (FTPS). This has to be enabled
|
FTP supports implicit FTP over TLS servers (FTPS). This has to be enabled
|
||||||
in the config for the remote. The default FTPS port is `990` so the
|
in the config for the remote. The default FTPS port is `990` so the
|
||||||
port will likely have to be explictly set in the config for the remote.
|
port will likely have to be explicitly set in the config for the remote.
|
||||||
|
|
||||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/ftp/ftp.go then run make backenddocs -->
|
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/ftp/ftp.go then run make backenddocs -->
|
||||||
### Standard Options
|
### Standard Options
|
||||||
|
|
|
@ -180,7 +180,7 @@ into albums.
|
||||||
```
|
```
|
||||||
|
|
||||||
There are two writable parts of the tree, the `upload` directory and
|
There are two writable parts of the tree, the `upload` directory and
|
||||||
sub directories of the the `album` directory.
|
sub directories of the `album` directory.
|
||||||
|
|
||||||
The `upload` directory is for uploading files you don't want to put
|
The `upload` directory is for uploading files you don't want to put
|
||||||
into albums. This will be empty to start with and will contain the
|
into albums. This will be empty to start with and will contain the
|
||||||
|
|
|
@ -73,7 +73,7 @@ The flag `--rc-web-gui` is shorthand for
|
||||||
- `--rc-pass <random password>`
|
- `--rc-pass <random password>`
|
||||||
- `--rc-serve`
|
- `--rc-serve`
|
||||||
|
|
||||||
These flags can be overidden as desired.
|
These flags can be overridden as desired.
|
||||||
|
|
||||||
See also the [rclone rcd documentation](https://rclone.org/commands/rclone_rcd/).
|
See also the [rclone rcd documentation](https://rclone.org/commands/rclone_rcd/).
|
||||||
|
|
||||||
|
@ -95,14 +95,14 @@ If you want to run the GUI behind a proxy at `/rclone` you could use these flags
|
||||||
- `--rc-baseurl rclone`
|
- `--rc-baseurl rclone`
|
||||||
- `--rc-htpasswd /path/to/htpasswd`
|
- `--rc-htpasswd /path/to/htpasswd`
|
||||||
|
|
||||||
Or instead of htpassword if you just want a single user and password:
|
Or instead of htpasswd if you just want a single user and password:
|
||||||
|
|
||||||
- `--rc-user me`
|
- `--rc-user me`
|
||||||
- `--rc-pass mypassword`
|
- `--rc-pass mypassword`
|
||||||
|
|
||||||
## Project
|
## Project
|
||||||
|
|
||||||
The GUI is being developed in the: [rclone/rclone-webui-react respository](https://github.com/rclone/rclone-webui-react).
|
The GUI is being developed in the: [rclone/rclone-webui-react repository](https://github.com/rclone/rclone-webui-react).
|
||||||
|
|
||||||
Bug reports and contributions are very welcome :-)
|
Bug reports and contributions are very welcome :-)
|
||||||
|
|
||||||
|
|
|
@ -9,7 +9,7 @@
|
||||||
|
|
||||||
set -e
|
set -e
|
||||||
|
|
||||||
#when adding a tool to the list make sure to also add it's corresponding command further in the script
|
#when adding a tool to the list make sure to also add its corresponding command further in the script
|
||||||
unzip_tools_list=('unzip' '7z' 'busybox')
|
unzip_tools_list=('unzip' '7z' 'busybox')
|
||||||
|
|
||||||
usage() { echo "Usage: curl https://rclone.org/install.sh | sudo bash [-s beta]" 1>&2; exit 1; }
|
usage() { echo "Usage: curl https://rclone.org/install.sh | sudo bash [-s beta]" 1>&2; exit 1; }
|
||||||
|
|
|
@ -161,7 +161,7 @@ as they can't be used in XML strings.
|
||||||
### Deleting files
|
### Deleting files
|
||||||
|
|
||||||
By default rclone will send all files to the trash when deleting files. They will be permanently
|
By default rclone will send all files to the trash when deleting files. They will be permanently
|
||||||
deleted automatically after 30 days. You may bypass the trash and permanently delete files immeditaly
|
deleted automatically after 30 days. You may bypass the trash and permanently delete files immediately
|
||||||
by using the [--jottacloud-hard-delete](#jottacloud-hard-delete) flag, or set the equivalent environment variable.
|
by using the [--jottacloud-hard-delete](#jottacloud-hard-delete) flag, or set the equivalent environment variable.
|
||||||
Emptying the trash is supported by the [cleanup](/commands/rclone_cleanup/) command.
|
Emptying the trash is supported by the [cleanup](/commands/rclone_cleanup/) command.
|
||||||
|
|
||||||
|
|
|
@ -428,7 +428,7 @@ See: the [encoding section in the overview](/overview/#encoding) for more info.
|
||||||
|
|
||||||
Here are the commands specific to the local backend.
|
Here are the commands specific to the local backend.
|
||||||
|
|
||||||
Run them with with
|
Run them with
|
||||||
|
|
||||||
rclone backend COMMAND remote:
|
rclone backend COMMAND remote:
|
||||||
|
|
||||||
|
|
|
@ -18,7 +18,7 @@ Currently it is recommended to disable 2FA on Mail.ru accounts intended for rclo
|
||||||
- Deleted files are by default moved to the trash
|
- Deleted files are by default moved to the trash
|
||||||
- Files and directories can be shared via public links
|
- Files and directories can be shared via public links
|
||||||
- Partial uploads or streaming are not supported, file size must be known before upload
|
- Partial uploads or streaming are not supported, file size must be known before upload
|
||||||
- Maximum file size is limited to 2G for a free acount, unlimited for paid accounts
|
- Maximum file size is limited to 2G for a free account, unlimited for paid accounts
|
||||||
- Storage keeps hash for all files and performs transparent deduplication,
|
- Storage keeps hash for all files and performs transparent deduplication,
|
||||||
the hash algorithm is a modified SHA1
|
the hash algorithm is a modified SHA1
|
||||||
- If a particular file is already present in storage, one can quickly submit file hash
|
- If a particular file is already present in storage, one can quickly submit file hash
|
||||||
|
|
|
@ -109,7 +109,7 @@ Use `rclone dedupe` to fix duplicated files.
|
||||||
|
|
||||||
Mega remotes seem to get blocked (reject logins) under "heavy use".
|
Mega remotes seem to get blocked (reject logins) under "heavy use".
|
||||||
We haven't worked out the exact blocking rules but it seems to be
|
We haven't worked out the exact blocking rules but it seems to be
|
||||||
related to fast paced, sucessive rclone commands.
|
related to fast paced, successive rclone commands.
|
||||||
|
|
||||||
For example, executing this command 90 times in a row `rclone link
|
For example, executing this command 90 times in a row `rclone link
|
||||||
remote:file` will cause the remote to become "blocked". This is not an
|
remote:file` will cause the remote to become "blocked". This is not an
|
||||||
|
@ -140,7 +140,7 @@ approach.
|
||||||
|
|
||||||
Note that once blocked, the use of other tools (such as megacmd) is
|
Note that once blocked, the use of other tools (such as megacmd) is
|
||||||
not a sure workaround: following megacmd login times have been
|
not a sure workaround: following megacmd login times have been
|
||||||
observed in sucession for blocked remote: 7 minutes, 20 min, 30min, 30
|
observed in succession for blocked remote: 7 minutes, 20 min, 30min, 30
|
||||||
min, 30min. Web access looks unaffected though.
|
min, 30min. Web access looks unaffected though.
|
||||||
|
|
||||||
Investigation is continuing in relation to workarounds based on
|
Investigation is continuing in relation to workarounds based on
|
||||||
|
|
|
@ -132,7 +132,7 @@ Client ID and Key by following the steps below:
|
||||||
2. Enter a name for your app, choose account type `Accounts in any organizational directory (Any Azure AD directory - Multitenant) and personal Microsoft accounts (e.g. Skype, Xbox)`, select `Web` in `Redirect URI` Enter `http://localhost:53682/` and click Register. Copy and keep the `Application (client) ID` under the app name for later use.
|
2. Enter a name for your app, choose account type `Accounts in any organizational directory (Any Azure AD directory - Multitenant) and personal Microsoft accounts (e.g. Skype, Xbox)`, select `Web` in `Redirect URI` Enter `http://localhost:53682/` and click Register. Copy and keep the `Application (client) ID` under the app name for later use.
|
||||||
3. Under `manage` select `Certificates & secrets`, click `New client secret`. Copy and keep that secret for later use.
|
3. Under `manage` select `Certificates & secrets`, click `New client secret`. Copy and keep that secret for later use.
|
||||||
4. Under `manage` select `API permissions`, click `Add a permission` and select `Microsoft Graph` then select `delegated permissions`.
|
4. Under `manage` select `API permissions`, click `Add a permission` and select `Microsoft Graph` then select `delegated permissions`.
|
||||||
5. Search and select the follwing permssions: `Files.Read`, `Files.ReadWrite`, `Files.Read.All`, `Files.ReadWrite.All`, `offline_access`, `User.Read`. Once selected click `Add permissions` at the bottom.
|
5. Search and select the following permissions: `Files.Read`, `Files.ReadWrite`, `Files.Read.All`, `Files.ReadWrite.All`, `offline_access`, `User.Read`. Once selected click `Add permissions` at the bottom.
|
||||||
|
|
||||||
Now the application is complete. Run `rclone config` to create or edit a OneDrive remote.
|
Now the application is complete. Run `rclone config` to create or edit a OneDrive remote.
|
||||||
Supply the app ID and password as Client ID and Secret, respectively. rclone will walk you through the remaining steps.
|
Supply the app ID and password as Client ID and Secret, respectively. rclone will walk you through the remaining steps.
|
||||||
|
|
|
@ -43,7 +43,7 @@ QingStor Access Key ID - leave blank for anonymous access or runtime credentials
|
||||||
access_key_id> access_key
|
access_key_id> access_key
|
||||||
QingStor Secret Access Key (password) - leave blank for anonymous access or runtime credentials.
|
QingStor Secret Access Key (password) - leave blank for anonymous access or runtime credentials.
|
||||||
secret_access_key> secret_key
|
secret_access_key> secret_key
|
||||||
Enter a endpoint URL to connection QingStor API.
|
Enter an endpoint URL to connection QingStor API.
|
||||||
Leave blank will use the default value "https://qingstor.com:443"
|
Leave blank will use the default value "https://qingstor.com:443"
|
||||||
endpoint>
|
endpoint>
|
||||||
Zone connect to. Default is "pek3a".
|
Zone connect to. Default is "pek3a".
|
||||||
|
@ -182,7 +182,7 @@ Leave blank for anonymous access or runtime credentials.
|
||||||
|
|
||||||
#### --qingstor-endpoint
|
#### --qingstor-endpoint
|
||||||
|
|
||||||
Enter a endpoint URL to connection QingStor API.
|
Enter an endpoint URL to connection QingStor API.
|
||||||
Leave blank will use the default value "https://qingstor.com:443"
|
Leave blank will use the default value "https://qingstor.com:443"
|
||||||
|
|
||||||
- Config: endpoint
|
- Config: endpoint
|
||||||
|
@ -260,7 +260,7 @@ Concurrency for multipart uploads.
|
||||||
This is the number of chunks of the same file that are uploaded
|
This is the number of chunks of the same file that are uploaded
|
||||||
concurrently.
|
concurrently.
|
||||||
|
|
||||||
NB if you set this to > 1 then the checksums of multpart uploads
|
NB if you set this to > 1 then the checksums of multipart uploads
|
||||||
become corrupted (the uploads themselves are not corrupted though).
|
become corrupted (the uploads themselves are not corrupted though).
|
||||||
|
|
||||||
If you are uploading small numbers of large file over high speed link
|
If you are uploading small numbers of large file over high speed link
|
||||||
|
|
|
@ -260,7 +260,7 @@ $ rclone rc job/list
|
||||||
|
|
||||||
### Assigning operations to groups with _group = <value>
|
### Assigning operations to groups with _group = <value>
|
||||||
|
|
||||||
Each rc call has it's own stats group for tracking it's metrics. By default
|
Each rc call has its own stats group for tracking its metrics. By default
|
||||||
grouping is done by the composite group name from prefix `job/` and id of the
|
grouping is done by the composite group name from prefix `job/` and id of the
|
||||||
job like so `job/1`.
|
job like so `job/1`.
|
||||||
|
|
||||||
|
@ -426,7 +426,7 @@ Eg
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
If the rate parameter is not suppied then the bandwidth is queried
|
If the rate parameter is not supplied then the bandwidth is queried
|
||||||
|
|
||||||
rclone rc core/bwlimit
|
rclone rc core/bwlimit
|
||||||
{
|
{
|
||||||
|
|
|
@ -1320,7 +1320,7 @@ storage_class =
|
||||||
|
|
||||||
[Spaces](https://www.digitalocean.com/products/object-storage/) is an [S3-interoperable](https://developers.digitalocean.com/documentation/spaces/) object storage service from cloud provider DigitalOcean.
|
[Spaces](https://www.digitalocean.com/products/object-storage/) is an [S3-interoperable](https://developers.digitalocean.com/documentation/spaces/) object storage service from cloud provider DigitalOcean.
|
||||||
|
|
||||||
To connect to DigitalOcean Spaces you will need an access key and secret key. These can be retrieved on the "[Applications & API](https://cloud.digitalocean.com/settings/api/tokens)" page of the DigitalOcean control panel. They will be needed when promted by `rclone config` for your `access_key_id` and `secret_access_key`.
|
To connect to DigitalOcean Spaces you will need an access key and secret key. These can be retrieved on the "[Applications & API](https://cloud.digitalocean.com/settings/api/tokens)" page of the DigitalOcean control panel. They will be needed when prompted by `rclone config` for your `access_key_id` and `secret_access_key`.
|
||||||
|
|
||||||
When prompted for a `region` or `location_constraint`, press enter to use the default value. The region must be included in the `endpoint` setting (e.g. `nyc3.digitaloceanspaces.com`). The default values can be used for other settings.
|
When prompted for a `region` or `location_constraint`, press enter to use the default value. The region must be included in the `endpoint` setting (e.g. `nyc3.digitaloceanspaces.com`). The default values can be used for other settings.
|
||||||
|
|
||||||
|
|
|
@ -480,7 +480,7 @@ The modified time is stored as metadata on the object as
|
||||||
`X-Object-Meta-Mtime` as floating point since the epoch accurate to 1
|
`X-Object-Meta-Mtime` as floating point since the epoch accurate to 1
|
||||||
ns.
|
ns.
|
||||||
|
|
||||||
This is a defacto standard (used in the official python-swiftclient
|
This is a de facto standard (used in the official python-swiftclient
|
||||||
amongst others) for storing the modification time for an object.
|
amongst others) for storing the modification time for an object.
|
||||||
|
|
||||||
### Restricted filename characters
|
### Restricted filename characters
|
||||||
|
|
|
@ -60,7 +60,7 @@ Some policies rely on quota information. These policies should be used only if y
|
||||||
| lus, eplus | Used |
|
| lus, eplus | Used |
|
||||||
| lno, eplno | Objects |
|
| lno, eplno | Objects |
|
||||||
|
|
||||||
To check if your upstream support the field, run `rclone about remote: [flags]` and see if the reuqired field exists.
|
To check if your upstream supports the field, run `rclone about remote: [flags]` and see if the required field exists.
|
||||||
|
|
||||||
#### Filters
|
#### Filters
|
||||||
|
|
||||||
|
@ -74,7 +74,7 @@ If all remotes are filtered an error will be returned.
|
||||||
|
|
||||||
#### Policy descriptions
|
#### Policy descriptions
|
||||||
|
|
||||||
THe policies definition are inspired by [trapexit/mergerfs](https://github.com/trapexit/mergerfs) but not exactly the same. Some policy definition could be different due to the much larger latency of remote file systems.
|
The policies definition are inspired by [trapexit/mergerfs](https://github.com/trapexit/mergerfs) but not exactly the same. Some policy definition could be different due to the much larger latency of remote file systems.
|
||||||
|
|
||||||
| Policy | Description |
|
| Policy | Description |
|
||||||
|------------------|------------------------------------------------------------|
|
|------------------|------------------------------------------------------------|
|
||||||
|
|
|
@ -222,7 +222,7 @@ First, you need to get your remote's URL:
|
||||||
- Now take a look at your address bar, the URL should look like this:
|
- Now take a look at your address bar, the URL should look like this:
|
||||||
`https://[YOUR-DOMAIN]-my.sharepoint.com/personal/[YOUR-EMAIL]/_layouts/15/onedrive.aspx`
|
`https://[YOUR-DOMAIN]-my.sharepoint.com/personal/[YOUR-EMAIL]/_layouts/15/onedrive.aspx`
|
||||||
|
|
||||||
You'll only need this URL upto the email address. After that, you'll
|
You'll only need this URL up to the email address. After that, you'll
|
||||||
most likely want to add "/Documents". That subdirectory contains
|
most likely want to add "/Documents". That subdirectory contains
|
||||||
the actual data stored on your OneDrive.
|
the actual data stored on your OneDrive.
|
||||||
|
|
||||||
|
|
|
@ -177,7 +177,7 @@ Eg
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
If the rate parameter is not suppied then the bandwidth is queried
|
If the rate parameter is not supplied then the bandwidth is queried
|
||||||
|
|
||||||
rclone rc core/bwlimit
|
rclone rc core/bwlimit
|
||||||
{
|
{
|
||||||
|
|
Loading…
Add table
Reference in a new issue