Use cobra autogenerated docs

* put the most up to date docs into the code
  * generate command docs using rclone gendocs
  * put command docs into own directory
  * remake them into MANUAL.md
This commit is contained in:
Nick Craig-Wood 2016-08-03 21:36:28 +01:00
parent ba0b41dd92
commit 549cac90af
32 changed files with 3520 additions and 499 deletions

View file

@ -12,7 +12,7 @@
<div id="header"> <div id="header">
<h1 class="title">rclone(1) User Manual</h1> <h1 class="title">rclone(1) User Manual</h1>
<h2 class="author">Nick Craig-Wood</h2> <h2 class="author">Nick Craig-Wood</h2>
<h3 class="date">Jul 13, 2016</h3> <h3 class="date">Aug 04, 2016</h3>
</div> </div>
<h1 id="rclone">Rclone</h1> <h1 id="rclone">Rclone</h1>
<p><a href="http://rclone.org/"><img src="http://rclone.org/img/rclone-120x120.png" alt="Logo" /></a></p> <p><a href="http://rclone.org/"><img src="http://rclone.org/img/rclone-120x120.png" alt="Logo" /></a></p>
@ -102,7 +102,18 @@ sudo mandb </code></pre>
<p>Source and destination paths are specified by the name you gave the storage system in the config file then the sub path, eg &quot;drive:myfolder&quot; to look at &quot;myfolder&quot; in Google drive.</p> <p>Source and destination paths are specified by the name you gave the storage system in the config file then the sub path, eg &quot;drive:myfolder&quot; to look at &quot;myfolder&quot; in Google drive.</p>
<p>You can define as many storage paths as you like in the config file.</p> <p>You can define as many storage paths as you like in the config file.</p>
<h2 id="subcommands">Subcommands</h2> <h2 id="subcommands">Subcommands</h2>
<h3 id="rclone-copy-sourcepath-destpath">rclone copy source:path dest:path</h3> <p>rclone uses a system of subcommands. For example</p>
<pre><code>rclone ls remote:path # lists a re
rclone copy /local/path remote:path # copies /local/path to the remote
rclone sync /local/path remote:path # syncs /local/path to the remote</code></pre>
<h2 id="rclone-config">rclone config</h2>
<p>Enter an interactive configuration session.</p>
<h3 id="synopsis">Synopsis</h3>
<p>Enter an interactive configuration session.</p>
<pre><code>rclone config</code></pre>
<h2 id="rclone-copy">rclone copy</h2>
<p>Copy files from source to dest, skipping already copied</p>
<h3 id="synopsis-1">Synopsis</h3>
<p>Copy the source to the destination. Doesn't transfer unchanged files, testing by size and modification time or MD5SUM. Doesn't delete files from the destination.</p> <p>Copy the source to the destination. Doesn't transfer unchanged files, testing by size and modification time or MD5SUM. Doesn't delete files from the destination.</p>
<p>Note that it is always the contents of the directory that is synced, not the directory so when source:path is a directory, it's the contents of source:path that are copied, not the directory name and contents.</p> <p>Note that it is always the contents of the directory that is synced, not the directory so when source:path is a directory, it's the contents of source:path that are copied, not the directory name and contents.</p>
<p>If dest:path doesn't exist, it is created and the source:path contents go there.</p> <p>If dest:path doesn't exist, it is created and the source:path contents go there.</p>
@ -119,36 +130,27 @@ destpath/two.txt</code></pre>
destpath/sourcepath/two.txt</code></pre> destpath/sourcepath/two.txt</code></pre>
<p>If you are familiar with <code>rsync</code>, rclone always works as if you had written a trailing / - meaning &quot;copy the contents of this directory&quot;. This applies to all commands and whether you are talking about the source or destination.</p> <p>If you are familiar with <code>rsync</code>, rclone always works as if you had written a trailing / - meaning &quot;copy the contents of this directory&quot;. This applies to all commands and whether you are talking about the source or destination.</p>
<p>See the <code>--no-traverse</code> option for controlling whether rclone lists the destination directory or not.</p> <p>See the <code>--no-traverse</code> option for controlling whether rclone lists the destination directory or not.</p>
<h3 id="rclone-sync-sourcepath-destpath">rclone sync source:path dest:path</h3> <pre><code>rclone copy source:path dest:path</code></pre>
<h2 id="rclone-sync">rclone sync</h2>
<p>Make source and dest identical, modifying destination only.</p>
<h3 id="synopsis-2">Synopsis</h3>
<p>Sync the source to the destination, changing the destination only. Doesn't transfer unchanged files, testing by size and modification time or MD5SUM. Destination is updated to match source, including deleting files if necessary.</p> <p>Sync the source to the destination, changing the destination only. Doesn't transfer unchanged files, testing by size and modification time or MD5SUM. Destination is updated to match source, including deleting files if necessary.</p>
<p><strong>Important</strong>: Since this can cause data loss, test first with the <code>--dry-run</code> flag to see exactly what would be copied and deleted.</p> <p><strong>Important</strong>: Since this can cause data loss, test first with the <code>--dry-run</code> flag to see exactly what would be copied and deleted.</p>
<p>Note that files in the destination won't be deleted if there were any errors at any point.</p> <p>Note that files in the destination won't be deleted if there were any errors at any point.</p>
<p>It is always the contents of the directory that is synced, not the directory so when source:path is a directory, it's the contents of source:path that are copied, not the directory name and contents. See extended explanation in the <code>copy</code> command above if unsure.</p> <p>It is always the contents of the directory that is synced, not the directory so when source:path is a directory, it's the contents of source:path that are copied, not the directory name and contents. See extended explanation in the <code>copy</code> command above if unsure.</p>
<p>If dest:path doesn't exist, it is created and the source:path contents go there.</p> <p>If dest:path doesn't exist, it is created and the source:path contents go there.</p>
<h3 id="move-sourcepath-destpath">move source:path dest:path</h3> <pre><code>rclone sync source:path dest:path</code></pre>
<h2 id="rclone-move">rclone move</h2>
<p>Move files from source to dest.</p>
<h3 id="synopsis-3">Synopsis</h3>
<p>Moves the contents of the source directory to the destination directory. Rclone will error if the source and destination overlap.</p> <p>Moves the contents of the source directory to the destination directory. Rclone will error if the source and destination overlap.</p>
<p>If no filters are in use and if possible this will server side move <code>source:path</code> into <code>dest:path</code>. After this <code>source:path</code> will no longer longer exist.</p> <p>If no filters are in use and if possible this will server side move <code>source:path</code> into <code>dest:path</code>. After this <code>source:path</code> will no longer longer exist.</p>
<p>Otherwise for each file in <code>source:path</code> selected by the filters (if any) this will move it into <code>dest:path</code>. If possible a server side move will be used, otherwise it will copy it (server side if possible) into <code>dest:path</code> then delete the original (if no errors on copy) in <code>source:path</code>.</p> <p>Otherwise for each file in <code>source:path</code> selected by the filters (if any) this will move it into <code>dest:path</code>. If possible a server side move will be used, otherwise it will copy it (server side if possible) into <code>dest:path</code> then delete the original (if no errors on copy) in <code>source:path</code>.</p>
<p><strong>Important</strong>: Since this can cause data loss, test first with the --dry-run flag.</p> <p><strong>Important</strong>: Since this can cause data loss, test first with the --dry-run flag.</p>
<h3 id="rclone-ls-remotepath">rclone ls remote:path</h3> <pre><code>rclone move source:path dest:path</code></pre>
<p>List all the objects in the path with size and path.</p> <h2 id="rclone-delete">rclone delete</h2>
<h3 id="rclone-lsd-remotepath">rclone lsd remote:path</h3> <p>Remove the contents of path.</p>
<p>List all directories/containers/buckets in the the path.</p> <h3 id="synopsis-4">Synopsis</h3>
<h3 id="rclone-lsl-remotepath">rclone lsl remote:path</h3>
<p>List all the objects in the the path with modification time, size and path.</p>
<h3 id="rclone-md5sum-remotepath">rclone md5sum remote:path</h3>
<p>Produces an md5sum file for all the objects in the path. This is in the same format as the standard md5sum tool produces.</p>
<h3 id="rclone-sha1sum-remotepath">rclone sha1sum remote:path</h3>
<p>Produces an sha1sum file for all the objects in the path. This is in the same format as the standard sha1sum tool produces.</p>
<h3 id="rclone-size-remotepath">rclone size remote:path</h3>
<p>Prints the total size of objects in remote:path and the number of objects.</p>
<h3 id="rclone-mkdir-remotepath">rclone mkdir remote:path</h3>
<p>Make the path if it doesn't already exist</p>
<h3 id="rclone-rmdir-remotepath">rclone rmdir remote:path</h3>
<p>Remove the path. Note that you can't remove a path with objects in it, use purge for that.</p>
<h3 id="rclone-purge-remotepath">rclone purge remote:path</h3>
<p>Remove the path and all of its contents. Note that this does not obey include/exclude filters - everything will be removed. Use <code>delete</code> if you want to selectively delete files.</p>
<h3 id="rclone-delete-remotepath">rclone delete remote:path</h3>
<p>Remove the contents of path. Unlike <code>purge</code> it obeys include/exclude filters so can be used to selectively delete files.</p> <p>Remove the contents of path. Unlike <code>purge</code> it obeys include/exclude filters so can be used to selectively delete files.</p>
<p>Eg delete all files bigger than 100MBytes</p> <p>Eg delete all files bigger than 100MBytes</p>
<p>Check what would be deleted first (use either)</p> <p>Check what would be deleted first (use either)</p>
@ -157,12 +159,71 @@ rclone --dry-run --min-size 100M delete remote:path</code></pre>
<p>Then delete</p> <p>Then delete</p>
<pre><code>rclone --min-size 100M delete remote:path</code></pre> <pre><code>rclone --min-size 100M delete remote:path</code></pre>
<p>That reads &quot;delete everything with a minimum size of 100 MB&quot;, hence delete all files bigger than 100MBytes.</p> <p>That reads &quot;delete everything with a minimum size of 100 MB&quot;, hence delete all files bigger than 100MBytes.</p>
<h3 id="rclone-check-sourcepath-destpath">rclone check source:path dest:path</h3> <pre><code>rclone delete remote:path</code></pre>
<h2 id="rclone-purge">rclone purge</h2>
<p>Remove the path and all of its contents.</p>
<h3 id="synopsis-5">Synopsis</h3>
<p>Remove the path and all of its contents. Note that this does not obey include/exclude filters - everything will be removed. Use <code>delete</code> if you want to selectively delete files.</p>
<pre><code>rclone purge remote:path</code></pre>
<h2 id="rclone-mkdir">rclone mkdir</h2>
<p>Make the path if it doesn't already exist.</p>
<h3 id="synopsis-6">Synopsis</h3>
<p>Make the path if it doesn't already exist.</p>
<pre><code>rclone mkdir remote:path</code></pre>
<h2 id="rclone-rmdir">rclone rmdir</h2>
<p>Remove the path if empty.</p>
<h3 id="synopsis-7">Synopsis</h3>
<p>Remove the path. Note that you can't remove a path with objects in it, use purge for that.</p>
<pre><code>rclone rmdir remote:path</code></pre>
<h2 id="rclone-check">rclone check</h2>
<p>Checks the files in the source and destination match.</p>
<h3 id="synopsis-8">Synopsis</h3>
<p>Checks the files in the source and destination match. It compares sizes and MD5SUMs and prints a report of files which don't match. It doesn't alter the source or destination.</p> <p>Checks the files in the source and destination match. It compares sizes and MD5SUMs and prints a report of files which don't match. It doesn't alter the source or destination.</p>
<p><code>--size-only</code> may be used to only compare the sizes, not the MD5SUMs.</p> <p><code>--size-only</code> may be used to only compare the sizes, not the MD5SUMs.</p>
<h3 id="rclone-cleanup-remotepath">rclone cleanup remote:path</h3> <pre><code>rclone check source:path dest:path</code></pre>
<h2 id="rclone-ls">rclone ls</h2>
<p>List all the objects in the the path with size and path.</p>
<h3 id="synopsis-9">Synopsis</h3>
<p>List all the objects in the the path with size and path.</p>
<pre><code>rclone ls remote:path</code></pre>
<h2 id="rclone-lsd">rclone lsd</h2>
<p>List all directories/containers/buckets in the the path.</p>
<h3 id="synopsis-10">Synopsis</h3>
<p>List all directories/containers/buckets in the the path.</p>
<pre><code>rclone lsd remote:path</code></pre>
<h2 id="rclone-lsl">rclone lsl</h2>
<p>List all the objects path with modification time, size and path.</p>
<h3 id="synopsis-11">Synopsis</h3>
<p>List all the objects path with modification time, size and path.</p>
<pre><code>rclone lsl remote:path</code></pre>
<h2 id="rclone-md5sum">rclone md5sum</h2>
<p>Produces an md5sum file for all the objects in the path.</p>
<h3 id="synopsis-12">Synopsis</h3>
<p>Produces an md5sum file for all the objects in the path. This is in the same format as the standard md5sum tool produces.</p>
<pre><code>rclone md5sum remote:path</code></pre>
<h2 id="rclone-sha1sum">rclone sha1sum</h2>
<p>Produces an sha1sum file for all the objects in the path.</p>
<h3 id="synopsis-13">Synopsis</h3>
<p>Produces an sha1sum file for all the objects in the path. This is in the same format as the standard sha1sum tool produces.</p>
<pre><code>rclone sha1sum remote:path</code></pre>
<h2 id="rclone-size">rclone size</h2>
<p>Prints the total size and number of objects in remote:path.</p>
<h3 id="synopsis-14">Synopsis</h3>
<p>Prints the total size and number of objects in remote:path.</p>
<pre><code>rclone size remote:path</code></pre>
<h2 id="rclone-version">rclone version</h2>
<p>Show the version number.</p>
<h3 id="synopsis-15">Synopsis</h3>
<p>Show the version number.</p>
<pre><code>rclone version</code></pre>
<h2 id="rclone-cleanup">rclone cleanup</h2>
<p>Clean up the remote if possible</p>
<h3 id="synopsis-16">Synopsis</h3>
<p>Clean up the remote if possible. Empty the trash or delete old file versions. Not supported by all remotes.</p> <p>Clean up the remote if possible. Empty the trash or delete old file versions. Not supported by all remotes.</p>
<h3 id="rclone-dedupe-remotepath">rclone dedupe remote:path</h3> <pre><code>rclone cleanup remote:path</code></pre>
<h2 id="rclone-dedupe">rclone dedupe</h2>
<p>Interactively find duplicate files delete/rename them.</p>
<h3 id="synopsis-17">Synopsis</h3>
<p>By default <code>dedup</code> interactively finds duplicate files and offers to delete all but one or rename them to be different. Only useful with Google Drive which can have duplicate file names.</p> <p>By default <code>dedup</code> interactively finds duplicate files and offers to delete all but one or rename them to be different. Only useful with Google Drive which can have duplicate file names.</p>
<p>The <code>dedupe</code> command will delete all but one of any identical (same md5sum) files it finds without confirmation. This means that for most duplicated files the <code>dedupe</code> command will not be interactive. You can use <code>--dry-run</code> to see what would happen without doing anything.</p> <p>The <code>dedupe</code> command will delete all but one of any identical (same md5sum) files it finds without confirmation. This means that for most duplicated files the <code>dedupe</code> command will not be interactive. You can use <code>--dry-run</code> to see what would happen without doing anything.</p>
<p>Here is an example run.</p> <p>Here is an example run.</p>
@ -207,7 +268,7 @@ two-3.txt: renamed from: two.txt</code></pre>
564374 2016-03-05 16:22:52.118000000 two-1.txt 564374 2016-03-05 16:22:52.118000000 two-1.txt
6048320 2016-03-05 16:22:46.185000000 two-2.txt 6048320 2016-03-05 16:22:46.185000000 two-2.txt
1744073 2016-03-05 16:22:38.104000000 two-3.txt</code></pre> 1744073 2016-03-05 16:22:38.104000000 two-3.txt</code></pre>
<p>Dedupe can be run non interactively using the <code>--dedupe-mode</code> flag.</p> <p>Dedupe can be run non interactively using the <code>--dedupe-mode</code> flag or by using an extra parameter with the same value</p>
<ul> <ul>
<li><code>--dedupe-mode interactive</code> - interactive as above.</li> <li><code>--dedupe-mode interactive</code> - interactive as above.</li>
<li><code>--dedupe-mode skip</code> - removes identical files then skips anything left.</li> <li><code>--dedupe-mode skip</code> - removes identical files then skips anything left.</li>
@ -218,10 +279,31 @@ two-3.txt: renamed from: two.txt</code></pre>
</ul> </ul>
<p>For example to rename all the identically named photos in your Google Photos directory, do</p> <p>For example to rename all the identically named photos in your Google Photos directory, do</p>
<pre><code>rclone dedupe --dedupe-mode rename &quot;drive:Google Photos&quot;</code></pre> <pre><code>rclone dedupe --dedupe-mode rename &quot;drive:Google Photos&quot;</code></pre>
<h3 id="rclone-config">rclone config</h3> <p>Or</p>
<p>Enter an interactive configuration session.</p> <pre><code>rclone dedupe rename &quot;drive:Google Photos&quot;</code></pre>
<h3 id="rclone-help">rclone help</h3> <pre><code>rclone dedupe [mode] remote:path</code></pre>
<p>Prints help on rclone commands and options.</p> <h3 id="options">Options</h3>
<pre><code> --dedupe-mode value Dedupe mode interactive|skip|first|newest|oldest|rename. (default &quot;interactive&quot;)</code></pre>
<h2 id="rclone-authorize">rclone authorize</h2>
<p>Remote authorization.</p>
<h3 id="synopsis-18">Synopsis</h3>
<p>Remote authorization. Used to authorize a remote or headless rclone from a machine with a browser - use as instructed by rclone config.</p>
<pre><code>rclone authorize</code></pre>
<h2 id="rclone-genautocomplete">rclone genautocomplete</h2>
<p>Output bash completion script for rclone.</p>
<h3 id="synopsis-19">Synopsis</h3>
<p>Generates a bash shell autocompletion script for rclone.</p>
<p>This writes to /etc/bash_completion.d/rclone by default so will probably need to be run with sudo or as root, eg</p>
<pre><code>sudo rclone genautocomplete</code></pre>
<p>Logout and login again to use the autocompletion scripts, or source them directly</p>
<pre><code>. /etc/bash_completion</code></pre>
<p>If you supply a command line argument the script will be written there.</p>
<pre><code>rclone genautocomplete [output_file]</code></pre>
<h2 id="rclone-gendocs">rclone gendocs</h2>
<p>Output markdown docs for rclone to the directory supplied.</p>
<h3 id="synopsis-20">Synopsis</h3>
<p>This produces markdown docs for the rclone commands to the directory supplied. These are in a format suitable for hugo to render into the rclone.org website.</p>
<pre><code>rclone gendocs output_directory</code></pre>
<h2 id="copying-single-files">Copying single files</h2> <h2 id="copying-single-files">Copying single files</h2>
<p>rclone normally syncs or copies directories. However if the source remote points to a file, rclone will just copy that file. The destination remote must point to a directory - rclone will give the error <code>Failed to create file system for &quot;remote:file&quot;: is a file not a directory</code> if it isn't.</p> <p>rclone normally syncs or copies directories. However if the source remote points to a file, rclone will just copy that file. The destination remote must point to a directory - rclone will give the error <code>Failed to create file system for &quot;remote:file&quot;: is a file not a directory</code> if it isn't.</p>
<p>For example, suppose you have a remote with a file in called <code>test.jpg</code>, then you could copy just that file like this</p> <p>For example, suppose you have a remote with a file in called <code>test.jpg</code>, then you could copy just that file like this</p>
@ -258,7 +340,7 @@ two-3.txt: renamed from: two.txt</code></pre>
<p>This can be used when scripting to make aged backups efficiently, eg</p> <p>This can be used when scripting to make aged backups efficiently, eg</p>
<pre><code>rclone sync remote:current-backup remote:previous-backup <pre><code>rclone sync remote:current-backup remote:previous-backup
rclone sync /path/to/files remote:current-backup</code></pre> rclone sync /path/to/files remote:current-backup</code></pre>
<h2 id="options">Options</h2> <h2 id="options-1">Options</h2>
<p>Rclone has a number of options to control its behaviour.</p> <p>Rclone has a number of options to control its behaviour.</p>
<p>Options which use TIME use the go time parser. A duration string is a possibly signed sequence of decimal numbers, each with optional fraction and a unit suffix, such as &quot;300ms&quot;, &quot;-1.5h&quot; or &quot;2h45m&quot;. Valid time units are &quot;ns&quot;, &quot;us&quot; (or &quot;µs&quot;), &quot;ms&quot;, &quot;s&quot;, &quot;m&quot;, &quot;h&quot;.</p> <p>Options which use TIME use the go time parser. A duration string is a possibly signed sequence of decimal numbers, each with optional fraction and a unit suffix, such as &quot;300ms&quot;, &quot;-1.5h&quot; or &quot;2h45m&quot;. Valid time units are &quot;ns&quot;, &quot;us&quot; (or &quot;µs&quot;), &quot;ms&quot;, &quot;s&quot;, &quot;m&quot;, &quot;h&quot;.</p>
<p>Options which use SIZE use kByte by default. However a suffix of <code>b</code> for bytes, <code>k</code> for kBytes, <code>M</code> for MBytes and <code>G</code> for GBytes may be used. These are the binary units, eg 1, 2**10, 2**20, 2**30 respectively.</p> <p>Options which use SIZE use kByte by default. However a suffix of <code>b</code> for bytes, <code>k</code> for kBytes, <code>M</code> for MBytes and <code>G</code> for GBytes may be used. These are the binary units, eg 1, 2**10, 2**20, 2**30 respectively.</p>

440
MANUAL.md
View file

@ -1,6 +1,6 @@
% rclone(1) User Manual % rclone(1) User Manual
% Nick Craig-Wood % Nick Craig-Wood
% Jul 13, 2016 % Aug 04, 2016
Rclone Rclone
====== ======
@ -134,7 +134,32 @@ You can define as many storage paths as you like in the config file.
Subcommands Subcommands
----------- -----------
### rclone copy source:path dest:path ### rclone uses a system of subcommands. For example
rclone ls remote:path # lists a re
rclone copy /local/path remote:path # copies /local/path to the remote
rclone sync /local/path remote:path # syncs /local/path to the remote
## rclone config
Enter an interactive configuration session.
### Synopsis
Enter an interactive configuration session.
```
rclone config
```
## rclone copy
Copy files from source to dest, skipping already copied
### Synopsis
Copy the source to the destination. Doesn't transfer Copy the source to the destination. Doesn't transfer
unchanged files, testing by size and modification time or unchanged files, testing by size and modification time or
@ -175,7 +200,18 @@ source or destination.
See the `--no-traverse` option for controlling whether rclone lists See the `--no-traverse` option for controlling whether rclone lists
the destination directory or not. the destination directory or not.
### rclone sync source:path dest:path ###
```
rclone copy source:path dest:path
```
## rclone sync
Make source and dest identical, modifying destination only.
### Synopsis
Sync the source to the destination, changing the destination Sync the source to the destination, changing the destination
only. Doesn't transfer unchanged files, testing by size and only. Doesn't transfer unchanged files, testing by size and
@ -196,7 +232,18 @@ extended explanation in the `copy` command above if unsure.
If dest:path doesn't exist, it is created and the source:path contents If dest:path doesn't exist, it is created and the source:path contents
go there. go there.
### move source:path dest:path ###
```
rclone sync source:path dest:path
```
## rclone move
Move files from source to dest.
### Synopsis
Moves the contents of the source directory to the destination Moves the contents of the source directory to the destination
directory. Rclone will error if the source and destination overlap. directory. Rclone will error if the source and destination overlap.
@ -214,50 +261,18 @@ into `dest:path` then delete the original (if no errors on copy) in
**Important**: Since this can cause data loss, test first with the **Important**: Since this can cause data loss, test first with the
--dry-run flag. --dry-run flag.
### rclone ls remote:path ###
List all the objects in the path with size and path. ```
rclone move source:path dest:path
```
### rclone lsd remote:path ### ## rclone delete
List all directories/containers/buckets in the the path. Remove the contents of path.
### rclone lsl remote:path ### ### Synopsis
List all the objects in the the path with modification time,
size and path.
### rclone md5sum remote:path ###
Produces an md5sum file for all the objects in the path. This
is in the same format as the standard md5sum tool produces.
### rclone sha1sum remote:path ###
Produces an sha1sum file for all the objects in the path. This
is in the same format as the standard sha1sum tool produces.
### rclone size remote:path ###
Prints the total size of objects in remote:path and the number of
objects.
### rclone mkdir remote:path ###
Make the path if it doesn't already exist
### rclone rmdir remote:path ###
Remove the path. Note that you can't remove a path with
objects in it, use purge for that.
### rclone purge remote:path ###
Remove the path and all of its contents. Note that this does not obey
include/exclude filters - everything will be removed. Use `delete` if
you want to selectively delete files.
### rclone delete remote:path ###
Remove the contents of path. Unlike `purge` it obeys include/exclude Remove the contents of path. Unlike `purge` it obeys include/exclude
filters so can be used to selectively delete files. filters so can be used to selectively delete files.
@ -276,7 +291,63 @@ Then delete
That reads "delete everything with a minimum size of 100 MB", hence That reads "delete everything with a minimum size of 100 MB", hence
delete all files bigger than 100MBytes. delete all files bigger than 100MBytes.
### rclone check source:path dest:path ###
```
rclone delete remote:path
```
## rclone purge
Remove the path and all of its contents.
### Synopsis
Remove the path and all of its contents. Note that this does not obey
include/exclude filters - everything will be removed. Use `delete` if
you want to selectively delete files.
```
rclone purge remote:path
```
## rclone mkdir
Make the path if it doesn't already exist.
### Synopsis
Make the path if it doesn't already exist.
```
rclone mkdir remote:path
```
## rclone rmdir
Remove the path if empty.
### Synopsis
Remove the path. Note that you can't remove a path with
objects in it, use purge for that.
```
rclone rmdir remote:path
```
## rclone check
Checks the files in the source and destination match.
### Synopsis
Checks the files in the source and destination match. It Checks the files in the source and destination match. It
compares sizes and MD5SUMs and prints a report of files which compares sizes and MD5SUMs and prints a report of files which
@ -284,12 +355,131 @@ don't match. It doesn't alter the source or destination.
`--size-only` may be used to only compare the sizes, not the MD5SUMs. `--size-only` may be used to only compare the sizes, not the MD5SUMs.
### rclone cleanup remote:path ###
```
rclone check source:path dest:path
```
## rclone ls
List all the objects in the the path with size and path.
### Synopsis
List all the objects in the the path with size and path.
```
rclone ls remote:path
```
## rclone lsd
List all directories/containers/buckets in the the path.
### Synopsis
List all directories/containers/buckets in the the path.
```
rclone lsd remote:path
```
## rclone lsl
List all the objects path with modification time, size and path.
### Synopsis
List all the objects path with modification time, size and path.
```
rclone lsl remote:path
```
## rclone md5sum
Produces an md5sum file for all the objects in the path.
### Synopsis
Produces an md5sum file for all the objects in the path. This
is in the same format as the standard md5sum tool produces.
```
rclone md5sum remote:path
```
## rclone sha1sum
Produces an sha1sum file for all the objects in the path.
### Synopsis
Produces an sha1sum file for all the objects in the path. This
is in the same format as the standard sha1sum tool produces.
```
rclone sha1sum remote:path
```
## rclone size
Prints the total size and number of objects in remote:path.
### Synopsis
Prints the total size and number of objects in remote:path.
```
rclone size remote:path
```
## rclone version
Show the version number.
### Synopsis
Show the version number.
```
rclone version
```
## rclone cleanup
Clean up the remote if possible
### Synopsis
Clean up the remote if possible. Empty the trash or delete old file Clean up the remote if possible. Empty the trash or delete old file
versions. Not supported by all remotes. versions. Not supported by all remotes.
### rclone dedupe remote:path ###
```
rclone cleanup remote:path
```
## rclone dedupe
Interactively find duplicate files delete/rename them.
### Synopsis
By default `dedup` interactively finds duplicate files and offers to By default `dedup` interactively finds duplicate files and offers to
delete all but one or rename them to be different. Only useful with delete all but one or rename them to be different. Only useful with
@ -304,58 +494,52 @@ Here is an example run.
Before - with duplicates Before - with duplicates
``` $ rclone lsl drive:dupes
$ rclone lsl drive:dupes 6048320 2016-03-05 16:23:16.798000000 one.txt
6048320 2016-03-05 16:23:16.798000000 one.txt 6048320 2016-03-05 16:23:11.775000000 one.txt
6048320 2016-03-05 16:23:11.775000000 one.txt 564374 2016-03-05 16:23:06.731000000 one.txt
564374 2016-03-05 16:23:06.731000000 one.txt 6048320 2016-03-05 16:18:26.092000000 one.txt
6048320 2016-03-05 16:18:26.092000000 one.txt 6048320 2016-03-05 16:22:46.185000000 two.txt
6048320 2016-03-05 16:22:46.185000000 two.txt 1744073 2016-03-05 16:22:38.104000000 two.txt
1744073 2016-03-05 16:22:38.104000000 two.txt 564374 2016-03-05 16:22:52.118000000 two.txt
564374 2016-03-05 16:22:52.118000000 two.txt
```
Now the `dedupe` session Now the `dedupe` session
``` $ rclone dedupe drive:dupes
$ rclone dedupe drive:dupes 2016/03/05 16:24:37 Google drive root 'dupes': Looking for duplicates using interactive mode.
2016/03/05 16:24:37 Google drive root 'dupes': Looking for duplicates using interactive mode. one.txt: Found 4 duplicates - deleting identical copies
one.txt: Found 4 duplicates - deleting identical copies one.txt: Deleting 2/3 identical duplicates (md5sum "1eedaa9fe86fd4b8632e2ac549403b36")
one.txt: Deleting 2/3 identical duplicates (md5sum "1eedaa9fe86fd4b8632e2ac549403b36") one.txt: 2 duplicates remain
one.txt: 2 duplicates remain 1: 6048320 bytes, 2016-03-05 16:23:16.798000000, md5sum 1eedaa9fe86fd4b8632e2ac549403b36
1: 6048320 bytes, 2016-03-05 16:23:16.798000000, md5sum 1eedaa9fe86fd4b8632e2ac549403b36 2: 564374 bytes, 2016-03-05 16:23:06.731000000, md5sum 7594e7dc9fc28f727c42ee3e0749de81
2: 564374 bytes, 2016-03-05 16:23:06.731000000, md5sum 7594e7dc9fc28f727c42ee3e0749de81 s) Skip and do nothing
s) Skip and do nothing k) Keep just one (choose which in next step)
k) Keep just one (choose which in next step) r) Rename all to be different (by changing file.jpg to file-1.jpg)
r) Rename all to be different (by changing file.jpg to file-1.jpg) s/k/r> k
s/k/r> k Enter the number of the file to keep> 1
Enter the number of the file to keep> 1 one.txt: Deleted 1 extra copies
one.txt: Deleted 1 extra copies two.txt: Found 3 duplicates - deleting identical copies
two.txt: Found 3 duplicates - deleting identical copies two.txt: 3 duplicates remain
two.txt: 3 duplicates remain 1: 564374 bytes, 2016-03-05 16:22:52.118000000, md5sum 7594e7dc9fc28f727c42ee3e0749de81
1: 564374 bytes, 2016-03-05 16:22:52.118000000, md5sum 7594e7dc9fc28f727c42ee3e0749de81 2: 6048320 bytes, 2016-03-05 16:22:46.185000000, md5sum 1eedaa9fe86fd4b8632e2ac549403b36
2: 6048320 bytes, 2016-03-05 16:22:46.185000000, md5sum 1eedaa9fe86fd4b8632e2ac549403b36 3: 1744073 bytes, 2016-03-05 16:22:38.104000000, md5sum 851957f7fb6f0bc4ce76be966d336802
3: 1744073 bytes, 2016-03-05 16:22:38.104000000, md5sum 851957f7fb6f0bc4ce76be966d336802 s) Skip and do nothing
s) Skip and do nothing k) Keep just one (choose which in next step)
k) Keep just one (choose which in next step) r) Rename all to be different (by changing file.jpg to file-1.jpg)
r) Rename all to be different (by changing file.jpg to file-1.jpg) s/k/r> r
s/k/r> r two-1.txt: renamed from: two.txt
two-1.txt: renamed from: two.txt two-2.txt: renamed from: two.txt
two-2.txt: renamed from: two.txt two-3.txt: renamed from: two.txt
two-3.txt: renamed from: two.txt
```
The result being The result being
``` $ rclone lsl drive:dupes
$ rclone lsl drive:dupes 6048320 2016-03-05 16:23:16.798000000 one.txt
6048320 2016-03-05 16:23:16.798000000 one.txt 564374 2016-03-05 16:22:52.118000000 two-1.txt
564374 2016-03-05 16:22:52.118000000 two-1.txt 6048320 2016-03-05 16:22:46.185000000 two-2.txt
6048320 2016-03-05 16:22:46.185000000 two-2.txt 1744073 2016-03-05 16:22:38.104000000 two-3.txt
1744073 2016-03-05 16:22:38.104000000 two-3.txt
```
Dedupe can be run non interactively using the `--dedupe-mode` flag. Dedupe can be run non interactively using the `--dedupe-mode` flag or by using an extra parameter with the same value
* `--dedupe-mode interactive` - interactive as above. * `--dedupe-mode interactive` - interactive as above.
* `--dedupe-mode skip` - removes identical files then skips anything left. * `--dedupe-mode skip` - removes identical files then skips anything left.
@ -368,13 +552,81 @@ For example to rename all the identically named photos in your Google Photos dir
rclone dedupe --dedupe-mode rename "drive:Google Photos" rclone dedupe --dedupe-mode rename "drive:Google Photos"
### rclone config ### Or
Enter an interactive configuration session. rclone dedupe rename "drive:Google Photos"
### rclone help ###
Prints help on rclone commands and options. ```
rclone dedupe [mode] remote:path
```
### Options
```
--dedupe-mode value Dedupe mode interactive|skip|first|newest|oldest|rename. (default "interactive")
```
## rclone authorize
Remote authorization.
### Synopsis
Remote authorization. Used to authorize a remote or headless
rclone from a machine with a browser - use as instructed by
rclone config.
```
rclone authorize
```
## rclone genautocomplete
Output bash completion script for rclone.
### Synopsis
Generates a bash shell autocompletion script for rclone.
This writes to /etc/bash_completion.d/rclone by default so will
probably need to be run with sudo or as root, eg
sudo rclone genautocomplete
Logout and login again to use the autocompletion scripts, or source
them directly
. /etc/bash_completion
If you supply a command line argument the script will be written
there.
```
rclone genautocomplete [output_file]
```
## rclone gendocs
Output markdown docs for rclone to the directory supplied.
### Synopsis
This produces markdown docs for the rclone commands to the directory
supplied. These are in a format suitable for hugo to render into the
rclone.org website.
```
rclone gendocs output_directory
```
Copying single files Copying single files
-------------------- --------------------

View file

@ -1,6 +1,6 @@
rclone(1) User Manual rclone(1) User Manual
Nick Craig-Wood Nick Craig-Wood
Jul 13, 2016 Aug 04, 2016
@ -134,7 +134,29 @@ You can define as many storage paths as you like in the config file.
Subcommands Subcommands
rclone copy source:path dest:path rclone uses a system of subcommands. For example
rclone ls remote:path # lists a re
rclone copy /local/path remote:path # copies /local/path to the remote
rclone sync /local/path remote:path # syncs /local/path to the remote
rclone config
Enter an interactive configuration session.
Synopsis
Enter an interactive configuration session.
rclone config
rclone copy
Copy files from source to dest, skipping already copied
Synopsis
Copy the source to the destination. Doesn't transfer unchanged files, Copy the source to the destination. Doesn't transfer unchanged files,
testing by size and modification time or MD5SUM. Doesn't delete files testing by size and modification time or MD5SUM. Doesn't delete files
@ -174,7 +196,14 @@ source or destination.
See the --no-traverse option for controlling whether rclone lists the See the --no-traverse option for controlling whether rclone lists the
destination directory or not. destination directory or not.
rclone sync source:path dest:path rclone copy source:path dest:path
rclone sync
Make source and dest identical, modifying destination only.
Synopsis
Sync the source to the destination, changing the destination only. Sync the source to the destination, changing the destination only.
Doesn't transfer unchanged files, testing by size and modification time Doesn't transfer unchanged files, testing by size and modification time
@ -195,7 +224,14 @@ extended explanation in the copy command above if unsure.
If dest:path doesn't exist, it is created and the source:path contents If dest:path doesn't exist, it is created and the source:path contents
go there. go there.
move source:path dest:path rclone sync source:path dest:path
rclone move
Move files from source to dest.
Synopsis
Moves the contents of the source directory to the destination directory. Moves the contents of the source directory to the destination directory.
Rclone will error if the source and destination overlap. Rclone will error if the source and destination overlap.
@ -212,50 +248,14 @@ then delete the original (if no errors on copy) in source:path.
IMPORTANT: Since this can cause data loss, test first with the --dry-run IMPORTANT: Since this can cause data loss, test first with the --dry-run
flag. flag.
rclone ls remote:path rclone move source:path dest:path
List all the objects in the path with size and path.
rclone lsd remote:path rclone delete
List all directories/containers/buckets in the the path. Remove the contents of path.
rclone lsl remote:path Synopsis
List all the objects in the the path with modification time, size and
path.
rclone md5sum remote:path
Produces an md5sum file for all the objects in the path. This is in the
same format as the standard md5sum tool produces.
rclone sha1sum remote:path
Produces an sha1sum file for all the objects in the path. This is in the
same format as the standard sha1sum tool produces.
rclone size remote:path
Prints the total size of objects in remote:path and the number of
objects.
rclone mkdir remote:path
Make the path if it doesn't already exist
rclone rmdir remote:path
Remove the path. Note that you can't remove a path with objects in it,
use purge for that.
rclone purge remote:path
Remove the path and all of its contents. Note that this does not obey
include/exclude filters - everything will be removed. Use delete if you
want to selectively delete files.
rclone delete remote:path
Remove the contents of path. Unlike purge it obeys include/exclude Remove the contents of path. Unlike purge it obeys include/exclude
filters so can be used to selectively delete files. filters so can be used to selectively delete files.
@ -274,7 +274,50 @@ Then delete
That reads "delete everything with a minimum size of 100 MB", hence That reads "delete everything with a minimum size of 100 MB", hence
delete all files bigger than 100MBytes. delete all files bigger than 100MBytes.
rclone check source:path dest:path rclone delete remote:path
rclone purge
Remove the path and all of its contents.
Synopsis
Remove the path and all of its contents. Note that this does not obey
include/exclude filters - everything will be removed. Use delete if you
want to selectively delete files.
rclone purge remote:path
rclone mkdir
Make the path if it doesn't already exist.
Synopsis
Make the path if it doesn't already exist.
rclone mkdir remote:path
rclone rmdir
Remove the path if empty.
Synopsis
Remove the path. Note that you can't remove a path with objects in it,
use purge for that.
rclone rmdir remote:path
rclone check
Checks the files in the source and destination match.
Synopsis
Checks the files in the source and destination match. It compares sizes Checks the files in the source and destination match. It compares sizes
and MD5SUMs and prints a report of files which don't match. It doesn't and MD5SUMs and prints a report of files which don't match. It doesn't
@ -282,12 +325,105 @@ alter the source or destination.
--size-only may be used to only compare the sizes, not the MD5SUMs. --size-only may be used to only compare the sizes, not the MD5SUMs.
rclone cleanup remote:path rclone check source:path dest:path
rclone ls
List all the objects in the the path with size and path.
Synopsis
List all the objects in the the path with size and path.
rclone ls remote:path
rclone lsd
List all directories/containers/buckets in the the path.
Synopsis
List all directories/containers/buckets in the the path.
rclone lsd remote:path
rclone lsl
List all the objects path with modification time, size and path.
Synopsis
List all the objects path with modification time, size and path.
rclone lsl remote:path
rclone md5sum
Produces an md5sum file for all the objects in the path.
Synopsis
Produces an md5sum file for all the objects in the path. This is in the
same format as the standard md5sum tool produces.
rclone md5sum remote:path
rclone sha1sum
Produces an sha1sum file for all the objects in the path.
Synopsis
Produces an sha1sum file for all the objects in the path. This is in the
same format as the standard sha1sum tool produces.
rclone sha1sum remote:path
rclone size
Prints the total size and number of objects in remote:path.
Synopsis
Prints the total size and number of objects in remote:path.
rclone size remote:path
rclone version
Show the version number.
Synopsis
Show the version number.
rclone version
rclone cleanup
Clean up the remote if possible
Synopsis
Clean up the remote if possible. Empty the trash or delete old file Clean up the remote if possible. Empty the trash or delete old file
versions. Not supported by all remotes. versions. Not supported by all remotes.
rclone dedupe remote:path rclone cleanup remote:path
rclone dedupe
Interactively find duplicate files delete/rename them.
Synopsis
By default dedup interactively finds duplicate files and offers to By default dedup interactively finds duplicate files and offers to
delete all but one or rename them to be different. Only useful with delete all but one or rename them to be different. Only useful with
@ -347,7 +483,8 @@ The result being
6048320 2016-03-05 16:22:46.185000000 two-2.txt 6048320 2016-03-05 16:22:46.185000000 two-2.txt
1744073 2016-03-05 16:22:38.104000000 two-3.txt 1744073 2016-03-05 16:22:38.104000000 two-3.txt
Dedupe can be run non interactively using the --dedupe-mode flag. Dedupe can be run non interactively using the --dedupe-mode flag or by
using an extra parameter with the same value
- --dedupe-mode interactive - interactive as above. - --dedupe-mode interactive - interactive as above.
- --dedupe-mode skip - removes identical files then skips - --dedupe-mode skip - removes identical files then skips
@ -366,13 +503,63 @@ Photos directory, do
rclone dedupe --dedupe-mode rename "drive:Google Photos" rclone dedupe --dedupe-mode rename "drive:Google Photos"
rclone config Or
Enter an interactive configuration session. rclone dedupe rename "drive:Google Photos"
rclone help rclone dedupe [mode] remote:path
Prints help on rclone commands and options. Options
--dedupe-mode value Dedupe mode interactive|skip|first|newest|oldest|rename. (default "interactive")
rclone authorize
Remote authorization.
Synopsis
Remote authorization. Used to authorize a remote or headless rclone from
a machine with a browser - use as instructed by rclone config.
rclone authorize
rclone genautocomplete
Output bash completion script for rclone.
Synopsis
Generates a bash shell autocompletion script for rclone.
This writes to /etc/bash_completion.d/rclone by default so will probably
need to be run with sudo or as root, eg
sudo rclone genautocomplete
Logout and login again to use the autocompletion scripts, or source them
directly
. /etc/bash_completion
If you supply a command line argument the script will be written there.
rclone genautocomplete [output_file]
rclone gendocs
Output markdown docs for rclone to the directory supplied.
Synopsis
This produces markdown docs for the rclone commands to the directory
supplied. These are in a format suitable for hugo to render into the
rclone.org website.
rclone gendocs output_directory
Copying single files Copying single files

View file

@ -40,7 +40,7 @@ doc: rclone.1 MANUAL.html MANUAL.txt
rclone.1: MANUAL.md rclone.1: MANUAL.md
pandoc -s --from markdown --to man MANUAL.md -o rclone.1 pandoc -s --from markdown --to man MANUAL.md -o rclone.1
MANUAL.md: make_manual.py docs/content/*.md MANUAL.md: make_manual.py docs/content/*.md commanddocs
./make_manual.py ./make_manual.py
MANUAL.html: MANUAL.md MANUAL.html: MANUAL.md
@ -49,6 +49,9 @@ MANUAL.html: MANUAL.md
MANUAL.txt: MANUAL.md MANUAL.txt: MANUAL.md
pandoc -s --from markdown --to plain MANUAL.md -o MANUAL.txt pandoc -s --from markdown --to plain MANUAL.md -o MANUAL.txt
commanddocs: rclone
rclone gendocs docs/content/commands/
install: rclone install: rclone
install -d ${DESTDIR}/usr/bin install -d ${DESTDIR}/usr/bin
install -t ${DESTDIR}/usr/bin ${GOPATH}/bin/rclone install -t ${DESTDIR}/usr/bin ${GOPATH}/bin/rclone

View file

@ -0,0 +1,140 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone"
slug: rclone
url: /commands/rclone/
---
## rclone
Sync files and directories to and from local and remote object stores - v1.32
### Synopsis
Rclone is a command line program to sync files and directories to and
from various cloud storage systems, such as:
* Google Drive
* Amazon S3
* Openstack Swift / Rackspace cloud files / Memset Memstore
* Dropbox
* Google Cloud Storage
* Amazon Drive
* Microsoft One Drive
* Hubic
* Backblaze B2
* Yandex Disk
* The local filesystem
Features
* MD5/SHA1 hashes checked at all times for file integrity
* Timestamps preserved on files
* Partial syncs supported on a whole file basis
* Copy mode to just copy new/changed files
* Sync (one way) mode to make a directory identical
* Check mode to check for file hash equality
* Can sync to and from network, eg two different cloud accounts
See the home page for installation, usage, documentation, changelog
and configuration walkthroughs.
* http://rclone.org/
```
rclone
```
### Options
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
-V, --version Print the version number
```
### SEE ALSO
* [rclone authorize](/commands/rclone_authorize/) - Remote authorization.
* [rclone check](/commands/rclone_check/) - Checks the files in the source and destination match.
* [rclone cleanup](/commands/rclone_cleanup/) - Clean up the remote if possible
* [rclone config](/commands/rclone_config/) - Enter an interactive configuration session.
* [rclone copy](/commands/rclone_copy/) - Copy files from source to dest, skipping already copied
* [rclone dedupe](/commands/rclone_dedupe/) - Interactively find duplicate files delete/rename them.
* [rclone delete](/commands/rclone_delete/) - Remove the contents of path.
* [rclone genautocomplete](/commands/rclone_genautocomplete/) - Output bash completion script for rclone.
* [rclone gendocs](/commands/rclone_gendocs/) - Output markdown docs for rclone to the directory supplied.
* [rclone ls](/commands/rclone_ls/) - List all the objects in the the path with size and path.
* [rclone lsd](/commands/rclone_lsd/) - List all directories/containers/buckets in the the path.
* [rclone lsl](/commands/rclone_lsl/) - List all the objects path with modification time, size and path.
* [rclone md5sum](/commands/rclone_md5sum/) - Produces an md5sum file for all the objects in the path.
* [rclone mkdir](/commands/rclone_mkdir/) - Make the path if it doesn't already exist.
* [rclone move](/commands/rclone_move/) - Move files from source to dest.
* [rclone purge](/commands/rclone_purge/) - Remove the path and all of its contents.
* [rclone rmdir](/commands/rclone_rmdir/) - Remove the path if empty.
* [rclone sha1sum](/commands/rclone_sha1sum/) - Produces an sha1sum file for all the objects in the path.
* [rclone size](/commands/rclone_size/) - Prints the total size and number of objects in remote:path.
* [rclone sync](/commands/rclone_sync/) - Make source and dest identical, modifying destination only.
* [rclone version](/commands/rclone_version/) - Show the version number.
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,92 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone authorize"
slug: rclone_authorize
url: /commands/rclone_authorize/
---
## rclone authorize
Remote authorization.
### Synopsis
Remote authorization. Used to authorize a remote or headless
rclone from a machine with a browser - use as instructed by
rclone config.
```
rclone authorize
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,95 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone check"
slug: rclone_check
url: /commands/rclone_check/
---
## rclone check
Checks the files in the source and destination match.
### Synopsis
Checks the files in the source and destination match. It
compares sizes and MD5SUMs and prints a report of files which
don't match. It doesn't alter the source or destination.
`--size-only` may be used to only compare the sizes, not the MD5SUMs.
```
rclone check source:path dest:path
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,92 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone cleanup"
slug: rclone_cleanup
url: /commands/rclone_cleanup/
---
## rclone cleanup
Clean up the remote if possible
### Synopsis
Clean up the remote if possible. Empty the trash or delete old file
versions. Not supported by all remotes.
```
rclone cleanup remote:path
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,89 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone config"
slug: rclone_config
url: /commands/rclone_config/
---
## rclone config
Enter an interactive configuration session.
### Synopsis
Enter an interactive configuration session.
```
rclone config
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,128 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone copy"
slug: rclone_copy
url: /commands/rclone_copy/
---
## rclone copy
Copy files from source to dest, skipping already copied
### Synopsis
Copy the source to the destination. Doesn't transfer
unchanged files, testing by size and modification time or
MD5SUM. Doesn't delete files from the destination.
Note that it is always the contents of the directory that is synced,
not the directory so when source:path is a directory, it's the
contents of source:path that are copied, not the directory name and
contents.
If dest:path doesn't exist, it is created and the source:path contents
go there.
For example
rclone copy source:sourcepath dest:destpath
Let's say there are two files in sourcepath
sourcepath/one.txt
sourcepath/two.txt
This copies them to
destpath/one.txt
destpath/two.txt
Not to
destpath/sourcepath/one.txt
destpath/sourcepath/two.txt
If you are familiar with `rsync`, rclone always works as if you had
written a trailing / - meaning "copy the contents of this directory".
This applies to all commands and whether you are talking about the
source or destination.
See the `--no-traverse` option for controlling whether rclone lists
the destination directory or not.
```
rclone copy source:path dest:path
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,170 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone dedupe"
slug: rclone_dedupe
url: /commands/rclone_dedupe/
---
## rclone dedupe
Interactively find duplicate files delete/rename them.
### Synopsis
By default `dedup` interactively finds duplicate files and offers to
delete all but one or rename them to be different. Only useful with
Google Drive which can have duplicate file names.
The `dedupe` command will delete all but one of any identical (same
md5sum) files it finds without confirmation. This means that for most
duplicated files the `dedupe` command will not be interactive. You
can use `--dry-run` to see what would happen without doing anything.
Here is an example run.
Before - with duplicates
$ rclone lsl drive:dupes
6048320 2016-03-05 16:23:16.798000000 one.txt
6048320 2016-03-05 16:23:11.775000000 one.txt
564374 2016-03-05 16:23:06.731000000 one.txt
6048320 2016-03-05 16:18:26.092000000 one.txt
6048320 2016-03-05 16:22:46.185000000 two.txt
1744073 2016-03-05 16:22:38.104000000 two.txt
564374 2016-03-05 16:22:52.118000000 two.txt
Now the `dedupe` session
$ rclone dedupe drive:dupes
2016/03/05 16:24:37 Google drive root 'dupes': Looking for duplicates using interactive mode.
one.txt: Found 4 duplicates - deleting identical copies
one.txt: Deleting 2/3 identical duplicates (md5sum "1eedaa9fe86fd4b8632e2ac549403b36")
one.txt: 2 duplicates remain
1: 6048320 bytes, 2016-03-05 16:23:16.798000000, md5sum 1eedaa9fe86fd4b8632e2ac549403b36
2: 564374 bytes, 2016-03-05 16:23:06.731000000, md5sum 7594e7dc9fc28f727c42ee3e0749de81
s) Skip and do nothing
k) Keep just one (choose which in next step)
r) Rename all to be different (by changing file.jpg to file-1.jpg)
s/k/r> k
Enter the number of the file to keep> 1
one.txt: Deleted 1 extra copies
two.txt: Found 3 duplicates - deleting identical copies
two.txt: 3 duplicates remain
1: 564374 bytes, 2016-03-05 16:22:52.118000000, md5sum 7594e7dc9fc28f727c42ee3e0749de81
2: 6048320 bytes, 2016-03-05 16:22:46.185000000, md5sum 1eedaa9fe86fd4b8632e2ac549403b36
3: 1744073 bytes, 2016-03-05 16:22:38.104000000, md5sum 851957f7fb6f0bc4ce76be966d336802
s) Skip and do nothing
k) Keep just one (choose which in next step)
r) Rename all to be different (by changing file.jpg to file-1.jpg)
s/k/r> r
two-1.txt: renamed from: two.txt
two-2.txt: renamed from: two.txt
two-3.txt: renamed from: two.txt
The result being
$ rclone lsl drive:dupes
6048320 2016-03-05 16:23:16.798000000 one.txt
564374 2016-03-05 16:22:52.118000000 two-1.txt
6048320 2016-03-05 16:22:46.185000000 two-2.txt
1744073 2016-03-05 16:22:38.104000000 two-3.txt
Dedupe can be run non interactively using the `--dedupe-mode` flag or by using an extra parameter with the same value
* `--dedupe-mode interactive` - interactive as above.
* `--dedupe-mode skip` - removes identical files then skips anything left.
* `--dedupe-mode first` - removes identical files then keeps the first one.
* `--dedupe-mode newest` - removes identical files then keeps the newest one.
* `--dedupe-mode oldest` - removes identical files then keeps the oldest one.
* `--dedupe-mode rename` - removes identical files then renames the rest to be different.
For example to rename all the identically named photos in your Google Photos directory, do
rclone dedupe --dedupe-mode rename "drive:Google Photos"
Or
rclone dedupe rename "drive:Google Photos"
```
rclone dedupe [mode] remote:path
```
### Options
```
--dedupe-mode value Dedupe mode interactive|skip|first|newest|oldest|rename. (default "interactive")
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,106 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone delete"
slug: rclone_delete
url: /commands/rclone_delete/
---
## rclone delete
Remove the contents of path.
### Synopsis
Remove the contents of path. Unlike `purge` it obeys include/exclude
filters so can be used to selectively delete files.
Eg delete all files bigger than 100MBytes
Check what would be deleted first (use either)
rclone --min-size 100M lsl remote:path
rclone --dry-run --min-size 100M delete remote:path
Then delete
rclone --min-size 100M delete remote:path
That reads "delete everything with a minimum size of 100 MB", hence
delete all files bigger than 100MBytes.
```
rclone delete remote:path
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,104 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone genautocomplete"
slug: rclone_genautocomplete
url: /commands/rclone_genautocomplete/
---
## rclone genautocomplete
Output bash completion script for rclone.
### Synopsis
Generates a bash shell autocompletion script for rclone.
This writes to /etc/bash_completion.d/rclone by default so will
probably need to be run with sudo or as root, eg
sudo rclone genautocomplete
Logout and login again to use the autocompletion scripts, or source
them directly
. /etc/bash_completion
If you supply a command line argument the script will be written
there.
```
rclone genautocomplete [output_file]
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,92 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone gendocs"
slug: rclone_gendocs
url: /commands/rclone_gendocs/
---
## rclone gendocs
Output markdown docs for rclone to the directory supplied.
### Synopsis
This produces markdown docs for the rclone commands to the directory
supplied. These are in a format suitable for hugo to render into the
rclone.org website.
```
rclone gendocs output_directory
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,89 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone ls"
slug: rclone_ls
url: /commands/rclone_ls/
---
## rclone ls
List all the objects in the the path with size and path.
### Synopsis
List all the objects in the the path with size and path.
```
rclone ls remote:path
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,89 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone lsd"
slug: rclone_lsd
url: /commands/rclone_lsd/
---
## rclone lsd
List all directories/containers/buckets in the the path.
### Synopsis
List all directories/containers/buckets in the the path.
```
rclone lsd remote:path
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,89 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone lsl"
slug: rclone_lsl
url: /commands/rclone_lsl/
---
## rclone lsl
List all the objects path with modification time, size and path.
### Synopsis
List all the objects path with modification time, size and path.
```
rclone lsl remote:path
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,92 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone md5sum"
slug: rclone_md5sum
url: /commands/rclone_md5sum/
---
## rclone md5sum
Produces an md5sum file for all the objects in the path.
### Synopsis
Produces an md5sum file for all the objects in the path. This
is in the same format as the standard md5sum tool produces.
```
rclone md5sum remote:path
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,89 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone mkdir"
slug: rclone_mkdir
url: /commands/rclone_mkdir/
---
## rclone mkdir
Make the path if it doesn't already exist.
### Synopsis
Make the path if it doesn't already exist.
```
rclone mkdir remote:path
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,105 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone move"
slug: rclone_move
url: /commands/rclone_move/
---
## rclone move
Move files from source to dest.
### Synopsis
Moves the contents of the source directory to the destination
directory. Rclone will error if the source and destination overlap.
If no filters are in use and if possible this will server side move
`source:path` into `dest:path`. After this `source:path` will no
longer longer exist.
Otherwise for each file in `source:path` selected by the filters (if
any) this will move it into `dest:path`. If possible a server side
move will be used, otherwise it will copy it (server side if possible)
into `dest:path` then delete the original (if no errors on copy) in
`source:path`.
**Important**: Since this can cause data loss, test first with the
--dry-run flag.
```
rclone move source:path dest:path
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,93 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone purge"
slug: rclone_purge
url: /commands/rclone_purge/
---
## rclone purge
Remove the path and all of its contents.
### Synopsis
Remove the path and all of its contents. Note that this does not obey
include/exclude filters - everything will be removed. Use `delete` if
you want to selectively delete files.
```
rclone purge remote:path
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,91 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone rmdir"
slug: rclone_rmdir
url: /commands/rclone_rmdir/
---
## rclone rmdir
Remove the path if empty.
### Synopsis
Remove the path. Note that you can't remove a path with
objects in it, use purge for that.
```
rclone rmdir remote:path
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,92 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone sha1sum"
slug: rclone_sha1sum
url: /commands/rclone_sha1sum/
---
## rclone sha1sum
Produces an sha1sum file for all the objects in the path.
### Synopsis
Produces an sha1sum file for all the objects in the path. This
is in the same format as the standard sha1sum tool produces.
```
rclone sha1sum remote:path
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,89 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone size"
slug: rclone_size
url: /commands/rclone_size/
---
## rclone size
Prints the total size and number of objects in remote:path.
### Synopsis
Prints the total size and number of objects in remote:path.
```
rclone size remote:path
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,108 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone sync"
slug: rclone_sync
url: /commands/rclone_sync/
---
## rclone sync
Make source and dest identical, modifying destination only.
### Synopsis
Sync the source to the destination, changing the destination
only. Doesn't transfer unchanged files, testing by size and
modification time or MD5SUM. Destination is updated to match
source, including deleting files if necessary.
**Important**: Since this can cause data loss, test first with the
`--dry-run` flag to see exactly what would be copied and deleted.
Note that files in the destination won't be deleted if there were any
errors at any point.
It is always the contents of the directory that is synced, not the
directory so when source:path is a directory, it's the contents of
source:path that are copied, not the directory name and contents. See
extended explanation in the `copy` command above if unsure.
If dest:path doesn't exist, it is created and the source:path contents
go there.
```
rclone sync source:path dest:path
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -0,0 +1,89 @@
---
date: 2016-08-04T21:37:09+01:00
title: "rclone version"
slug: rclone_version
url: /commands/rclone_version/
---
## rclone version
Show the version number.
### Synopsis
Show the version number.
```
rclone version
```
### Options inherited from parent commands
```
--acd-templink-threshold value Files >= this size will be downloaded via their tempLink. (default 9G)
--ask-password Allow prompt for password for encrypted configuration. (default true)
--b2-chunk-size value Upload chunk size. Must fit in memory. (default 96M)
--b2-test-mode string A flag string for X-Bz-Test-Mode header.
--b2-upload-cutoff value Cutoff for switching to chunked upload (default 190.735M)
--b2-versions Include old versions in directory listings.
--bwlimit value Bandwidth limit in kBytes/s, or use suffix b|k|M|G
--checkers int Number of checkers to run in parallel. (default 8)
-c, --checksum Skip based on checksum & size, not mod-time & size
--config string Config file. (default "/home/ncw/.rclone.conf")
--contimeout duration Connect timeout (default 1m0s)
--cpuprofile string Write cpu profile to file
--delete-after When synchronizing, delete files on destination after transfering
--delete-before When synchronizing, delete files on destination before transfering
--delete-during When synchronizing, delete files during transfer (default)
--delete-excluded Delete files on dest excluded from sync
--drive-auth-owner-only Only consider files owned by the authenticated user. Requires drive-full-list.
--drive-chunk-size value Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
--drive-full-list Use a full listing for directory list. More data but usually quicker. (obsolete)
--drive-upload-cutoff value Cutoff for switching to chunked upload (default 8M)
--drive-use-trash Send files to the trash instead of deleting permanently.
--dropbox-chunk-size value Upload chunk size. Max 150M. (default 128M)
-n, --dry-run Do a trial run with no permanent changes
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
--dump-filters Dump the filters to the output
--dump-headers Dump HTTP headers - may contain sensitive info
--exclude string Exclude files matching pattern
--exclude-from string Read exclude patterns from file
--files-from string Read list of source-file names from file
-f, --filter string Add a file-filtering rule
--filter-from string Read filtering patterns from a file
--ignore-existing Skip all files that exist on destination
--ignore-size Ignore size when skipping use mod-time or checksum.
-I, --ignore-times Don't skip files that match size and time - transfer all files
--include string Include files matching pattern
--include-from string Read include patterns from file
--log-file string Log everything to this file
--low-level-retries int Number of low level retries to do. (default 10)
--max-age string Don't transfer any file older than this in s or suffix ms|s|m|h|d|w|M|y
--max-depth int If set limits the recursion depth to this. (default -1)
--max-size value Don't transfer any file larger than this in k or suffix b|k|M|G (default off)
--memprofile string Write memory profile to file
--min-age string Don't transfer any file younger than this in s or suffix ms|s|m|h|d|w|M|y
--min-size value Don't transfer any file smaller than this in k or suffix b|k|M|G (default off)
--modify-window duration Max time diff to be considered the same (default 1ns)
--no-check-certificate Do not verify the server SSL certificate. Insecure.
--no-gzip-encoding Don't set Accept-Encoding: gzip.
--no-traverse Don't traverse destination file system on copy.
--no-update-modtime Don't update destination mod-time if files identical.
--onedrive-chunk-size value Above this size files will be chunked - must be multiple of 320k. (default 10M)
--onedrive-upload-cutoff value Cutoff for switching to chunked upload - must be <= 100MB (default 10M)
-q, --quiet Print as little stuff as possible
--retries int Retry operations this many times if they fail (default 3)
--size-only Skip based on size only, not mod-time or checksum
--stats duration Interval to print stats (0 to disable) (default 1m0s)
--swift-chunk-size value Above this size files will be chunked into a _segments container. (default 5G)
--timeout duration IO idle timeout (default 5m0s)
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
-v, --verbose Print lots more stuff
```
### SEE ALSO
* [rclone](/commands/rclone/) - Sync files and directories to and from local and remote object stores - v1.32
###### Auto generated by spf13/cobra on 4-Aug-2016

View file

@ -49,251 +49,34 @@ You can define as many storage paths as you like in the config file.
Subcommands Subcommands
----------- -----------
### rclone copy source:path dest:path ### rclone uses a system of subcommands. For example
Copy the source to the destination. Doesn't transfer rclone ls remote:path # lists a re
unchanged files, testing by size and modification time or rclone copy /local/path remote:path # copies /local/path to the remote
MD5SUM. Doesn't delete files from the destination. rclone sync /local/path remote:path # syncs /local/path to the remote
Note that it is always the contents of the directory that is synced, The main rclone commands with most used first
not the directory so when source:path is a directory, it's the
contents of source:path that are copied, not the directory name and * [rclone config](/commands/rclone_config/) - Enter an interactive configuration session.
contents. * [rclone copy](/commands/rclone_copy/) - Copy files from source to dest, skipping already copied
* [rclone sync](/commands/rclone_sync/) - Make source and dest identical, modifying destination only.
If dest:path doesn't exist, it is created and the source:path contents * [rclone move](/commands/rclone_move/) - Move files from source to dest.
go there. * [rclone delete](/commands/rclone_delete/) - Remove the contents of path.
* [rclone purge](/commands/rclone_purge/) - Remove the path and all of its contents.
For example * [rclone mkdir](/commands/rclone_mkdir/) - Make the path if it doesn't already exist.
* [rclone rmdir](/commands/rclone_rmdir/) - Remove the path.
rclone copy source:sourcepath dest:destpath * [rclone check](/commands/rclone_check/) - Checks the files in the source and destination match.
* [rclone ls](/commands/rclone_ls/) - List all the objects in the the path with size and path.
Let's say there are two files in sourcepath * [rclone lsd](/commands/rclone_lsd/) - List all directories/containers/buckets in the the path.
* [rclone lsl](/commands/rclone_lsl/) - List all the objects path with modification time, size and path.
sourcepath/one.txt * [rclone md5sum](/commands/rclone_md5sum/) - Produces an md5sum file for all the objects in the path.
sourcepath/two.txt * [rclone sha1sum](/commands/rclone_sha1sum/) - Produces an sha1sum file for all the objects in the path.
* [rclone size](/commands/rclone_size/) - Returns the total size and number of objects in remote:path.
This copies them to * [rclone version](/commands/rclone_version/) - Show the version number.
* [rclone cleanup](/commands/rclone_cleanup/) - Clean up the remote if possible
destpath/one.txt * [rclone dedupe](/commands/rclone_dedupe/) - Interactively find duplicate files delete/rename them.
destpath/two.txt
See the [commands index](/commands/) for the full list.
Not to
destpath/sourcepath/one.txt
destpath/sourcepath/two.txt
If you are familiar with `rsync`, rclone always works as if you had
written a trailing / - meaning "copy the contents of this directory".
This applies to all commands and whether you are talking about the
source or destination.
See the `--no-traverse` option for controlling whether rclone lists
the destination directory or not.
### rclone sync source:path dest:path ###
Sync the source to the destination, changing the destination
only. Doesn't transfer unchanged files, testing by size and
modification time or MD5SUM. Destination is updated to match
source, including deleting files if necessary.
**Important**: Since this can cause data loss, test first with the
`--dry-run` flag to see exactly what would be copied and deleted.
Note that files in the destination won't be deleted if there were any
errors at any point.
It is always the contents of the directory that is synced, not the
directory so when source:path is a directory, it's the contents of
source:path that are copied, not the directory name and contents. See
extended explanation in the `copy` command above if unsure.
If dest:path doesn't exist, it is created and the source:path contents
go there.
### move source:path dest:path ###
Moves the contents of the source directory to the destination
directory. Rclone will error if the source and destination overlap.
If no filters are in use and if possible this will server side move
`source:path` into `dest:path`. After this `source:path` will no
longer longer exist.
Otherwise for each file in `source:path` selected by the filters (if
any) this will move it into `dest:path`. If possible a server side
move will be used, otherwise it will copy it (server side if possible)
into `dest:path` then delete the original (if no errors on copy) in
`source:path`.
**Important**: Since this can cause data loss, test first with the
--dry-run flag.
### rclone ls remote:path ###
List all the objects in the path with size and path.
### rclone lsd remote:path ###
List all directories/containers/buckets in the the path.
### rclone lsl remote:path ###
List all the objects in the the path with modification time,
size and path.
### rclone md5sum remote:path ###
Produces an md5sum file for all the objects in the path. This
is in the same format as the standard md5sum tool produces.
### rclone sha1sum remote:path ###
Produces an sha1sum file for all the objects in the path. This
is in the same format as the standard sha1sum tool produces.
### rclone size remote:path ###
Prints the total size of objects in remote:path and the number of
objects.
### rclone mkdir remote:path ###
Make the path if it doesn't already exist
### rclone rmdir remote:path ###
Remove the path. Note that you can't remove a path with
objects in it, use purge for that.
### rclone purge remote:path ###
Remove the path and all of its contents. Note that this does not obey
include/exclude filters - everything will be removed. Use `delete` if
you want to selectively delete files.
### rclone delete remote:path ###
Remove the contents of path. Unlike `purge` it obeys include/exclude
filters so can be used to selectively delete files.
Eg delete all files bigger than 100MBytes
Check what would be deleted first (use either)
rclone --min-size 100M lsl remote:path
rclone --dry-run --min-size 100M delete remote:path
Then delete
rclone --min-size 100M delete remote:path
That reads "delete everything with a minimum size of 100 MB", hence
delete all files bigger than 100MBytes.
### rclone check source:path dest:path ###
Checks the files in the source and destination match. It
compares sizes and MD5SUMs and prints a report of files which
don't match. It doesn't alter the source or destination.
`--size-only` may be used to only compare the sizes, not the MD5SUMs.
### rclone cleanup remote:path ###
Clean up the remote if possible. Empty the trash or delete old file
versions. Not supported by all remotes.
### rclone dedupe remote:path ###
By default `dedup` interactively finds duplicate files and offers to
delete all but one or rename them to be different. Only useful with
Google Drive which can have duplicate file names.
The `dedupe` command will delete all but one of any identical (same
md5sum) files it finds without confirmation. This means that for most
duplicated files the `dedupe` command will not be interactive. You
can use `--dry-run` to see what would happen without doing anything.
Here is an example run.
Before - with duplicates
```
$ rclone lsl drive:dupes
6048320 2016-03-05 16:23:16.798000000 one.txt
6048320 2016-03-05 16:23:11.775000000 one.txt
564374 2016-03-05 16:23:06.731000000 one.txt
6048320 2016-03-05 16:18:26.092000000 one.txt
6048320 2016-03-05 16:22:46.185000000 two.txt
1744073 2016-03-05 16:22:38.104000000 two.txt
564374 2016-03-05 16:22:52.118000000 two.txt
```
Now the `dedupe` session
```
$ rclone dedupe drive:dupes
2016/03/05 16:24:37 Google drive root 'dupes': Looking for duplicates using interactive mode.
one.txt: Found 4 duplicates - deleting identical copies
one.txt: Deleting 2/3 identical duplicates (md5sum "1eedaa9fe86fd4b8632e2ac549403b36")
one.txt: 2 duplicates remain
1: 6048320 bytes, 2016-03-05 16:23:16.798000000, md5sum 1eedaa9fe86fd4b8632e2ac549403b36
2: 564374 bytes, 2016-03-05 16:23:06.731000000, md5sum 7594e7dc9fc28f727c42ee3e0749de81
s) Skip and do nothing
k) Keep just one (choose which in next step)
r) Rename all to be different (by changing file.jpg to file-1.jpg)
s/k/r> k
Enter the number of the file to keep> 1
one.txt: Deleted 1 extra copies
two.txt: Found 3 duplicates - deleting identical copies
two.txt: 3 duplicates remain
1: 564374 bytes, 2016-03-05 16:22:52.118000000, md5sum 7594e7dc9fc28f727c42ee3e0749de81
2: 6048320 bytes, 2016-03-05 16:22:46.185000000, md5sum 1eedaa9fe86fd4b8632e2ac549403b36
3: 1744073 bytes, 2016-03-05 16:22:38.104000000, md5sum 851957f7fb6f0bc4ce76be966d336802
s) Skip and do nothing
k) Keep just one (choose which in next step)
r) Rename all to be different (by changing file.jpg to file-1.jpg)
s/k/r> r
two-1.txt: renamed from: two.txt
two-2.txt: renamed from: two.txt
two-3.txt: renamed from: two.txt
```
The result being
```
$ rclone lsl drive:dupes
6048320 2016-03-05 16:23:16.798000000 one.txt
564374 2016-03-05 16:22:52.118000000 two-1.txt
6048320 2016-03-05 16:22:46.185000000 two-2.txt
1744073 2016-03-05 16:22:38.104000000 two-3.txt
```
Dedupe can be run non interactively using the `--dedupe-mode` flag.
* `--dedupe-mode interactive` - interactive as above.
* `--dedupe-mode skip` - removes identical files then skips anything left.
* `--dedupe-mode first` - removes identical files then keeps the first one.
* `--dedupe-mode newest` - removes identical files then keeps the newest one.
* `--dedupe-mode oldest` - removes identical files then keeps the oldest one.
* `--dedupe-mode rename` - removes identical files then renames the rest to be different.
For example to rename all the identically named photos in your Google Photos directory, do
rclone dedupe --dedupe-mode rename "drive:Google Photos"
The modes can also be passed as an extra parameter, eg
rclone dedupe rename "drive:Google Photos"
### rclone config ###
Enter an interactive configuration session.
### rclone help ###
Prints help on rclone commands and options.
Copying single files Copying single files
-------------------- --------------------

View file

@ -27,6 +27,24 @@
<li><a href="/privacy/"><i class="fa fa-book"></i> Privacy Policy</a></li> <li><a href="/privacy/"><i class="fa fa-book"></i> Privacy Policy</a></li>
</ul> </ul>
</li> </li>
<li class="dropdown">
<a href="#" class="dropdown-toggle" data-toggle="dropdown"><b class="caret"></b> Commands</a>
<ul class="dropdown-menu">
<li><a href="/commands/rclone_config/"><i class="fa fa-book"></i> rclone config</a></li>
<li><a href="/commands/rclone_copy/"><i class="fa fa-book"></i> rclone copy</a></li>
<li><a href="/commands/rclone_sync/"><i class="fa fa-book"></i> rclone sync</a></li>
<li><a href="/commands/rclone_move/"><i class="fa fa-book"></i> rclone move</a></li>
<li><a href="/commands/rclone_purge/"><i class="fa fa-book"></i> rclone purge</a></li>
<li><a href="/commands/rclone_mkdir/"><i class="fa fa-book"></i> rclone mkdir</a></li>
<li><a href="/commands/rclone_rmdir/"><i class="fa fa-book"></i> rclone rmdir</a></li>
<li><a href="/commands/rclone_check/"><i class="fa fa-book"></i> rclone check</a></li>
<li><a href="/commands/rclone_ls/"><i class="fa fa-book"></i> rclone ls</a></li>
<li><a href="/commands/rclone_lsd/"><i class="fa fa-book"></i> rclone lsd</a></li>
<li><a href="/commands/rclone_delete/"><i class="fa fa-book"></i> rclone delete</a></li>
<li><a href="/commands/rclone_size/"><i class="fa fa-book"></i> rclone size</a></li>
<li><a href="/commands/"><i class="fa fa-book"></i> ...and the rest</a></li>
</ul>
</li>
<li class="dropdown"> <li class="dropdown">
<a href="#" class="dropdown-toggle" data-toggle="dropdown"><b class="caret"></b> Storage Systems</a> <a href="#" class="dropdown-toggle" data-toggle="dropdown"><b class="caret"></b> Storage Systems</a>
<ul class="dropdown-menu"> <ul class="dropdown-menu">

View file

@ -0,0 +1,33 @@
{{ template "chrome/header.html" . }}
<body>
{{ template "chrome/navbar.html" . }}
<div class="container">
<div class="row">
<div class="col-md-9">
<div class="well well-sm">
<h3>{{ .Title }}<br> <small>{{ .Description }}</small></h3>
<hr>
<!-- FIXME factor this properly -->
<body>
<h1>Rclone Commands</h1>
<p>This is an index of all commands in rclone.</p>
<ul>
{{ range .Data.Pages }}
<li><a href="{{ .URL }}">{{ .Title }}</a></li>
{{ end }}
</ul>
<p>Docs autogenerated by <a href="https://github.com/spf13/cobra">Cobra</a>.</p>
<!-- FIXME factor this properly -->
</div>
</div>
<!-- Sidebar -->
<div class="col-md-3">
{{ template "chrome/menu.html" . }}
</div>
</div>
{{ template "chrome/footer.copyright.html" . }}
</div>
{{ template "chrome/footer.html" . }}

View file

@ -38,6 +38,28 @@ docs = [
"contact.md", "contact.md",
] ]
# Order to put the commands in - any not on here will be in sorted order
commands_order = [
"rclone_config.md",
"rclone_copy.md",
"rclone_sync.md",
"rclone_move.md",
"rclone_delete.md",
"rclone_purge.md",
"rclone_mkdir.md",
"rclone_rmdir.md",
"rclone_check.md",
"rclone_ls.md",
"rclone_lsd.md",
"rclone_lsl.md",
"rclone_md5sum.md",
"rclone_sha1sum.md",
"rclone_size.md",
"rclone_version.md",
"rclone_cleanup.md",
"rclone_dedupe.md",
]
# Docs which aren't made into outfile # Docs which aren't made into outfile
ignore_docs = [ ignore_docs = [
"downloads.md", "downloads.md",
@ -71,8 +93,27 @@ def check_docs(docpath):
print "Files in docs variable but not on disk: %s" % ", ".join(docs_set - files) print "Files in docs variable but not on disk: %s" % ", ".join(docs_set - files)
raise ValueError("Missing files") raise ValueError("Missing files")
def read_command(command):
doc = read_doc("commands/"+command)
doc = re.sub(r"### Options inherited from parent commands.*$", "", doc, 0, re.S)
doc = doc.strip()+"\n"
return doc
def read_commands(docpath):
"""Reads the commands an makes them into a single page"""
files = set(f for f in os.listdir(docpath + "/commands") if f.endswith(".md"))
docs = []
for command in commands_order:
docs.append(read_command(command))
files.remove(command)
for command in sorted(files):
if command != "rclone.md":
docs.append(read_command(command))
return "\n".join(docs)
def main(): def main():
check_docs(docpath) check_docs(docpath)
command_docs = read_commands(docpath)
with open(outfile, "w") as out: with open(outfile, "w") as out:
out.write("""\ out.write("""\
%% rclone(1) User Manual %% rclone(1) User Manual
@ -81,7 +122,11 @@ def main():
""" % datetime.now().strftime("%b %d, %Y")) """ % datetime.now().strftime("%b %d, %Y"))
for doc in docs: for doc in docs:
out.write(read_doc(doc)) contents = read_doc(doc)
# Substitute the commands into doc.md
if doc == "docs.md":
contents = re.sub(r"The main rclone commands.*?for the full list.", command_docs, contents, 0, re.S)
out.write(contents)
print "Written '%s'" % outfile print "Written '%s'" % outfile
if __name__ == "__main__": if __name__ == "__main__":

343
rclone.1
View file

@ -1,7 +1,7 @@
.\"t .\"t
.\" Automatically generated by Pandoc 1.16.0.2 .\" Automatically generated by Pandoc 1.16.0.2
.\" .\"
.TH "rclone" "1" "Jul 13, 2016" "User Manual" "" .TH "rclone" "1" "Aug 04, 2016" "User Manual" ""
.hy .hy
.SH Rclone .SH Rclone
.PP .PP
@ -181,7 +181,33 @@ to look at "myfolder" in Google drive.
.PP .PP
You can define as many storage paths as you like in the config file. You can define as many storage paths as you like in the config file.
.SS Subcommands .SS Subcommands
.SS rclone copy source:path dest:path .PP
rclone uses a system of subcommands.
For example
.IP
.nf
\f[C]
rclone\ ls\ remote:path\ #\ lists\ a\ re
rclone\ copy\ /local/path\ remote:path\ #\ copies\ /local/path\ to\ the\ remote
rclone\ sync\ /local/path\ remote:path\ #\ syncs\ /local/path\ to\ the\ remote
\f[]
.fi
.SS rclone config
.PP
Enter an interactive configuration session.
.SS Synopsis
.PP
Enter an interactive configuration session.
.IP
.nf
\f[C]
rclone\ config
\f[]
.fi
.SS rclone copy
.PP
Copy files from source to dest, skipping already copied
.SS Synopsis
.PP .PP
Copy the source to the destination. Copy the source to the destination.
Doesn\[aq]t transfer unchanged files, testing by size and modification Doesn\[aq]t transfer unchanged files, testing by size and modification
@ -238,7 +264,16 @@ source or destination.
.PP .PP
See the \f[C]\-\-no\-traverse\f[] option for controlling whether rclone See the \f[C]\-\-no\-traverse\f[] option for controlling whether rclone
lists the destination directory or not. lists the destination directory or not.
.SS rclone sync source:path dest:path .IP
.nf
\f[C]
rclone\ copy\ source:path\ dest:path
\f[]
.fi
.SS rclone sync
.PP
Make source and dest identical, modifying destination only.
.SS Synopsis
.PP .PP
Sync the source to the destination, changing the destination only. Sync the source to the destination, changing the destination only.
Doesn\[aq]t transfer unchanged files, testing by size and modification Doesn\[aq]t transfer unchanged files, testing by size and modification
@ -260,7 +295,16 @@ See extended explanation in the \f[C]copy\f[] command above if unsure.
.PP .PP
If dest:path doesn\[aq]t exist, it is created and the source:path If dest:path doesn\[aq]t exist, it is created and the source:path
contents go there. contents go there.
.SS move source:path dest:path .IP
.nf
\f[C]
rclone\ sync\ source:path\ dest:path
\f[]
.fi
.SS rclone move
.PP
Move files from source to dest.
.SS Synopsis
.PP .PP
Moves the contents of the source directory to the destination directory. Moves the contents of the source directory to the destination directory.
Rclone will error if the source and destination overlap. Rclone will error if the source and destination overlap.
@ -277,43 +321,16 @@ original (if no errors on copy) in \f[C]source:path\f[].
.PP .PP
\f[B]Important\f[]: Since this can cause data loss, test first with the \f[B]Important\f[]: Since this can cause data loss, test first with the
\-\-dry\-run flag. \-\-dry\-run flag.
.SS rclone ls remote:path .IP
.nf
\f[C]
rclone\ move\ source:path\ dest:path
\f[]
.fi
.SS rclone delete
.PP .PP
List all the objects in the path with size and path. Remove the contents of path.
.SS rclone lsd remote:path .SS Synopsis
.PP
List all directories/containers/buckets in the the path.
.SS rclone lsl remote:path
.PP
List all the objects in the the path with modification time, size and
path.
.SS rclone md5sum remote:path
.PP
Produces an md5sum file for all the objects in the path.
This is in the same format as the standard md5sum tool produces.
.SS rclone sha1sum remote:path
.PP
Produces an sha1sum file for all the objects in the path.
This is in the same format as the standard sha1sum tool produces.
.SS rclone size remote:path
.PP
Prints the total size of objects in remote:path and the number of
objects.
.SS rclone mkdir remote:path
.PP
Make the path if it doesn\[aq]t already exist
.SS rclone rmdir remote:path
.PP
Remove the path.
Note that you can\[aq]t remove a path with objects in it, use purge for
that.
.SS rclone purge remote:path
.PP
Remove the path and all of its contents.
Note that this does not obey include/exclude filters \- everything will
be removed.
Use \f[C]delete\f[] if you want to selectively delete files.
.SS rclone delete remote:path
.PP .PP
Remove the contents of path. Remove the contents of path.
Unlike \f[C]purge\f[] it obeys include/exclude filters so can be used to Unlike \f[C]purge\f[] it obeys include/exclude filters so can be used to
@ -340,7 +357,57 @@ rclone\ \-\-min\-size\ 100M\ delete\ remote:path
.PP .PP
That reads "delete everything with a minimum size of 100 MB", hence That reads "delete everything with a minimum size of 100 MB", hence
delete all files bigger than 100MBytes. delete all files bigger than 100MBytes.
.SS rclone check source:path dest:path .IP
.nf
\f[C]
rclone\ delete\ remote:path
\f[]
.fi
.SS rclone purge
.PP
Remove the path and all of its contents.
.SS Synopsis
.PP
Remove the path and all of its contents.
Note that this does not obey include/exclude filters \- everything will
be removed.
Use \f[C]delete\f[] if you want to selectively delete files.
.IP
.nf
\f[C]
rclone\ purge\ remote:path
\f[]
.fi
.SS rclone mkdir
.PP
Make the path if it doesn\[aq]t already exist.
.SS Synopsis
.PP
Make the path if it doesn\[aq]t already exist.
.IP
.nf
\f[C]
rclone\ mkdir\ remote:path
\f[]
.fi
.SS rclone rmdir
.PP
Remove the path if empty.
.SS Synopsis
.PP
Remove the path.
Note that you can\[aq]t remove a path with objects in it, use purge for
that.
.IP
.nf
\f[C]
rclone\ rmdir\ remote:path
\f[]
.fi
.SS rclone check
.PP
Checks the files in the source and destination match.
.SS Synopsis
.PP .PP
Checks the files in the source and destination match. Checks the files in the source and destination match.
It compares sizes and MD5SUMs and prints a report of files which It compares sizes and MD5SUMs and prints a report of files which
@ -349,12 +416,116 @@ It doesn\[aq]t alter the source or destination.
.PP .PP
\f[C]\-\-size\-only\f[] may be used to only compare the sizes, not the \f[C]\-\-size\-only\f[] may be used to only compare the sizes, not the
MD5SUMs. MD5SUMs.
.SS rclone cleanup remote:path .IP
.nf
\f[C]
rclone\ check\ source:path\ dest:path
\f[]
.fi
.SS rclone ls
.PP
List all the objects in the the path with size and path.
.SS Synopsis
.PP
List all the objects in the the path with size and path.
.IP
.nf
\f[C]
rclone\ ls\ remote:path
\f[]
.fi
.SS rclone lsd
.PP
List all directories/containers/buckets in the the path.
.SS Synopsis
.PP
List all directories/containers/buckets in the the path.
.IP
.nf
\f[C]
rclone\ lsd\ remote:path
\f[]
.fi
.SS rclone lsl
.PP
List all the objects path with modification time, size and path.
.SS Synopsis
.PP
List all the objects path with modification time, size and path.
.IP
.nf
\f[C]
rclone\ lsl\ remote:path
\f[]
.fi
.SS rclone md5sum
.PP
Produces an md5sum file for all the objects in the path.
.SS Synopsis
.PP
Produces an md5sum file for all the objects in the path.
This is in the same format as the standard md5sum tool produces.
.IP
.nf
\f[C]
rclone\ md5sum\ remote:path
\f[]
.fi
.SS rclone sha1sum
.PP
Produces an sha1sum file for all the objects in the path.
.SS Synopsis
.PP
Produces an sha1sum file for all the objects in the path.
This is in the same format as the standard sha1sum tool produces.
.IP
.nf
\f[C]
rclone\ sha1sum\ remote:path
\f[]
.fi
.SS rclone size
.PP
Prints the total size and number of objects in remote:path.
.SS Synopsis
.PP
Prints the total size and number of objects in remote:path.
.IP
.nf
\f[C]
rclone\ size\ remote:path
\f[]
.fi
.SS rclone version
.PP
Show the version number.
.SS Synopsis
.PP
Show the version number.
.IP
.nf
\f[C]
rclone\ version
\f[]
.fi
.SS rclone cleanup
.PP
Clean up the remote if possible
.SS Synopsis
.PP .PP
Clean up the remote if possible. Clean up the remote if possible.
Empty the trash or delete old file versions. Empty the trash or delete old file versions.
Not supported by all remotes. Not supported by all remotes.
.SS rclone dedupe remote:path .IP
.nf
\f[C]
rclone\ cleanup\ remote:path
\f[]
.fi
.SS rclone dedupe
.PP
Interactively find duplicate files delete/rename them.
.SS Synopsis
.PP .PP
By default \f[C]dedup\f[] interactively finds duplicate files and offers By default \f[C]dedup\f[] interactively finds duplicate files and offers
to delete all but one or rename them to be different. to delete all but one or rename them to be different.
@ -429,7 +600,7 @@ $\ rclone\ lsl\ drive:dupes
.fi .fi
.PP .PP
Dedupe can be run non interactively using the \f[C]\-\-dedupe\-mode\f[] Dedupe can be run non interactively using the \f[C]\-\-dedupe\-mode\f[]
flag. flag or by using an extra parameter with the same value
.IP \[bu] 2 .IP \[bu] 2
\f[C]\-\-dedupe\-mode\ interactive\f[] \- interactive as above. \f[C]\-\-dedupe\-mode\ interactive\f[] \- interactive as above.
.IP \[bu] 2 .IP \[bu] 2
@ -456,12 +627,88 @@ Photos directory, do
rclone\ dedupe\ \-\-dedupe\-mode\ rename\ "drive:Google\ Photos" rclone\ dedupe\ \-\-dedupe\-mode\ rename\ "drive:Google\ Photos"
\f[] \f[]
.fi .fi
.SS rclone config
.PP .PP
Enter an interactive configuration session. Or
.SS rclone help .IP
.nf
\f[C]
rclone\ dedupe\ rename\ "drive:Google\ Photos"
\f[]
.fi
.IP
.nf
\f[C]
rclone\ dedupe\ [mode]\ remote:path
\f[]
.fi
.SS Options
.IP
.nf
\f[C]
\ \ \ \ \ \ \-\-dedupe\-mode\ value\ \ \ Dedupe\ mode\ interactive|skip|first|newest|oldest|rename.\ (default\ "interactive")
\f[]
.fi
.SS rclone authorize
.PP .PP
Prints help on rclone commands and options. Remote authorization.
.SS Synopsis
.PP
Remote authorization.
Used to authorize a remote or headless rclone from a machine with a
browser \- use as instructed by rclone config.
.IP
.nf
\f[C]
rclone\ authorize
\f[]
.fi
.SS rclone genautocomplete
.PP
Output bash completion script for rclone.
.SS Synopsis
.PP
Generates a bash shell autocompletion script for rclone.
.PP
This writes to /etc/bash_completion.d/rclone by default so will probably
need to be run with sudo or as root, eg
.IP
.nf
\f[C]
sudo\ rclone\ genautocomplete
\f[]
.fi
.PP
Logout and login again to use the autocompletion scripts, or source them
directly
.IP
.nf
\f[C]
\&.\ /etc/bash_completion
\f[]
.fi
.PP
If you supply a command line argument the script will be written there.
.IP
.nf
\f[C]
rclone\ genautocomplete\ [output_file]
\f[]
.fi
.SS rclone gendocs
.PP
Output markdown docs for rclone to the directory supplied.
.SS Synopsis
.PP
This produces markdown docs for the rclone commands to the directory
supplied.
These are in a format suitable for hugo to render into the rclone.org
website.
.IP
.nf
\f[C]
rclone\ gendocs\ output_directory
\f[]
.fi
.SS Copying single files .SS Copying single files
.PP .PP
rclone normally syncs or copies directories. rclone normally syncs or copies directories.

200
rclone.go
View file

@ -240,7 +240,43 @@ var copyCmd = &cobra.Command{
Long: ` Long: `
Copy the source to the destination. Doesn't transfer Copy the source to the destination. Doesn't transfer
unchanged files, testing by size and modification time or unchanged files, testing by size and modification time or
MD5SUM. Doesn't delete files from the destination.`, MD5SUM. Doesn't delete files from the destination.
Note that it is always the contents of the directory that is synced,
not the directory so when source:path is a directory, it's the
contents of source:path that are copied, not the directory name and
contents.
If dest:path doesn't exist, it is created and the source:path contents
go there.
For example
rclone copy source:sourcepath dest:destpath
Let's say there are two files in sourcepath
sourcepath/one.txt
sourcepath/two.txt
This copies them to
destpath/one.txt
destpath/two.txt
Not to
destpath/sourcepath/one.txt
destpath/sourcepath/two.txt
If you are familiar with ` + "`" + `rsync` + "`" + `, rclone always works as if you had
written a trailing / - meaning "copy the contents of this directory".
This applies to all commands and whether you are talking about the
source or destination.
See the ` + "`" + `--no-traverse` + "`" + ` option for controlling whether rclone lists
the destination directory or not.
`,
Run: func(cmd *cobra.Command, args []string) { Run: func(cmd *cobra.Command, args []string) {
checkArgs(2, 2, cmd, args) checkArgs(2, 2, cmd, args)
fsrc, fdst := newFsSrcDst(args) fsrc, fdst := newFsSrcDst(args)
@ -257,8 +293,22 @@ var syncCmd = &cobra.Command{
Sync the source to the destination, changing the destination Sync the source to the destination, changing the destination
only. Doesn't transfer unchanged files, testing by size and only. Doesn't transfer unchanged files, testing by size and
modification time or MD5SUM. Destination is updated to match modification time or MD5SUM. Destination is updated to match
source, including deleting files if necessary. Since this can source, including deleting files if necessary.
cause data loss, test first with the --dry-run flag.`,
**Important**: Since this can cause data loss, test first with the
` + "`" + `--dry-run` + "`" + ` flag to see exactly what would be copied and deleted.
Note that files in the destination won't be deleted if there were any
errors at any point.
It is always the contents of the directory that is synced, not the
directory so when source:path is a directory, it's the contents of
source:path that are copied, not the directory name and contents. See
extended explanation in the ` + "`" + `copy` + "`" + ` command above if unsure.
If dest:path doesn't exist, it is created and the source:path contents
go there.
`,
Run: func(cmd *cobra.Command, args []string) { Run: func(cmd *cobra.Command, args []string) {
checkArgs(2, 2, cmd, args) checkArgs(2, 2, cmd, args)
fsrc, fdst := newFsSrcDst(args) fsrc, fdst := newFsSrcDst(args)
@ -272,10 +322,22 @@ var moveCmd = &cobra.Command{
Use: "move source:path dest:path", Use: "move source:path dest:path",
Short: `Move files from source to dest.`, Short: `Move files from source to dest.`,
Long: ` Long: `
Moves the source to the destination. This is equivalent to a Moves the contents of the source directory to the destination
copy followed by a purge, but may use server side operations directory. Rclone will error if the source and destination overlap.
to speed it up. Since this can cause data loss, test first
with the --dry-run flag.`, If no filters are in use and if possible this will server side move
` + "`" + `source:path` + "`" + ` into ` + "`" + `dest:path` + "`" + `. After this ` + "`" + `source:path` + "`" + ` will no
longer longer exist.
Otherwise for each file in ` + "`" + `source:path` + "`" + ` selected by the filters (if
any) this will move it into ` + "`" + `dest:path` + "`" + `. If possible a server side
move will be used, otherwise it will copy it (server side if possible)
into ` + "`" + `dest:path` + "`" + ` then delete the original (if no errors on copy) in
` + "`" + `source:path` + "`" + `.
**Important**: Since this can cause data loss, test first with the
--dry-run flag.
`,
Run: func(cmd *cobra.Command, args []string) { Run: func(cmd *cobra.Command, args []string) {
checkArgs(2, 2, cmd, args) checkArgs(2, 2, cmd, args)
fsrc, fdst := newFsSrcDst(args) fsrc, fdst := newFsSrcDst(args)
@ -326,7 +388,8 @@ var md5sumCmd = &cobra.Command{
Short: `Produces an md5sum file for all the objects in the path.`, Short: `Produces an md5sum file for all the objects in the path.`,
Long: ` Long: `
Produces an md5sum file for all the objects in the path. This Produces an md5sum file for all the objects in the path. This
is in the same format as the standard md5sum tool produces.`, is in the same format as the standard md5sum tool produces.
`,
Run: func(cmd *cobra.Command, args []string) { Run: func(cmd *cobra.Command, args []string) {
checkArgs(1, 1, cmd, args) checkArgs(1, 1, cmd, args)
fsrc := newFsSrc(args) fsrc := newFsSrc(args)
@ -341,7 +404,8 @@ var sha1sumCmd = &cobra.Command{
Short: `Produces an sha1sum file for all the objects in the path.`, Short: `Produces an sha1sum file for all the objects in the path.`,
Long: ` Long: `
Produces an sha1sum file for all the objects in the path. This Produces an sha1sum file for all the objects in the path. This
is in the same format as the standard sha1sum tool produces.`, is in the same format as the standard sha1sum tool produces.
`,
Run: func(cmd *cobra.Command, args []string) { Run: func(cmd *cobra.Command, args []string) {
checkArgs(1, 1, cmd, args) checkArgs(1, 1, cmd, args)
fsrc := newFsSrc(args) fsrc := newFsSrc(args)
@ -353,7 +417,7 @@ is in the same format as the standard sha1sum tool produces.`,
var sizeCmd = &cobra.Command{ var sizeCmd = &cobra.Command{
Use: "size remote:path", Use: "size remote:path",
Short: `Returns the total size and number of objects in remote:path.`, Short: `Prints the total size and number of objects in remote:path.`,
Run: func(cmd *cobra.Command, args []string) { Run: func(cmd *cobra.Command, args []string) {
checkArgs(1, 1, cmd, args) checkArgs(1, 1, cmd, args)
fsrc := newFsSrc(args) fsrc := newFsSrc(args)
@ -383,7 +447,7 @@ var mkdirCmd = &cobra.Command{
var rmdirCmd = &cobra.Command{ var rmdirCmd = &cobra.Command{
Use: "rmdir remote:path", Use: "rmdir remote:path",
Short: `Remove the path.`, Short: `Remove the path if empty.`,
Long: ` Long: `
Remove the path. Note that you can't remove a path with Remove the path. Note that you can't remove a path with
objects in it, use purge for that.`, objects in it, use purge for that.`,
@ -400,8 +464,10 @@ var purgeCmd = &cobra.Command{
Use: "purge remote:path", Use: "purge remote:path",
Short: `Remove the path and all of its contents.`, Short: `Remove the path and all of its contents.`,
Long: ` Long: `
Remove the path and all of its contents. Does not obey Remove the path and all of its contents. Note that this does not obey
filters - use remove for that.`, include/exclude filters - everything will be removed. Use ` + "`" + `delete` + "`" + ` if
you want to selectively delete files.
`,
Run: func(cmd *cobra.Command, args []string) { Run: func(cmd *cobra.Command, args []string) {
checkArgs(1, 1, cmd, args) checkArgs(1, 1, cmd, args)
fdst := newFsDst(args) fdst := newFsDst(args)
@ -415,7 +481,23 @@ var deleteCmd = &cobra.Command{
Use: "delete remote:path", Use: "delete remote:path",
Short: `Remove the contents of path.`, Short: `Remove the contents of path.`,
Long: ` Long: `
Remove the contents of path. Obeys include/exclude filters.`, Remove the contents of path. Unlike ` + "`" + `purge` + "`" + ` it obeys include/exclude
filters so can be used to selectively delete files.
Eg delete all files bigger than 100MBytes
Check what would be deleted first (use either)
rclone --min-size 100M lsl remote:path
rclone --dry-run --min-size 100M delete remote:path
Then delete
rclone --min-size 100M delete remote:path
That reads "delete everything with a minimum size of 100 MB", hence
delete all files bigger than 100MBytes.
`,
Run: func(cmd *cobra.Command, args []string) { Run: func(cmd *cobra.Command, args []string) {
checkArgs(1, 1, cmd, args) checkArgs(1, 1, cmd, args)
fsrc := newFsSrc(args) fsrc := newFsSrc(args)
@ -431,7 +513,10 @@ var checkCmd = &cobra.Command{
Long: ` Long: `
Checks the files in the source and destination match. It Checks the files in the source and destination match. It
compares sizes and MD5SUMs and prints a report of files which compares sizes and MD5SUMs and prints a report of files which
don't match. It doesn't alter the source or destination.`, don't match. It doesn't alter the source or destination.
` + "`" + `--size-only` + "`" + ` may be used to only compare the sizes, not the MD5SUMs.
`,
Run: func(cmd *cobra.Command, args []string) { Run: func(cmd *cobra.Command, args []string) {
checkArgs(2, 2, cmd, args) checkArgs(2, 2, cmd, args)
fsrc, fdst := newFsSrcDst(args) fsrc, fdst := newFsSrcDst(args)
@ -445,18 +530,80 @@ var dedupeCmd = &cobra.Command{
Use: "dedupe [mode] remote:path", Use: "dedupe [mode] remote:path",
Short: `Interactively find duplicate files delete/rename them.`, Short: `Interactively find duplicate files delete/rename them.`,
Long: ` Long: `
Interactively find duplicate files and offer to delete all By default ` + "`" + `dedup` + "`" + ` interactively finds duplicate files and offers to
but one or rename them to be different. Only useful with delete all but one or rename them to be different. Only useful with
Google Drive which can have duplicate file names. Google Drive which can have duplicate file names.
Pass in an extra parameter to make the dedupe noninteractive The ` + "`" + `dedupe` + "`" + ` command will delete all but one of any identical (same
md5sum) files it finds without confirmation. This means that for most
duplicated files the ` + "`" + `dedupe` + "`" + ` command will not be interactive. You
can use ` + "`" + `--dry-run` + "`" + ` to see what would happen without doing anything.
interactive - interactive as above. Here is an example run.
skip - removes identical files then skips anything left.
first - removes identical files then keeps the first one. Before - with duplicates
newest - removes identical files then keeps the newest one.
oldest - removes identical files then keeps the oldest one. $ rclone lsl drive:dupes
rename - removes identical files then renames the rest to be different. 6048320 2016-03-05 16:23:16.798000000 one.txt
6048320 2016-03-05 16:23:11.775000000 one.txt
564374 2016-03-05 16:23:06.731000000 one.txt
6048320 2016-03-05 16:18:26.092000000 one.txt
6048320 2016-03-05 16:22:46.185000000 two.txt
1744073 2016-03-05 16:22:38.104000000 two.txt
564374 2016-03-05 16:22:52.118000000 two.txt
Now the ` + "`" + `dedupe` + "`" + ` session
$ rclone dedupe drive:dupes
2016/03/05 16:24:37 Google drive root 'dupes': Looking for duplicates using interactive mode.
one.txt: Found 4 duplicates - deleting identical copies
one.txt: Deleting 2/3 identical duplicates (md5sum "1eedaa9fe86fd4b8632e2ac549403b36")
one.txt: 2 duplicates remain
1: 6048320 bytes, 2016-03-05 16:23:16.798000000, md5sum 1eedaa9fe86fd4b8632e2ac549403b36
2: 564374 bytes, 2016-03-05 16:23:06.731000000, md5sum 7594e7dc9fc28f727c42ee3e0749de81
s) Skip and do nothing
k) Keep just one (choose which in next step)
r) Rename all to be different (by changing file.jpg to file-1.jpg)
s/k/r> k
Enter the number of the file to keep> 1
one.txt: Deleted 1 extra copies
two.txt: Found 3 duplicates - deleting identical copies
two.txt: 3 duplicates remain
1: 564374 bytes, 2016-03-05 16:22:52.118000000, md5sum 7594e7dc9fc28f727c42ee3e0749de81
2: 6048320 bytes, 2016-03-05 16:22:46.185000000, md5sum 1eedaa9fe86fd4b8632e2ac549403b36
3: 1744073 bytes, 2016-03-05 16:22:38.104000000, md5sum 851957f7fb6f0bc4ce76be966d336802
s) Skip and do nothing
k) Keep just one (choose which in next step)
r) Rename all to be different (by changing file.jpg to file-1.jpg)
s/k/r> r
two-1.txt: renamed from: two.txt
two-2.txt: renamed from: two.txt
two-3.txt: renamed from: two.txt
The result being
$ rclone lsl drive:dupes
6048320 2016-03-05 16:23:16.798000000 one.txt
564374 2016-03-05 16:22:52.118000000 two-1.txt
6048320 2016-03-05 16:22:46.185000000 two-2.txt
1744073 2016-03-05 16:22:38.104000000 two-3.txt
Dedupe can be run non interactively using the ` + "`" + `--dedupe-mode` + "`" + ` flag or by using an extra parameter with the same value
* ` + "`" + `--dedupe-mode interactive` + "`" + ` - interactive as above.
* ` + "`" + `--dedupe-mode skip` + "`" + ` - removes identical files then skips anything left.
* ` + "`" + `--dedupe-mode first` + "`" + ` - removes identical files then keeps the first one.
* ` + "`" + `--dedupe-mode newest` + "`" + ` - removes identical files then keeps the newest one.
* ` + "`" + `--dedupe-mode oldest` + "`" + ` - removes identical files then keeps the oldest one.
* ` + "`" + `--dedupe-mode rename` + "`" + ` - removes identical files then renames the rest to be different.
For example to rename all the identically named photos in your Google Photos directory, do
rclone dedupe --dedupe-mode rename "drive:Google Photos"
Or
rclone dedupe rename "drive:Google Photos"
`, `,
Run: func(cmd *cobra.Command, args []string) { Run: func(cmd *cobra.Command, args []string) {
checkArgs(1, 2, cmd, args) checkArgs(1, 2, cmd, args)
@ -569,8 +716,9 @@ var cleanupCmd = &cobra.Command{
Use: "cleanup remote:path", Use: "cleanup remote:path",
Short: `Clean up the remote if possible`, Short: `Clean up the remote if possible`,
Long: ` Long: `
Clean up the remote if possible. Empty the trash or delete Clean up the remote if possible. Empty the trash or delete old file
old file versions. Not supported by all remotes.`, versions. Not supported by all remotes.
`,
Run: func(cmd *cobra.Command, args []string) { Run: func(cmd *cobra.Command, args []string) {
checkArgs(1, 1, cmd, args) checkArgs(1, 1, cmd, args)
fsrc := newFsSrc(args) fsrc := newFsSrc(args)