mirror of
https://github.com/rclone/rclone.git
synced 2024-12-12 10:12:41 +01:00
610 lines
41 KiB
Markdown
610 lines
41 KiB
Markdown
---
|
|
date: 2018-11-24T13:43:29Z
|
|
title: "rclone mount"
|
|
slug: rclone_mount
|
|
url: /commands/rclone_mount/
|
|
---
|
|
## rclone mount
|
|
|
|
Mount the remote as file system on a mountpoint.
|
|
|
|
### Synopsis
|
|
|
|
|
|
rclone mount allows Linux, FreeBSD, macOS and Windows to
|
|
mount any of Rclone's cloud storage systems as a file system with
|
|
FUSE.
|
|
|
|
First set up your remote using `rclone config`. Check it works with `rclone ls` etc.
|
|
|
|
Start the mount like this
|
|
|
|
rclone mount remote:path/to/files /path/to/local/mount
|
|
|
|
Or on Windows like this where X: is an unused drive letter
|
|
|
|
rclone mount remote:path/to/files X:
|
|
|
|
When the program ends, either via Ctrl+C or receiving a SIGINT or SIGTERM signal,
|
|
the mount is automatically stopped.
|
|
|
|
The umount operation can fail, for example when the mountpoint is busy.
|
|
When that happens, it is the user's responsibility to stop the mount manually with
|
|
|
|
# Linux
|
|
fusermount -u /path/to/local/mount
|
|
# OS X
|
|
umount /path/to/local/mount
|
|
|
|
### Installing on Windows
|
|
|
|
To run rclone mount on Windows, you will need to
|
|
download and install [WinFsp](http://www.secfs.net/winfsp/).
|
|
|
|
WinFsp is an [open source](https://github.com/billziss-gh/winfsp)
|
|
Windows File System Proxy which makes it easy to write user space file
|
|
systems for Windows. It provides a FUSE emulation layer which rclone
|
|
uses combination with
|
|
[cgofuse](https://github.com/billziss-gh/cgofuse). Both of these
|
|
packages are by Bill Zissimopoulos who was very helpful during the
|
|
implementation of rclone mount for Windows.
|
|
|
|
#### Windows caveats
|
|
|
|
Note that drives created as Administrator are not visible by other
|
|
accounts (including the account that was elevated as
|
|
Administrator). So if you start a Windows drive from an Administrative
|
|
Command Prompt and then try to access the same drive from Explorer
|
|
(which does not run as Administrator), you will not be able to see the
|
|
new drive.
|
|
|
|
The easiest way around this is to start the drive from a normal
|
|
command prompt. It is also possible to start a drive from the SYSTEM
|
|
account (using [the WinFsp.Launcher
|
|
infrastructure](https://github.com/billziss-gh/winfsp/wiki/WinFsp-Service-Architecture))
|
|
which creates drives accessible for everyone on the system or
|
|
alternatively using [the nssm service manager](https://nssm.cc/usage).
|
|
|
|
### Limitations
|
|
|
|
Without the use of "--vfs-cache-mode" this can only write files
|
|
sequentially, it can only seek when reading. This means that many
|
|
applications won't work with their files on an rclone mount without
|
|
"--vfs-cache-mode writes" or "--vfs-cache-mode full". See the [File
|
|
Caching](#file-caching) section for more info.
|
|
|
|
The bucket based remotes (eg Swift, S3, Google Compute Storage, B2,
|
|
Hubic) won't work from the root - you will need to specify a bucket,
|
|
or a path within the bucket. So `swift:` won't work whereas
|
|
`swift:bucket` will as will `swift:bucket/path`.
|
|
None of these support the concept of directories, so empty
|
|
directories will have a tendency to disappear once they fall out of
|
|
the directory cache.
|
|
|
|
Only supported on Linux, FreeBSD, OS X and Windows at the moment.
|
|
|
|
### rclone mount vs rclone sync/copy
|
|
|
|
File systems expect things to be 100% reliable, whereas cloud storage
|
|
systems are a long way from 100% reliable. The rclone sync/copy
|
|
commands cope with this with lots of retries. However rclone mount
|
|
can't use retries in the same way without making local copies of the
|
|
uploads. Look at the [file caching](#file-caching)
|
|
for solutions to make mount more reliable.
|
|
|
|
### Attribute caching
|
|
|
|
You can use the flag --attr-timeout to set the time the kernel caches
|
|
the attributes (size, modification time etc) for directory entries.
|
|
|
|
The default is "1s" which caches files just long enough to avoid
|
|
too many callbacks to rclone from the kernel.
|
|
|
|
In theory 0s should be the correct value for filesystems which can
|
|
change outside the control of the kernel. However this causes quite a
|
|
few problems such as
|
|
[rclone using too much memory](https://github.com/ncw/rclone/issues/2157),
|
|
[rclone not serving files to samba](https://forum.rclone.org/t/rclone-1-39-vs-1-40-mount-issue/5112)
|
|
and [excessive time listing directories](https://github.com/ncw/rclone/issues/2095#issuecomment-371141147).
|
|
|
|
The kernel can cache the info about a file for the time given by
|
|
"--attr-timeout". You may see corruption if the remote file changes
|
|
length during this window. It will show up as either a truncated file
|
|
or a file with garbage on the end. With "--attr-timeout 1s" this is
|
|
very unlikely but not impossible. The higher you set "--attr-timeout"
|
|
the more likely it is. The default setting of "1s" is the lowest
|
|
setting which mitigates the problems above.
|
|
|
|
If you set it higher ('10s' or '1m' say) then the kernel will call
|
|
back to rclone less often making it more efficient, however there is
|
|
more chance of the corruption issue above.
|
|
|
|
If files don't change on the remote outside of the control of rclone
|
|
then there is no chance of corruption.
|
|
|
|
This is the same as setting the attr_timeout option in mount.fuse.
|
|
|
|
### Filters
|
|
|
|
Note that all the rclone filters can be used to select a subset of the
|
|
files to be visible in the mount.
|
|
|
|
### systemd
|
|
|
|
When running rclone mount as a systemd service, it is possible
|
|
to use Type=notify. In this case the service will enter the started state
|
|
after the mountpoint has been successfully set up.
|
|
Units having the rclone mount service specified as a requirement
|
|
will see all files and folders immediately in this mode.
|
|
|
|
### chunked reading ###
|
|
|
|
--vfs-read-chunk-size will enable reading the source objects in parts.
|
|
This can reduce the used download quota for some remotes by requesting only chunks
|
|
from the remote that are actually read at the cost of an increased number of requests.
|
|
|
|
When --vfs-read-chunk-size-limit is also specified and greater than --vfs-read-chunk-size,
|
|
the chunk size for each open file will get doubled for each chunk read, until the
|
|
specified value is reached. A value of -1 will disable the limit and the chunk size will
|
|
grow indefinitely.
|
|
|
|
With --vfs-read-chunk-size 100M and --vfs-read-chunk-size-limit 0 the following
|
|
parts will be downloaded: 0-100M, 100M-200M, 200M-300M, 300M-400M and so on.
|
|
When --vfs-read-chunk-size-limit 500M is specified, the result would be
|
|
0-100M, 100M-300M, 300M-700M, 700M-1200M, 1200M-1700M and so on.
|
|
|
|
Chunked reading will only work with --vfs-cache-mode < full, as the file will always
|
|
be copied to the vfs cache before opening with --vfs-cache-mode full.
|
|
|
|
### Directory Cache
|
|
|
|
Using the `--dir-cache-time` flag, you can set how long a
|
|
directory should be considered up to date and not refreshed from the
|
|
backend. Changes made locally in the mount may appear immediately or
|
|
invalidate the cache. However, changes done on the remote will only
|
|
be picked up once the cache expires.
|
|
|
|
Alternatively, you can send a `SIGHUP` signal to rclone for
|
|
it to flush all directory caches, regardless of how old they are.
|
|
Assuming only one rclone instance is running, you can reset the cache
|
|
like this:
|
|
|
|
kill -SIGHUP $(pidof rclone)
|
|
|
|
If you configure rclone with a [remote control](/rc) then you can use
|
|
rclone rc to flush the whole directory cache:
|
|
|
|
rclone rc vfs/forget
|
|
|
|
Or individual files or directories:
|
|
|
|
rclone rc vfs/forget file=path/to/file dir=path/to/dir
|
|
|
|
### File Buffering
|
|
|
|
The `--buffer-size` flag determines the amount of memory,
|
|
that will be used to buffer data in advance.
|
|
|
|
Each open file descriptor will try to keep the specified amount of
|
|
data in memory at all times. The buffered data is bound to one file
|
|
descriptor and won't be shared between multiple open file descriptors
|
|
of the same file.
|
|
|
|
This flag is a upper limit for the used memory per file descriptor.
|
|
The buffer will only use memory for data that is downloaded but not
|
|
not yet read. If the buffer is empty, only a small amount of memory
|
|
will be used.
|
|
The maximum memory used by rclone for buffering can be up to
|
|
`--buffer-size * open files`.
|
|
|
|
### File Caching
|
|
|
|
These flags control the VFS file caching options. The VFS layer is
|
|
used by rclone mount to make a cloud storage system work more like a
|
|
normal file system.
|
|
|
|
You'll need to enable VFS caching if you want, for example, to read
|
|
and write simultaneously to a file. See below for more details.
|
|
|
|
Note that the VFS cache works in addition to the cache backend and you
|
|
may find that you need one or the other or both.
|
|
|
|
--cache-dir string Directory rclone will use for caching.
|
|
--vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s)
|
|
--vfs-cache-mode string Cache mode off|minimal|writes|full (default "off")
|
|
--vfs-cache-poll-interval duration Interval to poll the cache for stale objects. (default 1m0s)
|
|
|
|
If run with `-vv` rclone will print the location of the file cache. The
|
|
files are stored in the user cache file area which is OS dependent but
|
|
can be controlled with `--cache-dir` or setting the appropriate
|
|
environment variable.
|
|
|
|
The cache has 4 different modes selected by `--vfs-cache-mode`.
|
|
The higher the cache mode the more compatible rclone becomes at the
|
|
cost of using disk space.
|
|
|
|
Note that files are written back to the remote only when they are
|
|
closed so if rclone is quit or dies with open files then these won't
|
|
get written back to the remote. However they will still be in the on
|
|
disk cache.
|
|
|
|
#### --vfs-cache-mode off
|
|
|
|
In this mode the cache will read directly from the remote and write
|
|
directly to the remote without caching anything on disk.
|
|
|
|
This will mean some operations are not possible
|
|
|
|
* Files can't be opened for both read AND write
|
|
* Files opened for write can't be seeked
|
|
* Existing files opened for write must have O_TRUNC set
|
|
* Files open for read with O_TRUNC will be opened write only
|
|
* Files open for write only will behave as if O_TRUNC was supplied
|
|
* Open modes O_APPEND, O_TRUNC are ignored
|
|
* If an upload fails it can't be retried
|
|
|
|
#### --vfs-cache-mode minimal
|
|
|
|
This is very similar to "off" except that files opened for read AND
|
|
write will be buffered to disks. This means that files opened for
|
|
write will be a lot more compatible, but uses the minimal disk space.
|
|
|
|
These operations are not possible
|
|
|
|
* Files opened for write only can't be seeked
|
|
* Existing files opened for write must have O_TRUNC set
|
|
* Files opened for write only will ignore O_APPEND, O_TRUNC
|
|
* If an upload fails it can't be retried
|
|
|
|
#### --vfs-cache-mode writes
|
|
|
|
In this mode files opened for read only are still read directly from
|
|
the remote, write only and read/write files are buffered to disk
|
|
first.
|
|
|
|
This mode should support all normal file system operations.
|
|
|
|
If an upload fails it will be retried up to --low-level-retries times.
|
|
|
|
#### --vfs-cache-mode full
|
|
|
|
In this mode all reads and writes are buffered to and from disk. When
|
|
a file is opened for read it will be downloaded in its entirety first.
|
|
|
|
This may be appropriate for your needs, or you may prefer to look at
|
|
the cache backend which does a much more sophisticated job of caching,
|
|
including caching directory hierarchies and chunks of files.
|
|
|
|
In this mode, unlike the others, when a file is written to the disk,
|
|
it will be kept on the disk after it is written to the remote. It
|
|
will be purged on a schedule according to `--vfs-cache-max-age`.
|
|
|
|
This mode should support all normal file system operations.
|
|
|
|
If an upload or download fails it will be retried up to
|
|
--low-level-retries times.
|
|
|
|
|
|
```
|
|
rclone mount remote:path /path/to/mountpoint [flags]
|
|
```
|
|
|
|
### Options
|
|
|
|
```
|
|
--allow-non-empty Allow mounting over a non-empty directory.
|
|
--allow-other Allow access to other users.
|
|
--allow-root Allow access to root user.
|
|
--attr-timeout duration Time for which file/directory attributes are cached. (default 1s)
|
|
--daemon Run mount as a daemon (background mode).
|
|
--daemon-timeout duration Time limit for rclone to respond to kernel (not supported by all OSes).
|
|
--debug-fuse Debug the FUSE internals - needs -v.
|
|
--default-permissions Makes kernel enforce access control based on the file mode.
|
|
--dir-cache-time duration Time to cache directory entries for. (default 5m0s)
|
|
--fuse-flag stringArray Flags or arguments to be passed direct to libfuse/WinFsp. Repeat if required.
|
|
--gid uint32 Override the gid field set by the filesystem. (default 502)
|
|
-h, --help help for mount
|
|
--max-read-ahead int The number of bytes that can be prefetched for sequential reads. (default 128k)
|
|
--no-checksum Don't compare checksums on up/download.
|
|
--no-modtime Don't read/write the modification time (can speed things up).
|
|
--no-seek Don't allow seeking in files.
|
|
-o, --option stringArray Option for libfuse/WinFsp. Repeat if required.
|
|
--poll-interval duration Time to wait between polling for changes. Must be smaller than dir-cache-time. Only on supported remotes. Set to 0 to disable. (default 1m0s)
|
|
--read-only Mount read-only.
|
|
--uid uint32 Override the uid field set by the filesystem. (default 502)
|
|
--umask int Override the permission bits set by the filesystem.
|
|
--vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s)
|
|
--vfs-cache-mode string Cache mode off|minimal|writes|full (default "off")
|
|
--vfs-cache-poll-interval duration Interval to poll the cache for stale objects. (default 1m0s)
|
|
--vfs-read-chunk-size int Read the source objects in chunks. (default 128M)
|
|
--vfs-read-chunk-size-limit int If greater than --vfs-read-chunk-size, double the chunk size after each chunk read, until the limit is reached. 'off' is unlimited. (default off)
|
|
--volname string Set the volume name (not supported by all OSes).
|
|
--write-back-cache Makes kernel buffer writes before sending them to rclone. Without this, writethrough caching is used.
|
|
```
|
|
|
|
### Options inherited from parent commands
|
|
|
|
```
|
|
--acd-auth-url string Auth server URL.
|
|
--acd-client-id string Amazon Application Client ID.
|
|
--acd-client-secret string Amazon Application Client Secret.
|
|
--acd-templink-threshold SizeSuffix Files >= this size will be downloaded via their tempLink. (default 9G)
|
|
--acd-token-url string Token server url.
|
|
--acd-upload-wait-per-gb Duration Additional time per GB to wait after a failed complete upload to see if it appears. (default 3m0s)
|
|
--alias-remote string Remote or path to alias.
|
|
--ask-password Allow prompt for password for encrypted configuration. (default true)
|
|
--auto-confirm If enabled, do not request console confirmation.
|
|
--azureblob-access-tier string Access tier of blob: hot, cool or archive.
|
|
--azureblob-account string Storage Account Name (leave blank to use connection string or SAS URL)
|
|
--azureblob-chunk-size SizeSuffix Upload chunk size (<= 100MB). (default 4M)
|
|
--azureblob-endpoint string Endpoint for the service
|
|
--azureblob-key string Storage Account Key (leave blank to use connection string or SAS URL)
|
|
--azureblob-list-chunk int Size of blob list. (default 5000)
|
|
--azureblob-sas-url string SAS URL for container level access only
|
|
--azureblob-upload-cutoff SizeSuffix Cutoff for switching to chunked upload (<= 256MB). (default 256M)
|
|
--b2-account string Account ID or Application Key ID
|
|
--b2-chunk-size SizeSuffix Upload chunk size. Must fit in memory. (default 96M)
|
|
--b2-endpoint string Endpoint for the service.
|
|
--b2-hard-delete Permanently delete files on remote removal, otherwise hide files.
|
|
--b2-key string Application Key
|
|
--b2-test-mode string A flag string for X-Bz-Test-Mode header for debugging.
|
|
--b2-upload-cutoff SizeSuffix Cutoff for switching to chunked upload. (default 200M)
|
|
--b2-versions Include old versions in directory listings.
|
|
--backup-dir string Make backups into hierarchy based in DIR.
|
|
--bind string Local address to bind to for outgoing connections, IPv4, IPv6 or name.
|
|
--box-client-id string Box App Client Id.
|
|
--box-client-secret string Box App Client Secret
|
|
--box-commit-retries int Max number of times to try committing a multipart file. (default 100)
|
|
--box-upload-cutoff SizeSuffix Cutoff for switching to multipart upload (>= 50MB). (default 50M)
|
|
--buffer-size int In memory buffer size when reading files for each --transfer. (default 16M)
|
|
--bwlimit BwTimetable Bandwidth limit in kBytes/s, or use suffix b|k|M|G or a full timetable.
|
|
--cache-chunk-clean-interval Duration How often should the cache perform cleanups of the chunk storage. (default 1m0s)
|
|
--cache-chunk-no-memory Disable the in-memory cache for storing chunks during streaming.
|
|
--cache-chunk-path string Directory to cache chunk files. (default "$HOME/.cache/rclone/cache-backend")
|
|
--cache-chunk-size SizeSuffix The size of a chunk (partial file data). (default 5M)
|
|
--cache-chunk-total-size SizeSuffix The total size that the chunks can take up on the local disk. (default 10G)
|
|
--cache-db-path string Directory to store file structure metadata DB. (default "$HOME/.cache/rclone/cache-backend")
|
|
--cache-db-purge Clear all the cached data for this remote on start.
|
|
--cache-db-wait-time Duration How long to wait for the DB to be available - 0 is unlimited (default 1s)
|
|
--cache-dir string Directory rclone will use for caching. (default "$HOME/.cache/rclone")
|
|
--cache-info-age Duration How long to cache file structure information (directory listings, file size, times etc). (default 6h0m0s)
|
|
--cache-plex-insecure string Skip all certificate verifications when connecting to the Plex server
|
|
--cache-plex-password string The password of the Plex user
|
|
--cache-plex-url string The URL of the Plex server
|
|
--cache-plex-username string The username of the Plex user
|
|
--cache-read-retries int How many times to retry a read from a cache storage. (default 10)
|
|
--cache-remote string Remote to cache.
|
|
--cache-rps int Limits the number of requests per second to the source FS (-1 to disable) (default -1)
|
|
--cache-tmp-upload-path string Directory to keep temporary files until they are uploaded.
|
|
--cache-tmp-wait-time Duration How long should files be stored in local cache before being uploaded (default 15s)
|
|
--cache-workers int How many workers should run in parallel to download chunks. (default 4)
|
|
--cache-writes Cache file data on writes through the FS
|
|
--checkers int Number of checkers to run in parallel. (default 8)
|
|
-c, --checksum Skip based on checksum & size, not mod-time & size
|
|
--config string Config file. (default "/home/ncw/.rclone.conf")
|
|
--contimeout duration Connect timeout (default 1m0s)
|
|
-L, --copy-links Follow symlinks and copy the pointed to item.
|
|
--cpuprofile string Write cpu profile to file
|
|
--crypt-directory-name-encryption Option to either encrypt directory names or leave them intact. (default true)
|
|
--crypt-filename-encryption string How to encrypt the filenames. (default "standard")
|
|
--crypt-password string Password or pass phrase for encryption.
|
|
--crypt-password2 string Password or pass phrase for salt. Optional but recommended.
|
|
--crypt-remote string Remote to encrypt/decrypt.
|
|
--crypt-show-mapping For all files listed show how the names encrypt.
|
|
--delete-after When synchronizing, delete files on destination after transferring (default)
|
|
--delete-before When synchronizing, delete files on destination before transferring
|
|
--delete-during When synchronizing, delete files during transfer
|
|
--delete-excluded Delete files on dest excluded from sync
|
|
--disable string Disable a comma separated list of features. Use help to see a list.
|
|
--drive-acknowledge-abuse Set to allow files which return cannotDownloadAbusiveFile to be downloaded.
|
|
--drive-allow-import-name-change Allow the filetype to change when uploading Google docs (e.g. file.doc to file.docx). This will confuse sync and reupload every time.
|
|
--drive-alternate-export Use alternate export URLs for google documents export.,
|
|
--drive-auth-owner-only Only consider files owned by the authenticated user.
|
|
--drive-chunk-size SizeSuffix Upload chunk size. Must a power of 2 >= 256k. (default 8M)
|
|
--drive-client-id string Google Application Client Id
|
|
--drive-client-secret string Google Application Client Secret
|
|
--drive-export-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
|
|
--drive-formats string Deprecated: see export_formats
|
|
--drive-impersonate string Impersonate this user when using a service account.
|
|
--drive-import-formats string Comma separated list of preferred formats for uploading Google docs.
|
|
--drive-keep-revision-forever Keep new head revision of each file forever.
|
|
--drive-list-chunk int Size of listing chunk 100-1000. 0 to disable. (default 1000)
|
|
--drive-root-folder-id string ID of the root folder
|
|
--drive-scope string Scope that rclone should use when requesting access from drive.
|
|
--drive-service-account-credentials string Service Account Credentials JSON blob
|
|
--drive-service-account-file string Service Account Credentials JSON file path
|
|
--drive-shared-with-me Only show files that are shared with me.
|
|
--drive-skip-gdocs Skip google documents in all listings.
|
|
--drive-team-drive string ID of the Team Drive
|
|
--drive-trashed-only Only show files that are in the trash.
|
|
--drive-upload-cutoff SizeSuffix Cutoff for switching to chunked upload (default 8M)
|
|
--drive-use-created-date Use file created date instead of modified date.,
|
|
--drive-use-trash Send files to the trash instead of deleting permanently. (default true)
|
|
--drive-v2-download-min-size SizeSuffix If Object's are greater, use drive v2 API to download. (default off)
|
|
--dropbox-chunk-size SizeSuffix Upload chunk size. (< 150M). (default 48M)
|
|
--dropbox-client-id string Dropbox App Client Id
|
|
--dropbox-client-secret string Dropbox App Client Secret
|
|
--dropbox-impersonate string Impersonate this user when using a business account.
|
|
-n, --dry-run Do a trial run with no permanent changes
|
|
--dump string List of items to dump from: headers,bodies,requests,responses,auth,filters,goroutines,openfiles
|
|
--dump-bodies Dump HTTP headers and bodies - may contain sensitive info
|
|
--dump-headers Dump HTTP bodies - may contain sensitive info
|
|
--exclude stringArray Exclude files matching pattern
|
|
--exclude-from stringArray Read exclude patterns from file
|
|
--exclude-if-present string Exclude directories if filename is present
|
|
--fast-list Use recursive list if available. Uses more memory but fewer transactions.
|
|
--files-from stringArray Read list of source-file names from file
|
|
-f, --filter stringArray Add a file-filtering rule
|
|
--filter-from stringArray Read filtering patterns from a file
|
|
--ftp-host string FTP host to connect to
|
|
--ftp-pass string FTP password
|
|
--ftp-port string FTP port, leave blank to use default (21)
|
|
--ftp-user string FTP username, leave blank for current username, $USER
|
|
--gcs-bucket-acl string Access Control List for new buckets.
|
|
--gcs-client-id string Google Application Client Id
|
|
--gcs-client-secret string Google Application Client Secret
|
|
--gcs-location string Location for the newly created buckets.
|
|
--gcs-object-acl string Access Control List for new objects.
|
|
--gcs-project-number string Project number.
|
|
--gcs-service-account-file string Service Account Credentials JSON file path
|
|
--gcs-storage-class string The storage class to use when storing objects in Google Cloud Storage.
|
|
--http-url string URL of http host to connect to
|
|
--hubic-chunk-size SizeSuffix Above this size files will be chunked into a _segments container. (default 5G)
|
|
--hubic-client-id string Hubic Client Id
|
|
--hubic-client-secret string Hubic Client Secret
|
|
--ignore-case Ignore case in filters (case insensitive)
|
|
--ignore-checksum Skip post copy check of checksums.
|
|
--ignore-errors delete even if there are I/O errors
|
|
--ignore-existing Skip all files that exist on destination
|
|
--ignore-size Ignore size when skipping use mod-time or checksum.
|
|
-I, --ignore-times Don't skip files that match size and time - transfer all files
|
|
--immutable Do not modify files. Fail if existing files have been modified.
|
|
--include stringArray Include files matching pattern
|
|
--include-from stringArray Read include patterns from file
|
|
--jottacloud-hard-delete Delete files permanently rather than putting them into the trash.
|
|
--jottacloud-md5-memory-limit SizeSuffix Files bigger than this will be cached on disk to calculate the MD5 if required. (default 10M)
|
|
--jottacloud-mountpoint string The mountpoint to use.
|
|
--jottacloud-pass string Password.
|
|
--jottacloud-unlink Remove existing public link to file/folder with link command rather than creating.
|
|
--jottacloud-user string User Name
|
|
--local-no-check-updated Don't check to see if the files change during upload
|
|
--local-no-unicode-normalization Don't apply unicode normalization to paths and filenames (Deprecated)
|
|
--local-nounc string Disable UNC (long path names) conversion on Windows
|
|
--log-file string Log everything to this file
|
|
--log-format string Comma separated list of log format options (default "date,time")
|
|
--log-level string Log level DEBUG|INFO|NOTICE|ERROR (default "NOTICE")
|
|
--low-level-retries int Number of low level retries to do. (default 10)
|
|
--max-age duration Only transfer files younger than this in s or suffix ms|s|m|h|d|w|M|y (default off)
|
|
--max-backlog int Maximum number of objects in sync or check backlog. (default 10000)
|
|
--max-delete int When synchronizing, limit the number of deletes (default -1)
|
|
--max-depth int If set limits the recursion depth to this. (default -1)
|
|
--max-size int Only transfer files smaller than this in k or suffix b|k|M|G (default off)
|
|
--max-transfer int Maximum size of data to transfer. (default off)
|
|
--mega-debug Output more debug from Mega.
|
|
--mega-hard-delete Delete files permanently rather than putting them into the trash.
|
|
--mega-pass string Password.
|
|
--mega-user string User name
|
|
--memprofile string Write memory profile to file
|
|
--min-age duration Only transfer files older than this in s or suffix ms|s|m|h|d|w|M|y (default off)
|
|
--min-size int Only transfer files bigger than this in k or suffix b|k|M|G (default off)
|
|
--modify-window duration Max time diff to be considered the same (default 1ns)
|
|
--no-check-certificate Do not verify the server SSL certificate. Insecure.
|
|
--no-gzip-encoding Don't set Accept-Encoding: gzip.
|
|
--no-traverse Obsolete - does nothing.
|
|
--no-update-modtime Don't update destination mod-time if files identical.
|
|
-x, --one-file-system Don't cross filesystem boundaries (unix/macOS only).
|
|
--onedrive-chunk-size SizeSuffix Chunk size to upload files with - must be multiple of 320k. (default 10M)
|
|
--onedrive-client-id string Microsoft App Client Id
|
|
--onedrive-client-secret string Microsoft App Client Secret
|
|
--onedrive-drive-id string The ID of the drive to use
|
|
--onedrive-drive-type string The type of the drive ( personal | business | documentLibrary )
|
|
--onedrive-expose-onenote-files Set to make OneNote files show up in directory listings.
|
|
--opendrive-password string Password.
|
|
--opendrive-username string Username
|
|
--pcloud-client-id string Pcloud App Client Id
|
|
--pcloud-client-secret string Pcloud App Client Secret
|
|
-P, --progress Show progress during transfer.
|
|
--qingstor-access-key-id string QingStor Access Key ID
|
|
--qingstor-connection-retries int Number of connection retries. (default 3)
|
|
--qingstor-endpoint string Enter a endpoint URL to connection QingStor API.
|
|
--qingstor-env-auth Get QingStor credentials from runtime. Only applies if access_key_id and secret_access_key is blank.
|
|
--qingstor-secret-access-key string QingStor Secret Access Key (password)
|
|
--qingstor-zone string Zone to connect to.
|
|
-q, --quiet Print as little stuff as possible
|
|
--rc Enable the remote control server.
|
|
--rc-addr string IPaddress:Port or :Port to bind server to. (default "localhost:5572")
|
|
--rc-cert string SSL PEM key (concatenation of certificate and CA certificate)
|
|
--rc-client-ca string Client certificate authority to verify clients with
|
|
--rc-files string Path to local files to serve on the HTTP server.
|
|
--rc-htpasswd string htpasswd file - if not provided no authentication is done
|
|
--rc-key string SSL PEM Private key
|
|
--rc-max-header-bytes int Maximum size of request header (default 4096)
|
|
--rc-no-auth Don't require auth for certain methods.
|
|
--rc-pass string Password for authentication.
|
|
--rc-realm string realm for authentication (default "rclone")
|
|
--rc-serve Enable the serving of remote objects.
|
|
--rc-server-read-timeout duration Timeout for server reading data (default 1h0m0s)
|
|
--rc-server-write-timeout duration Timeout for server writing data (default 1h0m0s)
|
|
--rc-user string User name for authentication.
|
|
--retries int Retry operations this many times if they fail (default 3)
|
|
--retries-sleep duration Interval between retrying operations if they fail, e.g 500ms, 60s, 5m. (0 to disable)
|
|
--s3-access-key-id string AWS Access Key ID.
|
|
--s3-acl string Canned ACL used when creating buckets and storing or copying objects.
|
|
--s3-chunk-size SizeSuffix Chunk size to use for uploading. (default 5M)
|
|
--s3-disable-checksum Don't store MD5 checksum with object metadata
|
|
--s3-endpoint string Endpoint for S3 API.
|
|
--s3-env-auth Get AWS credentials from runtime (environment variables or EC2/ECS meta data if no env vars).
|
|
--s3-force-path-style If true use path style access if false use virtual hosted style. (default true)
|
|
--s3-location-constraint string Location constraint - must be set to match the Region.
|
|
--s3-provider string Choose your S3 provider.
|
|
--s3-region string Region to connect to.
|
|
--s3-secret-access-key string AWS Secret Access Key (password)
|
|
--s3-server-side-encryption string The server-side encryption algorithm used when storing this object in S3.
|
|
--s3-session-token string An AWS session token
|
|
--s3-sse-kms-key-id string If using KMS ID you must provide the ARN of Key.
|
|
--s3-storage-class string The storage class to use when storing new objects in S3.
|
|
--s3-upload-concurrency int Concurrency for multipart uploads. (default 2)
|
|
--s3-v2-auth If true use v2 authentication.
|
|
--sftp-ask-password Allow asking for SFTP password when needed.
|
|
--sftp-disable-hashcheck Disable the execution of SSH commands to determine if remote file hashing is available.
|
|
--sftp-host string SSH host to connect to
|
|
--sftp-key-file string Path to unencrypted PEM-encoded private key file, leave blank to use ssh-agent.
|
|
--sftp-pass string SSH password, leave blank to use ssh-agent.
|
|
--sftp-path-override string Override path used by SSH connection.
|
|
--sftp-port string SSH port, leave blank to use default (22)
|
|
--sftp-set-modtime Set the modified time on the remote if set. (default true)
|
|
--sftp-use-insecure-cipher Enable the use of the aes128-cbc cipher. This cipher is insecure and may allow plaintext data to be recovered by an attacker.
|
|
--sftp-user string SSH username, leave blank for current username, ncw
|
|
--size-only Skip based on size only, not mod-time or checksum
|
|
--skip-links Don't warn about skipped symlinks.
|
|
--stats duration Interval between printing stats, e.g 500ms, 60s, 5m. (0 to disable) (default 1m0s)
|
|
--stats-file-name-length int Max file name length in stats. 0 for no limit (default 40)
|
|
--stats-log-level string Log level to show --stats output DEBUG|INFO|NOTICE|ERROR (default "INFO")
|
|
--stats-one-line Make the stats fit on one line.
|
|
--stats-unit string Show data rate in stats as either 'bits' or 'bytes'/s (default "bytes")
|
|
--streaming-upload-cutoff int Cutoff for switching to chunked upload if file size is unknown. Upload starts after reaching cutoff or when file ends. (default 100k)
|
|
--suffix string Suffix for use with --backup-dir.
|
|
--swift-auth string Authentication URL for server (OS_AUTH_URL).
|
|
--swift-auth-token string Auth Token from alternate authentication - optional (OS_AUTH_TOKEN)
|
|
--swift-auth-version int AuthVersion - optional - set to (1,2,3) if your auth URL has no version (ST_AUTH_VERSION)
|
|
--swift-chunk-size SizeSuffix Above this size files will be chunked into a _segments container. (default 5G)
|
|
--swift-domain string User domain - optional (v3 auth) (OS_USER_DOMAIN_NAME)
|
|
--swift-endpoint-type string Endpoint type to choose from the service catalogue (OS_ENDPOINT_TYPE) (default "public")
|
|
--swift-env-auth Get swift credentials from environment variables in standard OpenStack form.
|
|
--swift-key string API key or password (OS_PASSWORD).
|
|
--swift-region string Region name - optional (OS_REGION_NAME)
|
|
--swift-storage-policy string The storage policy to use when creating a new container
|
|
--swift-storage-url string Storage URL - optional (OS_STORAGE_URL)
|
|
--swift-tenant string Tenant name - optional for v1 auth, this or tenant_id required otherwise (OS_TENANT_NAME or OS_PROJECT_NAME)
|
|
--swift-tenant-domain string Tenant domain - optional (v3 auth) (OS_PROJECT_DOMAIN_NAME)
|
|
--swift-tenant-id string Tenant ID - optional for v1 auth, this or tenant required otherwise (OS_TENANT_ID)
|
|
--swift-user string User name to log in (OS_USERNAME).
|
|
--swift-user-id string User ID to log in - optional - most swift systems use user and leave this blank (v3 auth) (OS_USER_ID).
|
|
--syslog Use Syslog for logging
|
|
--syslog-facility string Facility for syslog, eg KERN,USER,... (default "DAEMON")
|
|
--timeout duration IO idle timeout (default 5m0s)
|
|
--tpslimit float Limit HTTP transactions per second to this.
|
|
--tpslimit-burst int Max burst of transactions for --tpslimit. (default 1)
|
|
--track-renames When synchronizing, track file renames and do a server side move if possible
|
|
--transfers int Number of file transfers to run in parallel. (default 4)
|
|
--union-remotes string List of space separated remotes.
|
|
-u, --update Skip files that are newer on the destination.
|
|
--use-server-modtime Use server modified time instead of object metadata
|
|
--user-agent string Set the user-agent to a specified string. The default is rclone/ version (default "rclone/v1.45")
|
|
-v, --verbose count Print lots more stuff (repeat for more)
|
|
--webdav-bearer-token string Bearer token instead of user/pass (eg a Macaroon)
|
|
--webdav-pass string Password.
|
|
--webdav-url string URL of http host to connect to
|
|
--webdav-user string User name
|
|
--webdav-vendor string Name of the Webdav site/service/software you are using
|
|
--yandex-client-id string Yandex Client Id
|
|
--yandex-client-secret string Yandex Client Secret
|
|
--yandex-unlink Remove existing public link to file/folder with link command rather than creating.
|
|
```
|
|
|
|
### SEE ALSO
|
|
|
|
* [rclone](/commands/rclone/) - Show help for rclone commands, flags and backends.
|
|
|
|
###### Auto generated by spf13/cobra on 24-Nov-2018
|