mirror of
https://github.com/rclone/rclone.git
synced 2024-11-28 03:14:55 +01:00
docs: auto generate backend options documentation
This inserts the output of "rclone help backend xxx" into the help pages for each backend.
This commit is contained in:
parent
a9273c5da5
commit
78b9bd77f5
@ -182,10 +182,14 @@ with modules beneath.
|
||||
|
||||
If you are adding a new feature then please update the documentation.
|
||||
|
||||
If you add a new flag, then if it is a general flag, document it in
|
||||
If you add a new general flag (not for a backend), then document it in
|
||||
`docs/content/docs.md` - the flags there are supposed to be in
|
||||
alphabetical order. If it is a remote specific flag, then document it
|
||||
in `docs/content/remote.md`.
|
||||
alphabetical order.
|
||||
|
||||
If you add a new backend option/flag, then it should be documented in
|
||||
the source file in the `Help:` field. The first line of this is used
|
||||
for the flag help, the remainder is shown to the user in `rclone
|
||||
config` and is added to the docs with `make backenddocs`.
|
||||
|
||||
The only documentation you need to edit are the `docs/content/*.md`
|
||||
files. The MANUAL.*, rclone.1, web site etc are all auto generated
|
||||
@ -355,7 +359,7 @@ See the [testing](#testing) section for more information on integration tests.
|
||||
Add your fs to the docs - you'll need to pick an icon for it from [fontawesome](http://fontawesome.io/icons/). Keep lists of remotes in alphabetical order but with the local file system last.
|
||||
|
||||
* `README.md` - main Github page
|
||||
* `docs/content/remote.md` - main docs page
|
||||
* `docs/content/remote.md` - main docs page (note the backend options are automatically added to this file with `make backenddocs`)
|
||||
* `docs/content/overview.md` - overview docs
|
||||
* `docs/content/docs.md` - list of remotes in config section
|
||||
* `docs/content/about.md` - front page of rclone.org
|
||||
|
5
Makefile
5
Makefile
@ -107,7 +107,7 @@ doc: rclone.1 MANUAL.html MANUAL.txt rcdocs commanddocs
|
||||
rclone.1: MANUAL.md
|
||||
pandoc -s --from markdown --to man MANUAL.md -o rclone.1
|
||||
|
||||
MANUAL.md: bin/make_manual.py docs/content/*.md commanddocs
|
||||
MANUAL.md: bin/make_manual.py docs/content/*.md commanddocs backenddocs
|
||||
./bin/make_manual.py
|
||||
|
||||
MANUAL.html: MANUAL.md
|
||||
@ -119,6 +119,9 @@ MANUAL.txt: MANUAL.md
|
||||
commanddocs: rclone
|
||||
rclone gendocs docs/content/commands/
|
||||
|
||||
backenddocs: rclone bin/make_backend_docs.py
|
||||
./bin/make_backend_docs.py
|
||||
|
||||
rcdocs: rclone
|
||||
bin/make_rc_docs.sh
|
||||
|
||||
|
67
bin/make_backend_docs.py
Executable file
67
bin/make_backend_docs.py
Executable file
@ -0,0 +1,67 @@
|
||||
#!/usr/bin/env python
|
||||
"""
|
||||
Make backend documentation
|
||||
"""
|
||||
|
||||
import os
|
||||
import subprocess
|
||||
|
||||
marker = "<!--- autogenerated options"
|
||||
start = marker + " start"
|
||||
stop = marker + " stop"
|
||||
|
||||
# directory name to backend name
|
||||
dir_to_backend = {
|
||||
"googlecloudstorage": "google cloud storage",
|
||||
"amazonclouddrive": "amazon cloud drive",
|
||||
}
|
||||
|
||||
def find_backends():
|
||||
"""Return a list of all backends"""
|
||||
return [ x for x in os.listdir("backend") if x not in ("all",) ]
|
||||
|
||||
def output_docs(backend, out):
|
||||
"""Output documentation for backend options to out"""
|
||||
backend = dir_to_backend.get(backend, backend)
|
||||
out.flush()
|
||||
subprocess.check_call(["rclone", "help", "backend", backend], stdout=out)
|
||||
|
||||
def alter_doc(backend):
|
||||
"""Alter the documentation for backend"""
|
||||
doc_file = "docs/content/"+backend+".md"
|
||||
if not os.path.exists(doc_file):
|
||||
raise ValueError("Didn't find doc file %s" % (doc_file,))
|
||||
new_file = doc_file+"~new~"
|
||||
altered = False
|
||||
with open(doc_file, "r") as in_file, open(new_file, "w") as out_file:
|
||||
in_docs = False
|
||||
for line in in_file:
|
||||
if not in_docs:
|
||||
if start in line:
|
||||
in_docs = True
|
||||
start_full = start + " - DO NOT EDIT, instead edit fs.RegInfo in backend/%s/%s.go then run make backenddocs -->\n" % (backend, backend)
|
||||
out_file.write(start_full)
|
||||
output_docs(backend, out_file)
|
||||
out_file.write(stop+" -->\n")
|
||||
altered = True
|
||||
if not in_docs:
|
||||
out_file.write(line)
|
||||
if in_docs:
|
||||
if stop in line:
|
||||
in_docs = False
|
||||
os.rename(doc_file, doc_file+"~")
|
||||
os.rename(new_file, doc_file)
|
||||
if not altered:
|
||||
raise ValueError("Didn't find '%s' markers for in %s" % (start, doc_file))
|
||||
|
||||
if __name__ == "__main__":
|
||||
failed, success = 0, 0
|
||||
for backend in find_backends():
|
||||
try:
|
||||
alter_doc(backend)
|
||||
except Exception, e:
|
||||
print "Failed adding docs for %s backend: %s" % (backend, e)
|
||||
failed += 1
|
||||
else:
|
||||
success += 1
|
||||
print "Added docs for %d backends with %d failures" % (success, failed)
|
@ -128,5 +128,19 @@ Copy another local directory to the alias directory called source
|
||||
|
||||
rclone copy /home/source remote:source
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/alias/alias.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to alias (Alias for a existing remote).
|
||||
|
||||
#### --alias-remote
|
||||
|
||||
Remote or path to alias.
|
||||
Can be "myremote:path/to/dir", "myremote:bucket", "myremote:" or "/local/path".
|
||||
|
||||
- Config: remote
|
||||
- Env Var: RCLONE_ALIAS_REMOTE
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
@ -173,8 +173,110 @@ Let's say you usually use `amazon.co.uk`. When you authenticate with
|
||||
rclone it will take you to an `amazon.com` page to log in. Your
|
||||
`amazon.co.uk` email and password should work here just fine.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/amazonclouddrive/amazonclouddrive.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to amazon cloud drive (Amazon Drive).
|
||||
|
||||
#### --acd-client-id
|
||||
|
||||
Amazon Application Client ID.
|
||||
|
||||
- Config: client_id
|
||||
- Env Var: RCLONE_ACD_CLIENT_ID
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --acd-client-secret
|
||||
|
||||
Amazon Application Client Secret.
|
||||
|
||||
- Config: client_secret
|
||||
- Env Var: RCLONE_ACD_CLIENT_SECRET
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to amazon cloud drive (Amazon Drive).
|
||||
|
||||
#### --acd-auth-url
|
||||
|
||||
Auth server URL.
|
||||
Leave blank to use Amazon's.
|
||||
|
||||
- Config: auth_url
|
||||
- Env Var: RCLONE_ACD_AUTH_URL
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --acd-token-url
|
||||
|
||||
Token server url.
|
||||
leave blank to use Amazon's.
|
||||
|
||||
- Config: token_url
|
||||
- Env Var: RCLONE_ACD_TOKEN_URL
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --acd-checkpoint
|
||||
|
||||
Checkpoint for internal polling (debug).
|
||||
|
||||
- Config: checkpoint
|
||||
- Env Var: RCLONE_ACD_CHECKPOINT
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --acd-upload-wait-per-gb
|
||||
|
||||
Additional time per GB to wait after a failed complete upload to see if it appears.
|
||||
|
||||
Sometimes Amazon Drive gives an error when a file has been fully
|
||||
uploaded but the file appears anyway after a little while. This
|
||||
happens sometimes for files over 1GB in size and nearly every time for
|
||||
files bigger than 10GB. This parameter controls the time rclone waits
|
||||
for the file to appear.
|
||||
|
||||
The default value for this parameter is 3 minutes per GB, so by
|
||||
default it will wait 3 minutes for every GB uploaded to see if the
|
||||
file appears.
|
||||
|
||||
You can disable this feature by setting it to 0. This may cause
|
||||
conflict errors as rclone retries the failed upload but the file will
|
||||
most likely appear correctly eventually.
|
||||
|
||||
These values were determined empirically by observing lots of uploads
|
||||
of big files for a range of file sizes.
|
||||
|
||||
Upload with the "-v" flag to see more info about what rclone is doing
|
||||
in this situation.
|
||||
|
||||
- Config: upload_wait_per_gb
|
||||
- Env Var: RCLONE_ACD_UPLOAD_WAIT_PER_GB
|
||||
- Type: Duration
|
||||
- Default: 3m0s
|
||||
|
||||
#### --acd-templink-threshold
|
||||
|
||||
Files >= this size will be downloaded via their tempLink.
|
||||
|
||||
Files this size or more will be downloaded via their "tempLink". This
|
||||
is to work around a problem with Amazon Drive which blocks downloads
|
||||
of files bigger than about 10GB. The default for this is 9GB which
|
||||
shouldn't need to be changed.
|
||||
|
||||
To download files above this threshold, rclone requests a "tempLink"
|
||||
which downloads the file through a temporary URL directly from the
|
||||
underlying S3 storage.
|
||||
|
||||
- Config: templink_threshold
|
||||
- Env Var: RCLONE_ACD_TEMPLINK_THRESHOLD
|
||||
- Type: SizeSuffix
|
||||
- Default: 9G
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
### Limitations ###
|
||||
|
||||
|
@ -168,8 +168,112 @@ upload which means that there is a limit of 9.5TB of multipart uploads
|
||||
in progress as Azure won't allow more than that amount of uncommitted
|
||||
blocks.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/azureblob/azureblob.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to azureblob (Microsoft Azure Blob Storage).
|
||||
|
||||
#### --azureblob-account
|
||||
|
||||
Storage Account Name (leave blank to use connection string or SAS URL)
|
||||
|
||||
- Config: account
|
||||
- Env Var: RCLONE_AZUREBLOB_ACCOUNT
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --azureblob-key
|
||||
|
||||
Storage Account Key (leave blank to use connection string or SAS URL)
|
||||
|
||||
- Config: key
|
||||
- Env Var: RCLONE_AZUREBLOB_KEY
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --azureblob-sas-url
|
||||
|
||||
SAS URL for container level access only
|
||||
(leave blank if using account/key or connection string)
|
||||
|
||||
- Config: sas_url
|
||||
- Env Var: RCLONE_AZUREBLOB_SAS_URL
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to azureblob (Microsoft Azure Blob Storage).
|
||||
|
||||
#### --azureblob-endpoint
|
||||
|
||||
Endpoint for the service
|
||||
Leave blank normally.
|
||||
|
||||
- Config: endpoint
|
||||
- Env Var: RCLONE_AZUREBLOB_ENDPOINT
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --azureblob-upload-cutoff
|
||||
|
||||
Cutoff for switching to chunked upload (<= 256MB).
|
||||
|
||||
- Config: upload_cutoff
|
||||
- Env Var: RCLONE_AZUREBLOB_UPLOAD_CUTOFF
|
||||
- Type: SizeSuffix
|
||||
- Default: 256M
|
||||
|
||||
#### --azureblob-chunk-size
|
||||
|
||||
Upload chunk size (<= 100MB).
|
||||
|
||||
Note that this is stored in memory and there may be up to
|
||||
"--transfers" chunks stored at once in memory.
|
||||
|
||||
- Config: chunk_size
|
||||
- Env Var: RCLONE_AZUREBLOB_CHUNK_SIZE
|
||||
- Type: SizeSuffix
|
||||
- Default: 4M
|
||||
|
||||
#### --azureblob-list-chunk
|
||||
|
||||
Size of blob list.
|
||||
|
||||
This sets the number of blobs requested in each listing chunk. Default
|
||||
is the maximum, 5000. "List blobs" requests are permitted 2 minutes
|
||||
per megabyte to complete. If an operation is taking longer than 2
|
||||
minutes per megabyte on average, it will time out (
|
||||
[source](https://docs.microsoft.com/en-us/rest/api/storageservices/setting-timeouts-for-blob-service-operations#exceptions-to-default-timeout-interval)
|
||||
). This can be used to limit the number of blobs items to return, to
|
||||
avoid the time out.
|
||||
|
||||
- Config: list_chunk
|
||||
- Env Var: RCLONE_AZUREBLOB_LIST_CHUNK
|
||||
- Type: int
|
||||
- Default: 5000
|
||||
|
||||
#### --azureblob-access-tier
|
||||
|
||||
Access tier of blob: hot, cool or archive.
|
||||
|
||||
Archived blobs can be restored by setting access tier to hot or
|
||||
cool. Leave blank if you intend to use default access tier, which is
|
||||
set at account level
|
||||
|
||||
If there is no "access tier" specified, rclone doesn't apply any tier.
|
||||
rclone performs "Set Tier" operation on blobs while uploading, if objects
|
||||
are not modified, specifying "access tier" to new one will have no effect.
|
||||
If blobs are in "archive tier" at remote, trying to perform data transfer
|
||||
operations from remote will not be allowed. User should first restore by
|
||||
tiering blob to "Hot" or "Cool".
|
||||
|
||||
- Config: access_tier
|
||||
- Env Var: RCLONE_AZUREBLOB_ACCESS_TIER
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
### Limitations ###
|
||||
|
||||
|
@ -294,6 +294,108 @@ server to the nearest millisecond appended to them.
|
||||
Note that when using `--b2-versions` no file write operations are
|
||||
permitted, so you can't upload files or delete them.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/b2/b2.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to b2 (Backblaze B2).
|
||||
|
||||
#### --b2-account
|
||||
|
||||
Account ID or Application Key ID
|
||||
|
||||
- Config: account
|
||||
- Env Var: RCLONE_B2_ACCOUNT
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --b2-key
|
||||
|
||||
Application Key
|
||||
|
||||
- Config: key
|
||||
- Env Var: RCLONE_B2_KEY
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --b2-hard-delete
|
||||
|
||||
Permanently delete files on remote removal, otherwise hide files.
|
||||
|
||||
- Config: hard_delete
|
||||
- Env Var: RCLONE_B2_HARD_DELETE
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to b2 (Backblaze B2).
|
||||
|
||||
#### --b2-endpoint
|
||||
|
||||
Endpoint for the service.
|
||||
Leave blank normally.
|
||||
|
||||
- Config: endpoint
|
||||
- Env Var: RCLONE_B2_ENDPOINT
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --b2-test-mode
|
||||
|
||||
A flag string for X-Bz-Test-Mode header for debugging.
|
||||
|
||||
This is for debugging purposes only. Setting it to one of the strings
|
||||
below will cause b2 to return specific errors:
|
||||
|
||||
* "fail_some_uploads"
|
||||
* "expire_some_account_authorization_tokens"
|
||||
* "force_cap_exceeded"
|
||||
|
||||
These will be set in the "X-Bz-Test-Mode" header which is documented
|
||||
in the [b2 integrations checklist](https://www.backblaze.com/b2/docs/integration_checklist.html).
|
||||
|
||||
- Config: test_mode
|
||||
- Env Var: RCLONE_B2_TEST_MODE
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --b2-versions
|
||||
|
||||
Include old versions in directory listings.
|
||||
Note that when using this no file write operations are permitted,
|
||||
so you can't upload files or delete them.
|
||||
|
||||
- Config: versions
|
||||
- Env Var: RCLONE_B2_VERSIONS
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --b2-upload-cutoff
|
||||
|
||||
Cutoff for switching to chunked upload.
|
||||
|
||||
Files above this size will be uploaded in chunks of "--b2-chunk-size".
|
||||
|
||||
This value should be set no larger than 4.657GiB (== 5GB).
|
||||
|
||||
- Config: upload_cutoff
|
||||
- Env Var: RCLONE_B2_UPLOAD_CUTOFF
|
||||
- Type: SizeSuffix
|
||||
- Default: 190.735M
|
||||
|
||||
#### --b2-chunk-size
|
||||
|
||||
Upload chunk size. Must fit in memory.
|
||||
|
||||
When uploading large files, chunk the file into this size. Note that
|
||||
these chunks are buffered in memory and there might a maximum of
|
||||
"--transfers" chunks in progress at once. 5,000,000 Bytes is the
|
||||
minimim size.
|
||||
|
||||
- Config: chunk_size
|
||||
- Env Var: RCLONE_B2_CHUNK_SIZE
|
||||
- Type: SizeSuffix
|
||||
- Default: 96M
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
|
@ -217,8 +217,54 @@ normally 8MB so increasing `--transfers` will increase memory use.
|
||||
Depending on the enterprise settings for your user, the item will
|
||||
either be actually deleted from Box or moved to the trash.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/box/box.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to box (Box).
|
||||
|
||||
#### --box-client-id
|
||||
|
||||
Box App Client Id.
|
||||
Leave blank normally.
|
||||
|
||||
- Config: client_id
|
||||
- Env Var: RCLONE_BOX_CLIENT_ID
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --box-client-secret
|
||||
|
||||
Box App Client Secret
|
||||
Leave blank normally.
|
||||
|
||||
- Config: client_secret
|
||||
- Env Var: RCLONE_BOX_CLIENT_SECRET
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to box (Box).
|
||||
|
||||
#### --box-upload-cutoff
|
||||
|
||||
Cutoff for switching to multipart upload (>= 50MB).
|
||||
|
||||
- Config: upload_cutoff
|
||||
- Env Var: RCLONE_BOX_UPLOAD_CUTOFF
|
||||
- Type: SizeSuffix
|
||||
- Default: 50M
|
||||
|
||||
#### --box-commit-retries
|
||||
|
||||
Max number of times to try committing a multipart file.
|
||||
|
||||
- Config: commit_retries
|
||||
- Env Var: RCLONE_BOX_COMMIT_RETRIES
|
||||
- Type: int
|
||||
- Default: 100
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
### Limitations ###
|
||||
|
||||
|
@ -290,5 +290,315 @@ Params:
|
||||
- **remote** = path to remote **(required)**
|
||||
- **withData** = true/false to delete cached data (chunks) as well _(optional, false by default)_
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/cache/cache.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to cache (Cache a remote).
|
||||
|
||||
#### --cache-remote
|
||||
|
||||
Remote to cache.
|
||||
Normally should contain a ':' and a path, eg "myremote:path/to/dir",
|
||||
"myremote:bucket" or maybe "myremote:" (not recommended).
|
||||
|
||||
- Config: remote
|
||||
- Env Var: RCLONE_CACHE_REMOTE
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --cache-plex-url
|
||||
|
||||
The URL of the Plex server
|
||||
|
||||
- Config: plex_url
|
||||
- Env Var: RCLONE_CACHE_PLEX_URL
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --cache-plex-username
|
||||
|
||||
The username of the Plex user
|
||||
|
||||
- Config: plex_username
|
||||
- Env Var: RCLONE_CACHE_PLEX_USERNAME
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --cache-plex-password
|
||||
|
||||
The password of the Plex user
|
||||
|
||||
- Config: plex_password
|
||||
- Env Var: RCLONE_CACHE_PLEX_PASSWORD
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --cache-chunk-size
|
||||
|
||||
The size of a chunk (partial file data).
|
||||
|
||||
Use lower numbers for slower connections. If the chunk size is
|
||||
changed, any downloaded chunks will be invalid and cache-chunk-path
|
||||
will need to be cleared or unexpected EOF errors will occur.
|
||||
|
||||
- Config: chunk_size
|
||||
- Env Var: RCLONE_CACHE_CHUNK_SIZE
|
||||
- Type: SizeSuffix
|
||||
- Default: 5M
|
||||
- Examples:
|
||||
- "1m"
|
||||
- 1MB
|
||||
- "5M"
|
||||
- 5 MB
|
||||
- "10M"
|
||||
- 10 MB
|
||||
|
||||
#### --cache-info-age
|
||||
|
||||
How long to cache file structure information (directory listings, file size, times etc).
|
||||
If all write operations are done through the cache then you can safely make
|
||||
this value very large as the cache store will also be updated in real time.
|
||||
|
||||
- Config: info_age
|
||||
- Env Var: RCLONE_CACHE_INFO_AGE
|
||||
- Type: Duration
|
||||
- Default: 6h0m0s
|
||||
- Examples:
|
||||
- "1h"
|
||||
- 1 hour
|
||||
- "24h"
|
||||
- 24 hours
|
||||
- "48h"
|
||||
- 48 hours
|
||||
|
||||
#### --cache-chunk-total-size
|
||||
|
||||
The total size that the chunks can take up on the local disk.
|
||||
|
||||
If the cache exceeds this value then it will start to delete the
|
||||
oldest chunks until it goes under this value.
|
||||
|
||||
- Config: chunk_total_size
|
||||
- Env Var: RCLONE_CACHE_CHUNK_TOTAL_SIZE
|
||||
- Type: SizeSuffix
|
||||
- Default: 10G
|
||||
- Examples:
|
||||
- "500M"
|
||||
- 500 MB
|
||||
- "1G"
|
||||
- 1 GB
|
||||
- "10G"
|
||||
- 10 GB
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to cache (Cache a remote).
|
||||
|
||||
#### --cache-plex-token
|
||||
|
||||
The plex token for authentication - auto set normally
|
||||
|
||||
- Config: plex_token
|
||||
- Env Var: RCLONE_CACHE_PLEX_TOKEN
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --cache-plex-insecure
|
||||
|
||||
Skip all certificate verifications when connecting to the Plex server
|
||||
|
||||
- Config: plex_insecure
|
||||
- Env Var: RCLONE_CACHE_PLEX_INSECURE
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --cache-db-path
|
||||
|
||||
Directory to store file structure metadata DB.
|
||||
The remote name is used as the DB file name.
|
||||
|
||||
- Config: db_path
|
||||
- Env Var: RCLONE_CACHE_DB_PATH
|
||||
- Type: string
|
||||
- Default: "/home/ncw/.cache/rclone/cache-backend"
|
||||
|
||||
#### --cache-chunk-path
|
||||
|
||||
Directory to cache chunk files.
|
||||
|
||||
Path to where partial file data (chunks) are stored locally. The remote
|
||||
name is appended to the final path.
|
||||
|
||||
This config follows the "--cache-db-path". If you specify a custom
|
||||
location for "--cache-db-path" and don't specify one for "--cache-chunk-path"
|
||||
then "--cache-chunk-path" will use the same path as "--cache-db-path".
|
||||
|
||||
- Config: chunk_path
|
||||
- Env Var: RCLONE_CACHE_CHUNK_PATH
|
||||
- Type: string
|
||||
- Default: "/home/ncw/.cache/rclone/cache-backend"
|
||||
|
||||
#### --cache-db-purge
|
||||
|
||||
Clear all the cached data for this remote on start.
|
||||
|
||||
- Config: db_purge
|
||||
- Env Var: RCLONE_CACHE_DB_PURGE
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --cache-chunk-clean-interval
|
||||
|
||||
How often should the cache perform cleanups of the chunk storage.
|
||||
The default value should be ok for most people. If you find that the
|
||||
cache goes over "cache-chunk-total-size" too often then try to lower
|
||||
this value to force it to perform cleanups more often.
|
||||
|
||||
- Config: chunk_clean_interval
|
||||
- Env Var: RCLONE_CACHE_CHUNK_CLEAN_INTERVAL
|
||||
- Type: Duration
|
||||
- Default: 1m0s
|
||||
|
||||
#### --cache-read-retries
|
||||
|
||||
How many times to retry a read from a cache storage.
|
||||
|
||||
Since reading from a cache stream is independent from downloading file
|
||||
data, readers can get to a point where there's no more data in the
|
||||
cache. Most of the times this can indicate a connectivity issue if
|
||||
cache isn't able to provide file data anymore.
|
||||
|
||||
For really slow connections, increase this to a point where the stream is
|
||||
able to provide data but your experience will be very stuttering.
|
||||
|
||||
- Config: read_retries
|
||||
- Env Var: RCLONE_CACHE_READ_RETRIES
|
||||
- Type: int
|
||||
- Default: 10
|
||||
|
||||
#### --cache-workers
|
||||
|
||||
How many workers should run in parallel to download chunks.
|
||||
|
||||
Higher values will mean more parallel processing (better CPU needed)
|
||||
and more concurrent requests on the cloud provider. This impacts
|
||||
several aspects like the cloud provider API limits, more stress on the
|
||||
hardware that rclone runs on but it also means that streams will be
|
||||
more fluid and data will be available much more faster to readers.
|
||||
|
||||
**Note**: If the optional Plex integration is enabled then this
|
||||
setting will adapt to the type of reading performed and the value
|
||||
specified here will be used as a maximum number of workers to use.
|
||||
|
||||
- Config: workers
|
||||
- Env Var: RCLONE_CACHE_WORKERS
|
||||
- Type: int
|
||||
- Default: 4
|
||||
|
||||
#### --cache-chunk-no-memory
|
||||
|
||||
Disable the in-memory cache for storing chunks during streaming.
|
||||
|
||||
By default, cache will keep file data during streaming in RAM as well
|
||||
to provide it to readers as fast as possible.
|
||||
|
||||
This transient data is evicted as soon as it is read and the number of
|
||||
chunks stored doesn't exceed the number of workers. However, depending
|
||||
on other settings like "cache-chunk-size" and "cache-workers" this footprint
|
||||
can increase if there are parallel streams too (multiple files being read
|
||||
at the same time).
|
||||
|
||||
If the hardware permits it, use this feature to provide an overall better
|
||||
performance during streaming but it can also be disabled if RAM is not
|
||||
available on the local machine.
|
||||
|
||||
- Config: chunk_no_memory
|
||||
- Env Var: RCLONE_CACHE_CHUNK_NO_MEMORY
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --cache-rps
|
||||
|
||||
Limits the number of requests per second to the source FS (-1 to disable)
|
||||
|
||||
This setting places a hard limit on the number of requests per second
|
||||
that cache will be doing to the cloud provider remote and try to
|
||||
respect that value by setting waits between reads.
|
||||
|
||||
If you find that you're getting banned or limited on the cloud
|
||||
provider through cache and know that a smaller number of requests per
|
||||
second will allow you to work with it then you can use this setting
|
||||
for that.
|
||||
|
||||
A good balance of all the other settings should make this setting
|
||||
useless but it is available to set for more special cases.
|
||||
|
||||
**NOTE**: This will limit the number of requests during streams but
|
||||
other API calls to the cloud provider like directory listings will
|
||||
still pass.
|
||||
|
||||
- Config: rps
|
||||
- Env Var: RCLONE_CACHE_RPS
|
||||
- Type: int
|
||||
- Default: -1
|
||||
|
||||
#### --cache-writes
|
||||
|
||||
Cache file data on writes through the FS
|
||||
|
||||
If you need to read files immediately after you upload them through
|
||||
cache you can enable this flag to have their data stored in the
|
||||
cache store at the same time during upload.
|
||||
|
||||
- Config: writes
|
||||
- Env Var: RCLONE_CACHE_WRITES
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --cache-tmp-upload-path
|
||||
|
||||
Directory to keep temporary files until they are uploaded.
|
||||
|
||||
This is the path where cache will use as a temporary storage for new
|
||||
files that need to be uploaded to the cloud provider.
|
||||
|
||||
Specifying a value will enable this feature. Without it, it is
|
||||
completely disabled and files will be uploaded directly to the cloud
|
||||
provider
|
||||
|
||||
- Config: tmp_upload_path
|
||||
- Env Var: RCLONE_CACHE_TMP_UPLOAD_PATH
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --cache-tmp-wait-time
|
||||
|
||||
How long should files be stored in local cache before being uploaded
|
||||
|
||||
This is the duration that a file must wait in the temporary location
|
||||
_cache-tmp-upload-path_ before it is selected for upload.
|
||||
|
||||
Note that only one file is uploaded at a time and it can take longer
|
||||
to start the upload if a queue formed for this purpose.
|
||||
|
||||
- Config: tmp_wait_time
|
||||
- Env Var: RCLONE_CACHE_TMP_WAIT_TIME
|
||||
- Type: Duration
|
||||
- Default: 15s
|
||||
|
||||
#### --cache-db-wait-time
|
||||
|
||||
How long to wait for the DB to be available - 0 is unlimited
|
||||
|
||||
Only one process can have the DB open at any one time, so rclone waits
|
||||
for this duration for the DB to become available before it gives an
|
||||
error.
|
||||
|
||||
If you set it to 0 then it will wait forever.
|
||||
|
||||
- Config: db_wait_time
|
||||
- Env Var: RCLONE_CACHE_DB_WAIT_TIME
|
||||
- Type: Duration
|
||||
- Default: 1s
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
@ -294,8 +294,93 @@ Note that you should use the `rclone cryptcheck` command to check the
|
||||
integrity of a crypted remote instead of `rclone check` which can't
|
||||
check the checksums properly.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/crypt/crypt.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to crypt (Encrypt/Decrypt a remote).
|
||||
|
||||
#### --crypt-remote
|
||||
|
||||
Remote to encrypt/decrypt.
|
||||
Normally should contain a ':' and a path, eg "myremote:path/to/dir",
|
||||
"myremote:bucket" or maybe "myremote:" (not recommended).
|
||||
|
||||
- Config: remote
|
||||
- Env Var: RCLONE_CRYPT_REMOTE
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --crypt-filename-encryption
|
||||
|
||||
How to encrypt the filenames.
|
||||
|
||||
- Config: filename_encryption
|
||||
- Env Var: RCLONE_CRYPT_FILENAME_ENCRYPTION
|
||||
- Type: string
|
||||
- Default: "standard"
|
||||
- Examples:
|
||||
- "off"
|
||||
- Don't encrypt the file names. Adds a ".bin" extension only.
|
||||
- "standard"
|
||||
- Encrypt the filenames see the docs for the details.
|
||||
- "obfuscate"
|
||||
- Very simple filename obfuscation.
|
||||
|
||||
#### --crypt-directory-name-encryption
|
||||
|
||||
Option to either encrypt directory names or leave them intact.
|
||||
|
||||
- Config: directory_name_encryption
|
||||
- Env Var: RCLONE_CRYPT_DIRECTORY_NAME_ENCRYPTION
|
||||
- Type: bool
|
||||
- Default: true
|
||||
- Examples:
|
||||
- "true"
|
||||
- Encrypt directory names.
|
||||
- "false"
|
||||
- Don't encrypt directory names, leave them intact.
|
||||
|
||||
#### --crypt-password
|
||||
|
||||
Password or pass phrase for encryption.
|
||||
|
||||
- Config: password
|
||||
- Env Var: RCLONE_CRYPT_PASSWORD
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --crypt-password2
|
||||
|
||||
Password or pass phrase for salt. Optional but recommended.
|
||||
Should be different to the previous password.
|
||||
|
||||
- Config: password2
|
||||
- Env Var: RCLONE_CRYPT_PASSWORD2
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to crypt (Encrypt/Decrypt a remote).
|
||||
|
||||
#### --crypt-show-mapping
|
||||
|
||||
For all files listed show how the names encrypt.
|
||||
|
||||
If this flag is set then for each file that the remote is asked to
|
||||
list, it will log (at level INFO) a line stating the decrypted file
|
||||
name and the encrypted file name.
|
||||
|
||||
This is so you can work out which encrypted names are which decrypted
|
||||
names just in case you need to do something with the encrypted file
|
||||
names, or for debugging purposes.
|
||||
|
||||
- Config: show_mapping
|
||||
- Env Var: RCLONE_CRYPT_SHOW_MAPPING
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
## Backing up a crypted remote ##
|
||||
|
||||
|
@ -483,8 +483,311 @@ Google Documents.
|
||||
| url | INI style link file | macOS, Windows |
|
||||
| webloc | macOS specific XML format | macOS |
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/drive/drive.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to drive (Google Drive).
|
||||
|
||||
#### --drive-client-id
|
||||
|
||||
Google Application Client Id
|
||||
Leave blank normally.
|
||||
|
||||
- Config: client_id
|
||||
- Env Var: RCLONE_DRIVE_CLIENT_ID
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --drive-client-secret
|
||||
|
||||
Google Application Client Secret
|
||||
Leave blank normally.
|
||||
|
||||
- Config: client_secret
|
||||
- Env Var: RCLONE_DRIVE_CLIENT_SECRET
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --drive-scope
|
||||
|
||||
Scope that rclone should use when requesting access from drive.
|
||||
|
||||
- Config: scope
|
||||
- Env Var: RCLONE_DRIVE_SCOPE
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "drive"
|
||||
- Full access all files, excluding Application Data Folder.
|
||||
- "drive.readonly"
|
||||
- Read-only access to file metadata and file contents.
|
||||
- "drive.file"
|
||||
- Access to files created by rclone only.
|
||||
- These are visible in the drive website.
|
||||
- File authorization is revoked when the user deauthorizes the app.
|
||||
- "drive.appfolder"
|
||||
- Allows read and write access to the Application Data folder.
|
||||
- This is not visible in the drive website.
|
||||
- "drive.metadata.readonly"
|
||||
- Allows read-only access to file metadata but
|
||||
- does not allow any access to read or download file content.
|
||||
|
||||
#### --drive-root-folder-id
|
||||
|
||||
ID of the root folder
|
||||
Leave blank normally.
|
||||
Fill in to access "Computers" folders. (see docs).
|
||||
|
||||
- Config: root_folder_id
|
||||
- Env Var: RCLONE_DRIVE_ROOT_FOLDER_ID
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --drive-service-account-file
|
||||
|
||||
Service Account Credentials JSON file path
|
||||
Leave blank normally.
|
||||
Needed only if you want use SA instead of interactive login.
|
||||
|
||||
- Config: service_account_file
|
||||
- Env Var: RCLONE_DRIVE_SERVICE_ACCOUNT_FILE
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to drive (Google Drive).
|
||||
|
||||
#### --drive-service-account-credentials
|
||||
|
||||
Service Account Credentials JSON blob
|
||||
Leave blank normally.
|
||||
Needed only if you want use SA instead of interactive login.
|
||||
|
||||
- Config: service_account_credentials
|
||||
- Env Var: RCLONE_DRIVE_SERVICE_ACCOUNT_CREDENTIALS
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --drive-team-drive
|
||||
|
||||
ID of the Team Drive
|
||||
|
||||
- Config: team_drive
|
||||
- Env Var: RCLONE_DRIVE_TEAM_DRIVE
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --drive-auth-owner-only
|
||||
|
||||
Only consider files owned by the authenticated user.
|
||||
|
||||
- Config: auth_owner_only
|
||||
- Env Var: RCLONE_DRIVE_AUTH_OWNER_ONLY
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --drive-use-trash
|
||||
|
||||
Send files to the trash instead of deleting permanently.
|
||||
Defaults to true, namely sending files to the trash.
|
||||
Use `--drive-use-trash=false` to delete files permanently instead.
|
||||
|
||||
- Config: use_trash
|
||||
- Env Var: RCLONE_DRIVE_USE_TRASH
|
||||
- Type: bool
|
||||
- Default: true
|
||||
|
||||
#### --drive-skip-gdocs
|
||||
|
||||
Skip google documents in all listings.
|
||||
If given, gdocs practically become invisible to rclone.
|
||||
|
||||
- Config: skip_gdocs
|
||||
- Env Var: RCLONE_DRIVE_SKIP_GDOCS
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --drive-shared-with-me
|
||||
|
||||
Only show files that are shared with me.
|
||||
|
||||
Instructs rclone to operate on your "Shared with me" folder (where
|
||||
Google Drive lets you access the files and folders others have shared
|
||||
with you).
|
||||
|
||||
This works both with the "list" (lsd, lsl, etc) and the "copy"
|
||||
commands (copy, sync, etc), and with all other commands too.
|
||||
|
||||
- Config: shared_with_me
|
||||
- Env Var: RCLONE_DRIVE_SHARED_WITH_ME
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --drive-trashed-only
|
||||
|
||||
Only show files that are in the trash.
|
||||
This will show trashed files in their original directory structure.
|
||||
|
||||
- Config: trashed_only
|
||||
- Env Var: RCLONE_DRIVE_TRASHED_ONLY
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --drive-formats
|
||||
|
||||
Deprecated: see export_formats
|
||||
|
||||
- Config: formats
|
||||
- Env Var: RCLONE_DRIVE_FORMATS
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --drive-export-formats
|
||||
|
||||
Comma separated list of preferred formats for downloading Google docs.
|
||||
|
||||
- Config: export_formats
|
||||
- Env Var: RCLONE_DRIVE_EXPORT_FORMATS
|
||||
- Type: string
|
||||
- Default: "docx,xlsx,pptx,svg"
|
||||
|
||||
#### --drive-import-formats
|
||||
|
||||
Comma separated list of preferred formats for uploading Google docs.
|
||||
|
||||
- Config: import_formats
|
||||
- Env Var: RCLONE_DRIVE_IMPORT_FORMATS
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --drive-allow-import-name-change
|
||||
|
||||
Allow the filetype to change when uploading Google docs (e.g. file.doc to file.docx). This will confuse sync and reupload every time.
|
||||
|
||||
- Config: allow_import_name_change
|
||||
- Env Var: RCLONE_DRIVE_ALLOW_IMPORT_NAME_CHANGE
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --drive-use-created-date
|
||||
|
||||
Use file created date instead of modified date.,
|
||||
|
||||
Useful when downloading data and you want the creation date used in
|
||||
place of the last modified date.
|
||||
|
||||
**WARNING**: This flag may have some unexpected consequences.
|
||||
|
||||
When uploading to your drive all files will be overwritten unless they
|
||||
haven't been modified since their creation. And the inverse will occur
|
||||
while downloading. This side effect can be avoided by using the
|
||||
"--checksum" flag.
|
||||
|
||||
This feature was implemented to retain photos capture date as recorded
|
||||
by google photos. You will first need to check the "Create a Google
|
||||
Photos folder" option in your google drive settings. You can then copy
|
||||
or move the photos locally and use the date the image was taken
|
||||
(created) set as the modification date.
|
||||
|
||||
- Config: use_created_date
|
||||
- Env Var: RCLONE_DRIVE_USE_CREATED_DATE
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --drive-list-chunk
|
||||
|
||||
Size of listing chunk 100-1000. 0 to disable.
|
||||
|
||||
- Config: list_chunk
|
||||
- Env Var: RCLONE_DRIVE_LIST_CHUNK
|
||||
- Type: int
|
||||
- Default: 1000
|
||||
|
||||
#### --drive-impersonate
|
||||
|
||||
Impersonate this user when using a service account.
|
||||
|
||||
- Config: impersonate
|
||||
- Env Var: RCLONE_DRIVE_IMPERSONATE
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --drive-alternate-export
|
||||
|
||||
Use alternate export URLs for google documents export.,
|
||||
|
||||
If this option is set this instructs rclone to use an alternate set of
|
||||
export URLs for drive documents. Users have reported that the
|
||||
official export URLs can't export large documents, whereas these
|
||||
unofficial ones can.
|
||||
|
||||
See rclone issue [#2243](https://github.com/ncw/rclone/issues/2243) for background,
|
||||
[this google drive issue](https://issuetracker.google.com/issues/36761333) and
|
||||
[this helpful post](https://www.labnol.org/internet/direct-links-for-google-drive/28356/).
|
||||
|
||||
- Config: alternate_export
|
||||
- Env Var: RCLONE_DRIVE_ALTERNATE_EXPORT
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --drive-upload-cutoff
|
||||
|
||||
Cutoff for switching to chunked upload
|
||||
|
||||
- Config: upload_cutoff
|
||||
- Env Var: RCLONE_DRIVE_UPLOAD_CUTOFF
|
||||
- Type: SizeSuffix
|
||||
- Default: 8M
|
||||
|
||||
#### --drive-chunk-size
|
||||
|
||||
Upload chunk size. Must a power of 2 >= 256k.
|
||||
|
||||
Making this larger will improve performance, but note that each chunk
|
||||
is buffered in memory one per transfer.
|
||||
|
||||
Reducing this will reduce memory usage but decrease performance.
|
||||
|
||||
- Config: chunk_size
|
||||
- Env Var: RCLONE_DRIVE_CHUNK_SIZE
|
||||
- Type: SizeSuffix
|
||||
- Default: 8M
|
||||
|
||||
#### --drive-acknowledge-abuse
|
||||
|
||||
Set to allow files which return cannotDownloadAbusiveFile to be downloaded.
|
||||
|
||||
If downloading a file returns the error "This file has been identified
|
||||
as malware or spam and cannot be downloaded" with the error code
|
||||
"cannotDownloadAbusiveFile" then supply this flag to rclone to
|
||||
indicate you acknowledge the risks of downloading the file and rclone
|
||||
will download it anyway.
|
||||
|
||||
- Config: acknowledge_abuse
|
||||
- Env Var: RCLONE_DRIVE_ACKNOWLEDGE_ABUSE
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --drive-keep-revision-forever
|
||||
|
||||
Keep new head revision of each file forever.
|
||||
|
||||
- Config: keep_revision_forever
|
||||
- Env Var: RCLONE_DRIVE_KEEP_REVISION_FOREVER
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --drive-v2-download-min-size
|
||||
|
||||
If Object's are greater, use drive v2 API to download.
|
||||
|
||||
- Config: v2_download_min_size
|
||||
- Env Var: RCLONE_DRIVE_V2_DOWNLOAD_MIN_SIZE
|
||||
- Type: SizeSuffix
|
||||
- Default: off
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
### Limitations ###
|
||||
|
||||
|
@ -123,8 +123,52 @@ Dropbox supports [its own hash
|
||||
type](https://www.dropbox.com/developers/reference/content-hash) which
|
||||
is checked for all transfers.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/dropbox/dropbox.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to dropbox (Dropbox).
|
||||
|
||||
#### --dropbox-client-id
|
||||
|
||||
Dropbox App Client Id
|
||||
Leave blank normally.
|
||||
|
||||
- Config: client_id
|
||||
- Env Var: RCLONE_DROPBOX_CLIENT_ID
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --dropbox-client-secret
|
||||
|
||||
Dropbox App Client Secret
|
||||
Leave blank normally.
|
||||
|
||||
- Config: client_secret
|
||||
- Env Var: RCLONE_DROPBOX_CLIENT_SECRET
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to dropbox (Dropbox).
|
||||
|
||||
#### --dropbox-chunk-size
|
||||
|
||||
Upload chunk size. (< 150M).
|
||||
|
||||
Any files larger than this will be uploaded in chunks of this size.
|
||||
|
||||
Note that chunks are buffered in memory (one at a time) so rclone can
|
||||
deal with retries. Setting this larger will increase the speed
|
||||
slightly (at most 10% for 128MB in tests) at the cost of using more
|
||||
memory. It can be set smaller if you are tight on memory.
|
||||
|
||||
- Config: chunk_size
|
||||
- Env Var: RCLONE_DROPBOX_CHUNK_SIZE
|
||||
- Type: SizeSuffix
|
||||
- Default: 48M
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
### Limitations ###
|
||||
|
||||
|
@ -119,8 +119,51 @@ will be time of upload.
|
||||
|
||||
FTP does not support any checksums.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/ftp/ftp.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to ftp (FTP Connection).
|
||||
|
||||
#### --ftp-host
|
||||
|
||||
FTP host to connect to
|
||||
|
||||
- Config: host
|
||||
- Env Var: RCLONE_FTP_HOST
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "ftp.example.com"
|
||||
- Connect to ftp.example.com
|
||||
|
||||
#### --ftp-user
|
||||
|
||||
FTP username, leave blank for current username, ncw
|
||||
|
||||
- Config: user
|
||||
- Env Var: RCLONE_FTP_USER
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --ftp-port
|
||||
|
||||
FTP port, leave blank to use default (21)
|
||||
|
||||
- Config: port
|
||||
- Env Var: RCLONE_FTP_PORT
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --ftp-pass
|
||||
|
||||
FTP password
|
||||
|
||||
- Config: pass
|
||||
- Env Var: RCLONE_FTP_PASS
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
### Limitations ###
|
||||
|
||||
|
@ -229,5 +229,163 @@ Google google cloud storage stores md5sums natively and rclone stores
|
||||
modification times as metadata on the object, under the "mtime" key in
|
||||
RFC3339 format accurate to 1ns.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/googlecloudstorage/googlecloudstorage.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to google cloud storage (Google Cloud Storage (this is not Google Drive)).
|
||||
|
||||
#### --gcs-client-id
|
||||
|
||||
Google Application Client Id
|
||||
Leave blank normally.
|
||||
|
||||
- Config: client_id
|
||||
- Env Var: RCLONE_GCS_CLIENT_ID
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --gcs-client-secret
|
||||
|
||||
Google Application Client Secret
|
||||
Leave blank normally.
|
||||
|
||||
- Config: client_secret
|
||||
- Env Var: RCLONE_GCS_CLIENT_SECRET
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --gcs-project-number
|
||||
|
||||
Project number.
|
||||
Optional - needed only for list/create/delete buckets - see your developer console.
|
||||
|
||||
- Config: project_number
|
||||
- Env Var: RCLONE_GCS_PROJECT_NUMBER
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --gcs-service-account-file
|
||||
|
||||
Service Account Credentials JSON file path
|
||||
Leave blank normally.
|
||||
Needed only if you want use SA instead of interactive login.
|
||||
|
||||
- Config: service_account_file
|
||||
- Env Var: RCLONE_GCS_SERVICE_ACCOUNT_FILE
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --gcs-service-account-credentials
|
||||
|
||||
Service Account Credentials JSON blob
|
||||
Leave blank normally.
|
||||
Needed only if you want use SA instead of interactive login.
|
||||
|
||||
- Config: service_account_credentials
|
||||
- Env Var: RCLONE_GCS_SERVICE_ACCOUNT_CREDENTIALS
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --gcs-object-acl
|
||||
|
||||
Access Control List for new objects.
|
||||
|
||||
- Config: object_acl
|
||||
- Env Var: RCLONE_GCS_OBJECT_ACL
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "authenticatedRead"
|
||||
- Object owner gets OWNER access, and all Authenticated Users get READER access.
|
||||
- "bucketOwnerFullControl"
|
||||
- Object owner gets OWNER access, and project team owners get OWNER access.
|
||||
- "bucketOwnerRead"
|
||||
- Object owner gets OWNER access, and project team owners get READER access.
|
||||
- "private"
|
||||
- Object owner gets OWNER access [default if left blank].
|
||||
- "projectPrivate"
|
||||
- Object owner gets OWNER access, and project team members get access according to their roles.
|
||||
- "publicRead"
|
||||
- Object owner gets OWNER access, and all Users get READER access.
|
||||
|
||||
#### --gcs-bucket-acl
|
||||
|
||||
Access Control List for new buckets.
|
||||
|
||||
- Config: bucket_acl
|
||||
- Env Var: RCLONE_GCS_BUCKET_ACL
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "authenticatedRead"
|
||||
- Project team owners get OWNER access, and all Authenticated Users get READER access.
|
||||
- "private"
|
||||
- Project team owners get OWNER access [default if left blank].
|
||||
- "projectPrivate"
|
||||
- Project team members get access according to their roles.
|
||||
- "publicRead"
|
||||
- Project team owners get OWNER access, and all Users get READER access.
|
||||
- "publicReadWrite"
|
||||
- Project team owners get OWNER access, and all Users get WRITER access.
|
||||
|
||||
#### --gcs-location
|
||||
|
||||
Location for the newly created buckets.
|
||||
|
||||
- Config: location
|
||||
- Env Var: RCLONE_GCS_LOCATION
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- ""
|
||||
- Empty for default location (US).
|
||||
- "asia"
|
||||
- Multi-regional location for Asia.
|
||||
- "eu"
|
||||
- Multi-regional location for Europe.
|
||||
- "us"
|
||||
- Multi-regional location for United States.
|
||||
- "asia-east1"
|
||||
- Taiwan.
|
||||
- "asia-northeast1"
|
||||
- Tokyo.
|
||||
- "asia-southeast1"
|
||||
- Singapore.
|
||||
- "australia-southeast1"
|
||||
- Sydney.
|
||||
- "europe-west1"
|
||||
- Belgium.
|
||||
- "europe-west2"
|
||||
- London.
|
||||
- "us-central1"
|
||||
- Iowa.
|
||||
- "us-east1"
|
||||
- South Carolina.
|
||||
- "us-east4"
|
||||
- Northern Virginia.
|
||||
- "us-west1"
|
||||
- Oregon.
|
||||
|
||||
#### --gcs-storage-class
|
||||
|
||||
The storage class to use when storing objects in Google Cloud Storage.
|
||||
|
||||
- Config: storage_class
|
||||
- Env Var: RCLONE_GCS_STORAGE_CLASS
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- ""
|
||||
- Default
|
||||
- "MULTI_REGIONAL"
|
||||
- Multi-regional storage class
|
||||
- "REGIONAL"
|
||||
- Regional storage class
|
||||
- "NEARLINE"
|
||||
- Nearline storage class
|
||||
- "COLDLINE"
|
||||
- Coldline storage class
|
||||
- "DURABLE_REDUCED_AVAILABILITY"
|
||||
- Durable reduced availability storage class
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
@ -126,5 +126,21 @@ without a config file:
|
||||
|
||||
rclone lsd --http-url https://beta.rclone.org :http:
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/http/http.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to http (http Connection).
|
||||
|
||||
#### --http-url
|
||||
|
||||
URL of http host to connect to
|
||||
|
||||
- Config: url
|
||||
- Env Var: RCLONE_HTTP_URL
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "https://example.com"
|
||||
- Connect to example.com
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
@ -128,8 +128,48 @@ amongst others) for storing the modification time for an object.
|
||||
Note that Hubic wraps the Swift backend, so most of the properties of
|
||||
are the same.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/hubic/hubic.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to hubic (Hubic).
|
||||
|
||||
#### --hubic-client-id
|
||||
|
||||
Hubic Client Id
|
||||
Leave blank normally.
|
||||
|
||||
- Config: client_id
|
||||
- Env Var: RCLONE_HUBIC_CLIENT_ID
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --hubic-client-secret
|
||||
|
||||
Hubic Client Secret
|
||||
Leave blank normally.
|
||||
|
||||
- Config: client_secret
|
||||
- Env Var: RCLONE_HUBIC_CLIENT_SECRET
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to hubic (Hubic).
|
||||
|
||||
#### --hubic-chunk-size
|
||||
|
||||
Above this size files will be chunked into a _segments container.
|
||||
|
||||
Above this size files will be chunked into a _segments container. The
|
||||
default for this is 5GB which is its maximum value.
|
||||
|
||||
- Config: chunk_size
|
||||
- Env Var: RCLONE_HUBIC_CHUNK_SIZE
|
||||
- Type: SizeSuffix
|
||||
- Default: 5G
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
### Limitations ###
|
||||
|
||||
|
@ -124,8 +124,76 @@ To view your current quota you can use the `rclone about remote:`
|
||||
command which will display your usage limit (unless it is unlimited)
|
||||
and the current usage.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/jottacloud/jottacloud.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to jottacloud (JottaCloud).
|
||||
|
||||
#### --jottacloud-user
|
||||
|
||||
User Name
|
||||
|
||||
- Config: user
|
||||
- Env Var: RCLONE_JOTTACLOUD_USER
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --jottacloud-pass
|
||||
|
||||
Password.
|
||||
|
||||
- Config: pass
|
||||
- Env Var: RCLONE_JOTTACLOUD_PASS
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --jottacloud-mountpoint
|
||||
|
||||
The mountpoint to use.
|
||||
|
||||
- Config: mountpoint
|
||||
- Env Var: RCLONE_JOTTACLOUD_MOUNTPOINT
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "Sync"
|
||||
- Will be synced by the official client.
|
||||
- "Archive"
|
||||
- Archive
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to jottacloud (JottaCloud).
|
||||
|
||||
#### --jottacloud-md5-memory-limit
|
||||
|
||||
Files bigger than this will be cached on disk to calculate the MD5 if required.
|
||||
|
||||
- Config: md5_memory_limit
|
||||
- Env Var: RCLONE_JOTTACLOUD_MD5_MEMORY_LIMIT
|
||||
- Type: SizeSuffix
|
||||
- Default: 10M
|
||||
|
||||
#### --jottacloud-hard-delete
|
||||
|
||||
Delete files permanently rather than putting them into the trash.
|
||||
|
||||
- Config: hard_delete
|
||||
- Env Var: RCLONE_JOTTACLOUD_HARD_DELETE
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --jottacloud-unlink
|
||||
|
||||
Remove existing public link to file/folder with link command rather than creating.
|
||||
Default is false, meaning link command will create or retrieve public link.
|
||||
|
||||
- Config: unlink
|
||||
- Env Var: RCLONE_JOTTACLOUD_UNLINK
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
### Limitations ###
|
||||
|
||||
|
@ -159,5 +159,84 @@ filesystem.
|
||||
**NB** This flag is only available on Unix based systems. On systems
|
||||
where it isn't supported (eg Windows) it will be ignored.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/local/local.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to local (Local Disk).
|
||||
|
||||
#### --local-nounc
|
||||
|
||||
Disable UNC (long path names) conversion on Windows
|
||||
|
||||
- Config: nounc
|
||||
- Env Var: RCLONE_LOCAL_NOUNC
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "true"
|
||||
- Disables long file names
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to local (Local Disk).
|
||||
|
||||
#### --copy-links
|
||||
|
||||
Follow symlinks and copy the pointed to item.
|
||||
|
||||
- Config: copy_links
|
||||
- Env Var: RCLONE_LOCAL_COPY_LINKS
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --skip-links
|
||||
|
||||
Don't warn about skipped symlinks.
|
||||
This flag disables warning messages on skipped symlinks or junction
|
||||
points, as you explicitly acknowledge that they should be skipped.
|
||||
|
||||
- Config: skip_links
|
||||
- Env Var: RCLONE_LOCAL_SKIP_LINKS
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --local-no-unicode-normalization
|
||||
|
||||
Don't apply unicode normalization to paths and filenames (Deprecated)
|
||||
|
||||
This flag is deprecated now. Rclone no longer normalizes unicode file
|
||||
names, but it compares them with unicode normalization in the sync
|
||||
routine instead.
|
||||
|
||||
- Config: no_unicode_normalization
|
||||
- Env Var: RCLONE_LOCAL_NO_UNICODE_NORMALIZATION
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --local-no-check-updated
|
||||
|
||||
Don't check to see if the files change during upload
|
||||
|
||||
Normally rclone checks the size and modification time of files as they
|
||||
are being uploaded and aborts with a message which starts "can't copy
|
||||
- source file is being updated" if the file changes during upload.
|
||||
|
||||
However on some file systems this modification time check may fail (eg
|
||||
[Glusterfs #2206](https://github.com/ncw/rclone/issues/2206)) so this
|
||||
check can be disabled with this flag.
|
||||
|
||||
- Config: no_check_updated
|
||||
- Env Var: RCLONE_LOCAL_NO_CHECK_UPDATED
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --one-file-system
|
||||
|
||||
Don't cross filesystem boundaries (unix/macOS only).
|
||||
|
||||
- Config: one_file_system
|
||||
- Env Var: RCLONE_LOCAL_ONE_FILE_SYSTEM
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
@ -96,8 +96,59 @@ messages in the log about duplicates.
|
||||
|
||||
Use `rclone dedupe` to fix duplicated files.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/mega/mega.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to mega (Mega).
|
||||
|
||||
#### --mega-user
|
||||
|
||||
User name
|
||||
|
||||
- Config: user
|
||||
- Env Var: RCLONE_MEGA_USER
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --mega-pass
|
||||
|
||||
Password.
|
||||
|
||||
- Config: pass
|
||||
- Env Var: RCLONE_MEGA_PASS
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to mega (Mega).
|
||||
|
||||
#### --mega-debug
|
||||
|
||||
Output more debug from Mega.
|
||||
|
||||
If this flag is set (along with -vv) it will print further debugging
|
||||
information from the mega backend.
|
||||
|
||||
- Config: debug
|
||||
- Env Var: RCLONE_MEGA_DEBUG
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --mega-hard-delete
|
||||
|
||||
Delete files permanently rather than putting them into the trash.
|
||||
|
||||
Normally the mega backend will put all deletions into the trash rather
|
||||
than permanently deleting them. If you specify this then rclone will
|
||||
permanently delete objects instead.
|
||||
|
||||
- Config: hard_delete
|
||||
- Env Var: RCLONE_MEGA_HARD_DELETE
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
### Limitations ###
|
||||
|
||||
|
@ -155,8 +155,81 @@ doesn't provide an API to permanently delete files, nor to empty the
|
||||
trash, so you will have to do that with one of Microsoft's apps or via
|
||||
the OneDrive website.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/onedrive/onedrive.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to onedrive (Microsoft OneDrive).
|
||||
|
||||
#### --onedrive-client-id
|
||||
|
||||
Microsoft App Client Id
|
||||
Leave blank normally.
|
||||
|
||||
- Config: client_id
|
||||
- Env Var: RCLONE_ONEDRIVE_CLIENT_ID
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --onedrive-client-secret
|
||||
|
||||
Microsoft App Client Secret
|
||||
Leave blank normally.
|
||||
|
||||
- Config: client_secret
|
||||
- Env Var: RCLONE_ONEDRIVE_CLIENT_SECRET
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to onedrive (Microsoft OneDrive).
|
||||
|
||||
#### --onedrive-chunk-size
|
||||
|
||||
Chunk size to upload files with - must be multiple of 320k.
|
||||
|
||||
Above this size files will be chunked - must be multiple of 320k. Note
|
||||
that the chunks will be buffered into memory.
|
||||
|
||||
- Config: chunk_size
|
||||
- Env Var: RCLONE_ONEDRIVE_CHUNK_SIZE
|
||||
- Type: SizeSuffix
|
||||
- Default: 10M
|
||||
|
||||
#### --onedrive-drive-id
|
||||
|
||||
The ID of the drive to use
|
||||
|
||||
- Config: drive_id
|
||||
- Env Var: RCLONE_ONEDRIVE_DRIVE_ID
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --onedrive-drive-type
|
||||
|
||||
The type of the drive ( personal | business | documentLibrary )
|
||||
|
||||
- Config: drive_type
|
||||
- Env Var: RCLONE_ONEDRIVE_DRIVE_TYPE
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --onedrive-expose-onenote-files
|
||||
|
||||
Set to make OneNote files show up in directory listings.
|
||||
|
||||
By default rclone will hide OneNote files in directory listings because
|
||||
operations like "Open" and "Update" won't work on them. But this
|
||||
behaviour may also prevent you from deleting them. If you want to
|
||||
delete OneNote files or otherwise want them to show up in directory
|
||||
listing, set this option.
|
||||
|
||||
- Config: expose_onenote_files
|
||||
- Env Var: RCLONE_ONEDRIVE_EXPOSE_ONENOTE_FILES
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
### Limitations ###
|
||||
|
||||
|
@ -93,8 +93,30 @@ OpenDrive allows modification times to be set on objects accurate to 1
|
||||
second. These will be used to detect whether objects need syncing or
|
||||
not.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/opendrive/opendrive.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to opendrive (OpenDrive).
|
||||
|
||||
#### --opendrive-username
|
||||
|
||||
Username
|
||||
|
||||
- Config: username
|
||||
- Env Var: RCLONE_OPENDRIVE_USERNAME
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --opendrive-password
|
||||
|
||||
Password.
|
||||
|
||||
- Config: password
|
||||
- Env Var: RCLONE_OPENDRIVE_PASSWORD
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
### Limitations ###
|
||||
|
||||
|
@ -134,5 +134,29 @@ Deleted files will be moved to the trash. Your subscription level
|
||||
will determine how long items stay in the trash. `rclone cleanup` can
|
||||
be used to empty the trash.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/pcloud/pcloud.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to pcloud (Pcloud).
|
||||
|
||||
#### --pcloud-client-id
|
||||
|
||||
Pcloud App Client Id
|
||||
Leave blank normally.
|
||||
|
||||
- Config: client_id
|
||||
- Env Var: RCLONE_PCLOUD_CLIENT_ID
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --pcloud-client-secret
|
||||
|
||||
Pcloud App Client Secret
|
||||
Leave blank normally.
|
||||
|
||||
- Config: client_secret
|
||||
- Env Var: RCLONE_PCLOUD_CLIENT_SECRET
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
@ -152,5 +152,86 @@ credentials. In order of precedence:
|
||||
- Access Key ID: `QS_ACCESS_KEY_ID` or `QS_ACCESS_KEY`
|
||||
- Secret Access Key: `QS_SECRET_ACCESS_KEY` or `QS_SECRET_KEY`
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/qingstor/qingstor.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to qingstor (QingCloud Object Storage).
|
||||
|
||||
#### --qingstor-env-auth
|
||||
|
||||
Get QingStor credentials from runtime. Only applies if access_key_id and secret_access_key is blank.
|
||||
|
||||
- Config: env_auth
|
||||
- Env Var: RCLONE_QINGSTOR_ENV_AUTH
|
||||
- Type: bool
|
||||
- Default: false
|
||||
- Examples:
|
||||
- "false"
|
||||
- Enter QingStor credentials in the next step
|
||||
- "true"
|
||||
- Get QingStor credentials from the environment (env vars or IAM)
|
||||
|
||||
#### --qingstor-access-key-id
|
||||
|
||||
QingStor Access Key ID
|
||||
Leave blank for anonymous access or runtime credentials.
|
||||
|
||||
- Config: access_key_id
|
||||
- Env Var: RCLONE_QINGSTOR_ACCESS_KEY_ID
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --qingstor-secret-access-key
|
||||
|
||||
QingStor Secret Access Key (password)
|
||||
Leave blank for anonymous access or runtime credentials.
|
||||
|
||||
- Config: secret_access_key
|
||||
- Env Var: RCLONE_QINGSTOR_SECRET_ACCESS_KEY
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --qingstor-endpoint
|
||||
|
||||
Enter a endpoint URL to connection QingStor API.
|
||||
Leave blank will use the default value "https://qingstor.com:443"
|
||||
|
||||
- Config: endpoint
|
||||
- Env Var: RCLONE_QINGSTOR_ENDPOINT
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --qingstor-zone
|
||||
|
||||
Zone to connect to.
|
||||
Default is "pek3a".
|
||||
|
||||
- Config: zone
|
||||
- Env Var: RCLONE_QINGSTOR_ZONE
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "pek3a"
|
||||
- The Beijing (China) Three Zone
|
||||
- Needs location constraint pek3a.
|
||||
- "sh1a"
|
||||
- The Shanghai (China) First Zone
|
||||
- Needs location constraint sh1a.
|
||||
- "gd2a"
|
||||
- The Guangdong (China) Second Zone
|
||||
- Needs location constraint gd2a.
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to qingstor (QingCloud Object Storage).
|
||||
|
||||
#### --qingstor-connection-retries
|
||||
|
||||
Number of connnection retries.
|
||||
|
||||
- Config: connection_retries
|
||||
- Env Var: RCLONE_QINGSTOR_CONNECTION_RETRIES
|
||||
- Type: int
|
||||
- Default: 3
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
@ -370,8 +370,531 @@ tries to access the data you will see an error like below.
|
||||
In this case you need to [restore](http://docs.aws.amazon.com/AmazonS3/latest/user-guide/restore-archived-objects.html)
|
||||
the object(s) in question before using rclone.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/s3/s3.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to s3 (Amazon S3 Compliant Storage Providers (AWS, Ceph, Dreamhost, IBM COS, Minio)).
|
||||
|
||||
#### --s3-provider
|
||||
|
||||
Choose your S3 provider.
|
||||
|
||||
- Config: provider
|
||||
- Env Var: RCLONE_S3_PROVIDER
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "AWS"
|
||||
- Amazon Web Services (AWS) S3
|
||||
- "Ceph"
|
||||
- Ceph Object Storage
|
||||
- "DigitalOcean"
|
||||
- Digital Ocean Spaces
|
||||
- "Dreamhost"
|
||||
- Dreamhost DreamObjects
|
||||
- "IBMCOS"
|
||||
- IBM COS S3
|
||||
- "Minio"
|
||||
- Minio Object Storage
|
||||
- "Wasabi"
|
||||
- Wasabi Object Storage
|
||||
- "Other"
|
||||
- Any other S3 compatible provider
|
||||
|
||||
#### --s3-env-auth
|
||||
|
||||
Get AWS credentials from runtime (environment variables or EC2/ECS meta data if no env vars).
|
||||
Only applies if access_key_id and secret_access_key is blank.
|
||||
|
||||
- Config: env_auth
|
||||
- Env Var: RCLONE_S3_ENV_AUTH
|
||||
- Type: bool
|
||||
- Default: false
|
||||
- Examples:
|
||||
- "false"
|
||||
- Enter AWS credentials in the next step
|
||||
- "true"
|
||||
- Get AWS credentials from the environment (env vars or IAM)
|
||||
|
||||
#### --s3-access-key-id
|
||||
|
||||
AWS Access Key ID.
|
||||
Leave blank for anonymous access or runtime credentials.
|
||||
|
||||
- Config: access_key_id
|
||||
- Env Var: RCLONE_S3_ACCESS_KEY_ID
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --s3-secret-access-key
|
||||
|
||||
AWS Secret Access Key (password)
|
||||
Leave blank for anonymous access or runtime credentials.
|
||||
|
||||
- Config: secret_access_key
|
||||
- Env Var: RCLONE_S3_SECRET_ACCESS_KEY
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --s3-region
|
||||
|
||||
Region to connect to.
|
||||
|
||||
- Config: region
|
||||
- Env Var: RCLONE_S3_REGION
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "us-east-1"
|
||||
- The default endpoint - a good choice if you are unsure.
|
||||
- US Region, Northern Virginia or Pacific Northwest.
|
||||
- Leave location constraint empty.
|
||||
- "us-east-2"
|
||||
- US East (Ohio) Region
|
||||
- Needs location constraint us-east-2.
|
||||
- "us-west-2"
|
||||
- US West (Oregon) Region
|
||||
- Needs location constraint us-west-2.
|
||||
- "us-west-1"
|
||||
- US West (Northern California) Region
|
||||
- Needs location constraint us-west-1.
|
||||
- "ca-central-1"
|
||||
- Canada (Central) Region
|
||||
- Needs location constraint ca-central-1.
|
||||
- "eu-west-1"
|
||||
- EU (Ireland) Region
|
||||
- Needs location constraint EU or eu-west-1.
|
||||
- "eu-west-2"
|
||||
- EU (London) Region
|
||||
- Needs location constraint eu-west-2.
|
||||
- "eu-central-1"
|
||||
- EU (Frankfurt) Region
|
||||
- Needs location constraint eu-central-1.
|
||||
- "ap-southeast-1"
|
||||
- Asia Pacific (Singapore) Region
|
||||
- Needs location constraint ap-southeast-1.
|
||||
- "ap-southeast-2"
|
||||
- Asia Pacific (Sydney) Region
|
||||
- Needs location constraint ap-southeast-2.
|
||||
- "ap-northeast-1"
|
||||
- Asia Pacific (Tokyo) Region
|
||||
- Needs location constraint ap-northeast-1.
|
||||
- "ap-northeast-2"
|
||||
- Asia Pacific (Seoul)
|
||||
- Needs location constraint ap-northeast-2.
|
||||
- "ap-south-1"
|
||||
- Asia Pacific (Mumbai)
|
||||
- Needs location constraint ap-south-1.
|
||||
- "sa-east-1"
|
||||
- South America (Sao Paulo) Region
|
||||
- Needs location constraint sa-east-1.
|
||||
|
||||
#### --s3-region
|
||||
|
||||
Region to connect to.
|
||||
Leave blank if you are using an S3 clone and you don't have a region.
|
||||
|
||||
- Config: region
|
||||
- Env Var: RCLONE_S3_REGION
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- ""
|
||||
- Use this if unsure. Will use v4 signatures and an empty region.
|
||||
- "other-v2-signature"
|
||||
- Use this only if v4 signatures don't work, eg pre Jewel/v10 CEPH.
|
||||
|
||||
#### --s3-endpoint
|
||||
|
||||
Endpoint for S3 API.
|
||||
Leave blank if using AWS to use the default endpoint for the region.
|
||||
|
||||
- Config: endpoint
|
||||
- Env Var: RCLONE_S3_ENDPOINT
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --s3-endpoint
|
||||
|
||||
Endpoint for IBM COS S3 API.
|
||||
Specify if using an IBM COS On Premise.
|
||||
|
||||
- Config: endpoint
|
||||
- Env Var: RCLONE_S3_ENDPOINT
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "s3-api.us-geo.objectstorage.softlayer.net"
|
||||
- US Cross Region Endpoint
|
||||
- "s3-api.dal.us-geo.objectstorage.softlayer.net"
|
||||
- US Cross Region Dallas Endpoint
|
||||
- "s3-api.wdc-us-geo.objectstorage.softlayer.net"
|
||||
- US Cross Region Washington DC Endpoint
|
||||
- "s3-api.sjc-us-geo.objectstorage.softlayer.net"
|
||||
- US Cross Region San Jose Endpoint
|
||||
- "s3-api.us-geo.objectstorage.service.networklayer.com"
|
||||
- US Cross Region Private Endpoint
|
||||
- "s3-api.dal-us-geo.objectstorage.service.networklayer.com"
|
||||
- US Cross Region Dallas Private Endpoint
|
||||
- "s3-api.wdc-us-geo.objectstorage.service.networklayer.com"
|
||||
- US Cross Region Washington DC Private Endpoint
|
||||
- "s3-api.sjc-us-geo.objectstorage.service.networklayer.com"
|
||||
- US Cross Region San Jose Private Endpoint
|
||||
- "s3.us-east.objectstorage.softlayer.net"
|
||||
- US Region East Endpoint
|
||||
- "s3.us-east.objectstorage.service.networklayer.com"
|
||||
- US Region East Private Endpoint
|
||||
- "s3.us-south.objectstorage.softlayer.net"
|
||||
- US Region South Endpoint
|
||||
- "s3.us-south.objectstorage.service.networklayer.com"
|
||||
- US Region South Private Endpoint
|
||||
- "s3.eu-geo.objectstorage.softlayer.net"
|
||||
- EU Cross Region Endpoint
|
||||
- "s3.fra-eu-geo.objectstorage.softlayer.net"
|
||||
- EU Cross Region Frankfurt Endpoint
|
||||
- "s3.mil-eu-geo.objectstorage.softlayer.net"
|
||||
- EU Cross Region Milan Endpoint
|
||||
- "s3.ams-eu-geo.objectstorage.softlayer.net"
|
||||
- EU Cross Region Amsterdam Endpoint
|
||||
- "s3.eu-geo.objectstorage.service.networklayer.com"
|
||||
- EU Cross Region Private Endpoint
|
||||
- "s3.fra-eu-geo.objectstorage.service.networklayer.com"
|
||||
- EU Cross Region Frankfurt Private Endpoint
|
||||
- "s3.mil-eu-geo.objectstorage.service.networklayer.com"
|
||||
- EU Cross Region Milan Private Endpoint
|
||||
- "s3.ams-eu-geo.objectstorage.service.networklayer.com"
|
||||
- EU Cross Region Amsterdam Private Endpoint
|
||||
- "s3.eu-gb.objectstorage.softlayer.net"
|
||||
- Great Britan Endpoint
|
||||
- "s3.eu-gb.objectstorage.service.networklayer.com"
|
||||
- Great Britan Private Endpoint
|
||||
- "s3.ap-geo.objectstorage.softlayer.net"
|
||||
- APAC Cross Regional Endpoint
|
||||
- "s3.tok-ap-geo.objectstorage.softlayer.net"
|
||||
- APAC Cross Regional Tokyo Endpoint
|
||||
- "s3.hkg-ap-geo.objectstorage.softlayer.net"
|
||||
- APAC Cross Regional HongKong Endpoint
|
||||
- "s3.seo-ap-geo.objectstorage.softlayer.net"
|
||||
- APAC Cross Regional Seoul Endpoint
|
||||
- "s3.ap-geo.objectstorage.service.networklayer.com"
|
||||
- APAC Cross Regional Private Endpoint
|
||||
- "s3.tok-ap-geo.objectstorage.service.networklayer.com"
|
||||
- APAC Cross Regional Tokyo Private Endpoint
|
||||
- "s3.hkg-ap-geo.objectstorage.service.networklayer.com"
|
||||
- APAC Cross Regional HongKong Private Endpoint
|
||||
- "s3.seo-ap-geo.objectstorage.service.networklayer.com"
|
||||
- APAC Cross Regional Seoul Private Endpoint
|
||||
- "s3.mel01.objectstorage.softlayer.net"
|
||||
- Melbourne Single Site Endpoint
|
||||
- "s3.mel01.objectstorage.service.networklayer.com"
|
||||
- Melbourne Single Site Private Endpoint
|
||||
- "s3.tor01.objectstorage.softlayer.net"
|
||||
- Toronto Single Site Endpoint
|
||||
- "s3.tor01.objectstorage.service.networklayer.com"
|
||||
- Toronto Single Site Private Endpoint
|
||||
|
||||
#### --s3-endpoint
|
||||
|
||||
Endpoint for S3 API.
|
||||
Required when using an S3 clone.
|
||||
|
||||
- Config: endpoint
|
||||
- Env Var: RCLONE_S3_ENDPOINT
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "objects-us-west-1.dream.io"
|
||||
- Dream Objects endpoint
|
||||
- "nyc3.digitaloceanspaces.com"
|
||||
- Digital Ocean Spaces New York 3
|
||||
- "ams3.digitaloceanspaces.com"
|
||||
- Digital Ocean Spaces Amsterdam 3
|
||||
- "sgp1.digitaloceanspaces.com"
|
||||
- Digital Ocean Spaces Singapore 1
|
||||
- "s3.wasabisys.com"
|
||||
- Wasabi Object Storage
|
||||
|
||||
#### --s3-location-constraint
|
||||
|
||||
Location constraint - must be set to match the Region.
|
||||
Used when creating buckets only.
|
||||
|
||||
- Config: location_constraint
|
||||
- Env Var: RCLONE_S3_LOCATION_CONSTRAINT
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- ""
|
||||
- Empty for US Region, Northern Virginia or Pacific Northwest.
|
||||
- "us-east-2"
|
||||
- US East (Ohio) Region.
|
||||
- "us-west-2"
|
||||
- US West (Oregon) Region.
|
||||
- "us-west-1"
|
||||
- US West (Northern California) Region.
|
||||
- "ca-central-1"
|
||||
- Canada (Central) Region.
|
||||
- "eu-west-1"
|
||||
- EU (Ireland) Region.
|
||||
- "eu-west-2"
|
||||
- EU (London) Region.
|
||||
- "EU"
|
||||
- EU Region.
|
||||
- "ap-southeast-1"
|
||||
- Asia Pacific (Singapore) Region.
|
||||
- "ap-southeast-2"
|
||||
- Asia Pacific (Sydney) Region.
|
||||
- "ap-northeast-1"
|
||||
- Asia Pacific (Tokyo) Region.
|
||||
- "ap-northeast-2"
|
||||
- Asia Pacific (Seoul)
|
||||
- "ap-south-1"
|
||||
- Asia Pacific (Mumbai)
|
||||
- "sa-east-1"
|
||||
- South America (Sao Paulo) Region.
|
||||
|
||||
#### --s3-location-constraint
|
||||
|
||||
Location constraint - must match endpoint when using IBM Cloud Public.
|
||||
For on-prem COS, do not make a selection from this list, hit enter
|
||||
|
||||
- Config: location_constraint
|
||||
- Env Var: RCLONE_S3_LOCATION_CONSTRAINT
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "us-standard"
|
||||
- US Cross Region Standard
|
||||
- "us-vault"
|
||||
- US Cross Region Vault
|
||||
- "us-cold"
|
||||
- US Cross Region Cold
|
||||
- "us-flex"
|
||||
- US Cross Region Flex
|
||||
- "us-east-standard"
|
||||
- US East Region Standard
|
||||
- "us-east-vault"
|
||||
- US East Region Vault
|
||||
- "us-east-cold"
|
||||
- US East Region Cold
|
||||
- "us-east-flex"
|
||||
- US East Region Flex
|
||||
- "us-south-standard"
|
||||
- US Sout hRegion Standard
|
||||
- "us-south-vault"
|
||||
- US South Region Vault
|
||||
- "us-south-cold"
|
||||
- US South Region Cold
|
||||
- "us-south-flex"
|
||||
- US South Region Flex
|
||||
- "eu-standard"
|
||||
- EU Cross Region Standard
|
||||
- "eu-vault"
|
||||
- EU Cross Region Vault
|
||||
- "eu-cold"
|
||||
- EU Cross Region Cold
|
||||
- "eu-flex"
|
||||
- EU Cross Region Flex
|
||||
- "eu-gb-standard"
|
||||
- Great Britan Standard
|
||||
- "eu-gb-vault"
|
||||
- Great Britan Vault
|
||||
- "eu-gb-cold"
|
||||
- Great Britan Cold
|
||||
- "eu-gb-flex"
|
||||
- Great Britan Flex
|
||||
- "ap-standard"
|
||||
- APAC Standard
|
||||
- "ap-vault"
|
||||
- APAC Vault
|
||||
- "ap-cold"
|
||||
- APAC Cold
|
||||
- "ap-flex"
|
||||
- APAC Flex
|
||||
- "mel01-standard"
|
||||
- Melbourne Standard
|
||||
- "mel01-vault"
|
||||
- Melbourne Vault
|
||||
- "mel01-cold"
|
||||
- Melbourne Cold
|
||||
- "mel01-flex"
|
||||
- Melbourne Flex
|
||||
- "tor01-standard"
|
||||
- Toronto Standard
|
||||
- "tor01-vault"
|
||||
- Toronto Vault
|
||||
- "tor01-cold"
|
||||
- Toronto Cold
|
||||
- "tor01-flex"
|
||||
- Toronto Flex
|
||||
|
||||
#### --s3-location-constraint
|
||||
|
||||
Location constraint - must be set to match the Region.
|
||||
Leave blank if not sure. Used when creating buckets only.
|
||||
|
||||
- Config: location_constraint
|
||||
- Env Var: RCLONE_S3_LOCATION_CONSTRAINT
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --s3-acl
|
||||
|
||||
Canned ACL used when creating buckets and/or storing objects in S3.
|
||||
For more info visit https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html#canned-acl
|
||||
|
||||
- Config: acl
|
||||
- Env Var: RCLONE_S3_ACL
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "private"
|
||||
- Owner gets FULL_CONTROL. No one else has access rights (default).
|
||||
- "public-read"
|
||||
- Owner gets FULL_CONTROL. The AllUsers group gets READ access.
|
||||
- "public-read-write"
|
||||
- Owner gets FULL_CONTROL. The AllUsers group gets READ and WRITE access.
|
||||
- Granting this on a bucket is generally not recommended.
|
||||
- "authenticated-read"
|
||||
- Owner gets FULL_CONTROL. The AuthenticatedUsers group gets READ access.
|
||||
- "bucket-owner-read"
|
||||
- Object owner gets FULL_CONTROL. Bucket owner gets READ access.
|
||||
- If you specify this canned ACL when creating a bucket, Amazon S3 ignores it.
|
||||
- "bucket-owner-full-control"
|
||||
- Both the object owner and the bucket owner get FULL_CONTROL over the object.
|
||||
- If you specify this canned ACL when creating a bucket, Amazon S3 ignores it.
|
||||
- "private"
|
||||
- Owner gets FULL_CONTROL. No one else has access rights (default). This acl is available on IBM Cloud (Infra), IBM Cloud (Storage), On-Premise COS
|
||||
- "public-read"
|
||||
- Owner gets FULL_CONTROL. The AllUsers group gets READ access. This acl is available on IBM Cloud (Infra), IBM Cloud (Storage), On-Premise IBM COS
|
||||
- "public-read-write"
|
||||
- Owner gets FULL_CONTROL. The AllUsers group gets READ and WRITE access. This acl is available on IBM Cloud (Infra), On-Premise IBM COS
|
||||
- "authenticated-read"
|
||||
- Owner gets FULL_CONTROL. The AuthenticatedUsers group gets READ access. Not supported on Buckets. This acl is available on IBM Cloud (Infra) and On-Premise IBM COS
|
||||
|
||||
#### --s3-server-side-encryption
|
||||
|
||||
The server-side encryption algorithm used when storing this object in S3.
|
||||
|
||||
- Config: server_side_encryption
|
||||
- Env Var: RCLONE_S3_SERVER_SIDE_ENCRYPTION
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- ""
|
||||
- None
|
||||
- "AES256"
|
||||
- AES256
|
||||
- "aws:kms"
|
||||
- aws:kms
|
||||
|
||||
#### --s3-sse-kms-key-id
|
||||
|
||||
If using KMS ID you must provide the ARN of Key.
|
||||
|
||||
- Config: sse_kms_key_id
|
||||
- Env Var: RCLONE_S3_SSE_KMS_KEY_ID
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- ""
|
||||
- None
|
||||
- "arn:aws:kms:us-east-1:*"
|
||||
- arn:aws:kms:*
|
||||
|
||||
#### --s3-storage-class
|
||||
|
||||
The storage class to use when storing new objects in S3.
|
||||
|
||||
- Config: storage_class
|
||||
- Env Var: RCLONE_S3_STORAGE_CLASS
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- ""
|
||||
- Default
|
||||
- "STANDARD"
|
||||
- Standard storage class
|
||||
- "REDUCED_REDUNDANCY"
|
||||
- Reduced redundancy storage class
|
||||
- "STANDARD_IA"
|
||||
- Standard Infrequent Access storage class
|
||||
- "ONEZONE_IA"
|
||||
- One Zone Infrequent Access storage class
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to s3 (Amazon S3 Compliant Storage Providers (AWS, Ceph, Dreamhost, IBM COS, Minio)).
|
||||
|
||||
#### --s3-chunk-size
|
||||
|
||||
Chunk size to use for uploading.
|
||||
|
||||
Any files larger than this will be uploaded in chunks of this
|
||||
size. The default is 5MB. The minimum is 5MB.
|
||||
|
||||
Note that "--s3-upload-concurrency" chunks of this size are buffered
|
||||
in memory per transfer.
|
||||
|
||||
If you are transferring large files over high speed links and you have
|
||||
enough memory, then increasing this will speed up the transfers.
|
||||
|
||||
- Config: chunk_size
|
||||
- Env Var: RCLONE_S3_CHUNK_SIZE
|
||||
- Type: SizeSuffix
|
||||
- Default: 5M
|
||||
|
||||
#### --s3-disable-checksum
|
||||
|
||||
Don't store MD5 checksum with object metadata
|
||||
|
||||
- Config: disable_checksum
|
||||
- Env Var: RCLONE_S3_DISABLE_CHECKSUM
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --s3-session-token
|
||||
|
||||
An AWS session token
|
||||
|
||||
- Config: session_token
|
||||
- Env Var: RCLONE_S3_SESSION_TOKEN
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --s3-upload-concurrency
|
||||
|
||||
Concurrency for multipart uploads.
|
||||
|
||||
This is the number of chunks of the same file that are uploaded
|
||||
concurrently.
|
||||
|
||||
If you are uploading small numbers of large file over high speed link
|
||||
and these uploads do not fully utilize your bandwidth, then increasing
|
||||
this may help to speed up the transfers.
|
||||
|
||||
- Config: upload_concurrency
|
||||
- Env Var: RCLONE_S3_UPLOAD_CONCURRENCY
|
||||
- Type: int
|
||||
- Default: 2
|
||||
|
||||
#### --s3-force-path-style
|
||||
|
||||
If true use path style access if false use virtual hosted style.
|
||||
|
||||
If this is true (the default) then rclone will use path style access,
|
||||
if false then rclone will use virtual path style. See [the AWS S3
|
||||
docs](https://docs.aws.amazon.com/AmazonS3/latest/dev/UsingBucket.html#access-bucket-intro)
|
||||
for more info.
|
||||
|
||||
Some providers (eg Aliyun OSS or Netease COS) require this set to false.
|
||||
|
||||
- Config: force_path_style
|
||||
- Env Var: RCLONE_S3_FORCE_PATH_STYLE
|
||||
- Type: bool
|
||||
- Default: true
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
### Anonymous access to public buckets ###
|
||||
|
||||
|
@ -158,8 +158,126 @@ upload (for example, certain configurations of ProFTPd with mod_sftp). If you
|
||||
are using one of these servers, you can set the option `set_modtime = false` in
|
||||
your RClone backend configuration to disable this behaviour.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/sftp/sftp.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to sftp (SSH/SFTP Connection).
|
||||
|
||||
#### --sftp-host
|
||||
|
||||
SSH host to connect to
|
||||
|
||||
- Config: host
|
||||
- Env Var: RCLONE_SFTP_HOST
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "example.com"
|
||||
- Connect to example.com
|
||||
|
||||
#### --sftp-user
|
||||
|
||||
SSH username, leave blank for current username, ncw
|
||||
|
||||
- Config: user
|
||||
- Env Var: RCLONE_SFTP_USER
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --sftp-port
|
||||
|
||||
SSH port, leave blank to use default (22)
|
||||
|
||||
- Config: port
|
||||
- Env Var: RCLONE_SFTP_PORT
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --sftp-pass
|
||||
|
||||
SSH password, leave blank to use ssh-agent.
|
||||
|
||||
- Config: pass
|
||||
- Env Var: RCLONE_SFTP_PASS
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --sftp-key-file
|
||||
|
||||
Path to unencrypted PEM-encoded private key file, leave blank to use ssh-agent.
|
||||
|
||||
- Config: key_file
|
||||
- Env Var: RCLONE_SFTP_KEY_FILE
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --sftp-use-insecure-cipher
|
||||
|
||||
Enable the use of the aes128-cbc cipher. This cipher is insecure and may allow plaintext data to be recovered by an attacker.
|
||||
|
||||
- Config: use_insecure_cipher
|
||||
- Env Var: RCLONE_SFTP_USE_INSECURE_CIPHER
|
||||
- Type: bool
|
||||
- Default: false
|
||||
- Examples:
|
||||
- "false"
|
||||
- Use default Cipher list.
|
||||
- "true"
|
||||
- Enables the use of the aes128-cbc cipher.
|
||||
|
||||
#### --sftp-disable-hashcheck
|
||||
|
||||
Disable the execution of SSH commands to determine if remote file hashing is available.
|
||||
Leave blank or set to false to enable hashing (recommended), set to true to disable hashing.
|
||||
|
||||
- Config: disable_hashcheck
|
||||
- Env Var: RCLONE_SFTP_DISABLE_HASHCHECK
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to sftp (SSH/SFTP Connection).
|
||||
|
||||
#### --sftp-ask-password
|
||||
|
||||
Allow asking for SFTP password when needed.
|
||||
|
||||
- Config: ask_password
|
||||
- Env Var: RCLONE_SFTP_ASK_PASSWORD
|
||||
- Type: bool
|
||||
- Default: false
|
||||
|
||||
#### --sftp-path-override
|
||||
|
||||
Override path used by SSH connection.
|
||||
|
||||
This allows checksum calculation when SFTP and SSH paths are
|
||||
different. This issue affects among others Synology NAS boxes.
|
||||
|
||||
Shared folders can be found in directories representing volumes
|
||||
|
||||
rclone sync /home/local/directory remote:/directory --ssh-path-override /volume2/directory
|
||||
|
||||
Home directory can be found in a shared folder called "home"
|
||||
|
||||
rclone sync /home/local/directory remote:/home/directory --ssh-path-override /volume1/homes/USER/directory
|
||||
|
||||
- Config: path_override
|
||||
- Env Var: RCLONE_SFTP_PATH_OVERRIDE
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --sftp-set-modtime
|
||||
|
||||
Set the modified time on the remote if set.
|
||||
|
||||
- Config: set_modtime
|
||||
- Env Var: RCLONE_SFTP_SET_MODTIME
|
||||
- Type: bool
|
||||
- Default: true
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
### Limitations ###
|
||||
|
||||
|
@ -261,8 +261,200 @@ sufficient to determine if it is "dirty". By using `--update` along with
|
||||
`--use-server-modtime`, you can avoid the extra API call and simply upload
|
||||
files whose local modtime is newer than the time it was last uploaded.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/swift/swift.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to swift (Openstack Swift (Rackspace Cloud Files, Memset Memstore, OVH)).
|
||||
|
||||
#### --swift-env-auth
|
||||
|
||||
Get swift credentials from environment variables in standard OpenStack form.
|
||||
|
||||
- Config: env_auth
|
||||
- Env Var: RCLONE_SWIFT_ENV_AUTH
|
||||
- Type: bool
|
||||
- Default: false
|
||||
- Examples:
|
||||
- "false"
|
||||
- Enter swift credentials in the next step
|
||||
- "true"
|
||||
- Get swift credentials from environment vars. Leave other fields blank if using this.
|
||||
|
||||
#### --swift-user
|
||||
|
||||
User name to log in (OS_USERNAME).
|
||||
|
||||
- Config: user
|
||||
- Env Var: RCLONE_SWIFT_USER
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --swift-key
|
||||
|
||||
API key or password (OS_PASSWORD).
|
||||
|
||||
- Config: key
|
||||
- Env Var: RCLONE_SWIFT_KEY
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --swift-auth
|
||||
|
||||
Authentication URL for server (OS_AUTH_URL).
|
||||
|
||||
- Config: auth
|
||||
- Env Var: RCLONE_SWIFT_AUTH
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "https://auth.api.rackspacecloud.com/v1.0"
|
||||
- Rackspace US
|
||||
- "https://lon.auth.api.rackspacecloud.com/v1.0"
|
||||
- Rackspace UK
|
||||
- "https://identity.api.rackspacecloud.com/v2.0"
|
||||
- Rackspace v2
|
||||
- "https://auth.storage.memset.com/v1.0"
|
||||
- Memset Memstore UK
|
||||
- "https://auth.storage.memset.com/v2.0"
|
||||
- Memset Memstore UK v2
|
||||
- "https://auth.cloud.ovh.net/v2.0"
|
||||
- OVH
|
||||
|
||||
#### --swift-user-id
|
||||
|
||||
User ID to log in - optional - most swift systems use user and leave this blank (v3 auth) (OS_USER_ID).
|
||||
|
||||
- Config: user_id
|
||||
- Env Var: RCLONE_SWIFT_USER_ID
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --swift-domain
|
||||
|
||||
User domain - optional (v3 auth) (OS_USER_DOMAIN_NAME)
|
||||
|
||||
- Config: domain
|
||||
- Env Var: RCLONE_SWIFT_DOMAIN
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --swift-tenant
|
||||
|
||||
Tenant name - optional for v1 auth, this or tenant_id required otherwise (OS_TENANT_NAME or OS_PROJECT_NAME)
|
||||
|
||||
- Config: tenant
|
||||
- Env Var: RCLONE_SWIFT_TENANT
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --swift-tenant-id
|
||||
|
||||
Tenant ID - optional for v1 auth, this or tenant required otherwise (OS_TENANT_ID)
|
||||
|
||||
- Config: tenant_id
|
||||
- Env Var: RCLONE_SWIFT_TENANT_ID
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --swift-tenant-domain
|
||||
|
||||
Tenant domain - optional (v3 auth) (OS_PROJECT_DOMAIN_NAME)
|
||||
|
||||
- Config: tenant_domain
|
||||
- Env Var: RCLONE_SWIFT_TENANT_DOMAIN
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --swift-region
|
||||
|
||||
Region name - optional (OS_REGION_NAME)
|
||||
|
||||
- Config: region
|
||||
- Env Var: RCLONE_SWIFT_REGION
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --swift-storage-url
|
||||
|
||||
Storage URL - optional (OS_STORAGE_URL)
|
||||
|
||||
- Config: storage_url
|
||||
- Env Var: RCLONE_SWIFT_STORAGE_URL
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --swift-auth-token
|
||||
|
||||
Auth Token from alternate authentication - optional (OS_AUTH_TOKEN)
|
||||
|
||||
- Config: auth_token
|
||||
- Env Var: RCLONE_SWIFT_AUTH_TOKEN
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --swift-auth-version
|
||||
|
||||
AuthVersion - optional - set to (1,2,3) if your auth URL has no version (ST_AUTH_VERSION)
|
||||
|
||||
- Config: auth_version
|
||||
- Env Var: RCLONE_SWIFT_AUTH_VERSION
|
||||
- Type: int
|
||||
- Default: 0
|
||||
|
||||
#### --swift-endpoint-type
|
||||
|
||||
Endpoint type to choose from the service catalogue (OS_ENDPOINT_TYPE)
|
||||
|
||||
- Config: endpoint_type
|
||||
- Env Var: RCLONE_SWIFT_ENDPOINT_TYPE
|
||||
- Type: string
|
||||
- Default: "public"
|
||||
- Examples:
|
||||
- "public"
|
||||
- Public (default, choose this if not sure)
|
||||
- "internal"
|
||||
- Internal (use internal service net)
|
||||
- "admin"
|
||||
- Admin
|
||||
|
||||
#### --swift-storage-policy
|
||||
|
||||
The storage policy to use when creating a new container
|
||||
|
||||
This applies the specified storage policy when creating a new
|
||||
container. The policy cannot be changed afterwards. The allowed
|
||||
configuration values and their meaning depend on your Swift storage
|
||||
provider.
|
||||
|
||||
- Config: storage_policy
|
||||
- Env Var: RCLONE_SWIFT_STORAGE_POLICY
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- ""
|
||||
- Default
|
||||
- "pcs"
|
||||
- OVH Public Cloud Storage
|
||||
- "pca"
|
||||
- OVH Public Cloud Archive
|
||||
|
||||
### Advanced Options
|
||||
|
||||
Here are the advanced options specific to swift (Openstack Swift (Rackspace Cloud Files, Memset Memstore, OVH)).
|
||||
|
||||
#### --swift-chunk-size
|
||||
|
||||
Above this size files will be chunked into a _segments container.
|
||||
|
||||
Above this size files will be chunked into a _segments container. The
|
||||
default for this is 5GB which is its maximum value.
|
||||
|
||||
- Config: chunk_size
|
||||
- Env Var: RCLONE_SWIFT_CHUNK_SIZE
|
||||
- Type: SizeSuffix
|
||||
- Default: 5G
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
### Modified time ###
|
||||
|
||||
|
@ -142,5 +142,20 @@ Copy another local directory to the union directory called source, which will be
|
||||
|
||||
rclone copy C:\source remote:source
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/union/union.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to union (A stackable unification remote, which can appear to merge the contents of several remotes).
|
||||
|
||||
#### --union-remotes
|
||||
|
||||
List of space separated remotes.
|
||||
Can be 'remotea:test/dir remoteb:', '"remotea:test/space dir" remoteb:', etc.
|
||||
The last remote is used to write to.
|
||||
|
||||
- Config: remotes
|
||||
- Env Var: RCLONE_UNION_REMOTES
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
@ -101,8 +101,69 @@ Owncloud or Nextcloud rclone will support modified times.
|
||||
|
||||
Hashes are not supported.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/webdav/webdav.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to webdav (Webdav).
|
||||
|
||||
#### --webdav-url
|
||||
|
||||
URL of http host to connect to
|
||||
|
||||
- Config: url
|
||||
- Env Var: RCLONE_WEBDAV_URL
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "https://example.com"
|
||||
- Connect to example.com
|
||||
|
||||
#### --webdav-vendor
|
||||
|
||||
Name of the Webdav site/service/software you are using
|
||||
|
||||
- Config: vendor
|
||||
- Env Var: RCLONE_WEBDAV_VENDOR
|
||||
- Type: string
|
||||
- Default: ""
|
||||
- Examples:
|
||||
- "nextcloud"
|
||||
- Nextcloud
|
||||
- "owncloud"
|
||||
- Owncloud
|
||||
- "sharepoint"
|
||||
- Sharepoint
|
||||
- "other"
|
||||
- Other site/service or software
|
||||
|
||||
#### --webdav-user
|
||||
|
||||
User name
|
||||
|
||||
- Config: user
|
||||
- Env Var: RCLONE_WEBDAV_USER
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --webdav-pass
|
||||
|
||||
Password.
|
||||
|
||||
- Config: pass
|
||||
- Env Var: RCLONE_WEBDAV_PASS
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --webdav-bearer-token
|
||||
|
||||
Bearer token instead of user/pass (eg a Macaroon)
|
||||
|
||||
- Config: bearer_token
|
||||
- Env Var: RCLONE_WEBDAV_BEARER_TOKEN
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
||||
## Provider notes ##
|
||||
|
||||
|
@ -128,5 +128,29 @@ If you wish to empty your trash you can use the `rclone cleanup remote:`
|
||||
command which will permanently delete all your trashed files. This command
|
||||
does not take any path arguments.
|
||||
|
||||
<!--- autogenerated options start - edit in backend/backend.go options -->
|
||||
<!--- autogenerated options stop -->
|
||||
<!--- autogenerated options start - DO NOT EDIT, instead edit fs.RegInfo in backend/yandex/yandex.go then run make backenddocs -->
|
||||
### Standard Options
|
||||
|
||||
Here are the standard options specific to yandex (Yandex Disk).
|
||||
|
||||
#### --yandex-client-id
|
||||
|
||||
Yandex Client Id
|
||||
Leave blank normally.
|
||||
|
||||
- Config: client_id
|
||||
- Env Var: RCLONE_YANDEX_CLIENT_ID
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
#### --yandex-client-secret
|
||||
|
||||
Yandex Client Secret
|
||||
Leave blank normally.
|
||||
|
||||
- Config: client_secret
|
||||
- Env Var: RCLONE_YANDEX_CLIENT_SECRET
|
||||
- Type: string
|
||||
- Default: ""
|
||||
|
||||
<!--- autogenerated options stop -->
|
||||
|
Loading…
Reference in New Issue
Block a user