2015-09-03 00:37:42 +02:00
|
|
|
---
|
|
|
|
title: "Amazon Cloud Drive"
|
|
|
|
description: "Rclone docs for Amazon Cloud Drive"
|
|
|
|
date: "2015-09-06"
|
|
|
|
---
|
|
|
|
|
2015-10-15 17:57:21 +02:00
|
|
|
<i class="fa fa-amazon"></i> Amazon Cloud Drive
|
2015-09-03 00:37:42 +02:00
|
|
|
-----------------------------------------
|
|
|
|
|
|
|
|
Paths are specified as `remote:path`
|
|
|
|
|
|
|
|
Paths may be as deep as required, eg `remote:directory/subdirectory`.
|
|
|
|
|
|
|
|
The initial setup for Amazon cloud drive involves getting a token from
|
|
|
|
Amazon which you need to do in your browser. `rclone config` walks
|
|
|
|
you through it.
|
|
|
|
|
|
|
|
Here is an example of how to make a remote called `remote`. First run:
|
|
|
|
|
|
|
|
rclone config
|
|
|
|
|
|
|
|
This will guide you through an interactive setup process:
|
|
|
|
|
|
|
|
```
|
|
|
|
n) New remote
|
|
|
|
d) Delete remote
|
|
|
|
q) Quit config
|
|
|
|
e/n/d/q> n
|
|
|
|
name> remote
|
2016-02-21 14:39:04 +01:00
|
|
|
Type of storage to configure.
|
|
|
|
Choose a number from below, or type in your own value
|
|
|
|
1 / Amazon Cloud Drive
|
|
|
|
\ "amazon cloud drive"
|
|
|
|
2 / Amazon S3 (also Dreamhost, Ceph)
|
|
|
|
\ "s3"
|
|
|
|
3 / Backblaze B2
|
|
|
|
\ "b2"
|
|
|
|
4 / Dropbox
|
|
|
|
\ "dropbox"
|
|
|
|
5 / Google Cloud Storage (this is not Google Drive)
|
|
|
|
\ "google cloud storage"
|
|
|
|
6 / Google Drive
|
|
|
|
\ "drive"
|
|
|
|
7 / Hubic
|
|
|
|
\ "hubic"
|
|
|
|
8 / Local Disk
|
|
|
|
\ "local"
|
|
|
|
9 / Microsoft OneDrive
|
|
|
|
\ "onedrive"
|
|
|
|
10 / Openstack Swift (Rackspace Cloud Files, Memset Memstore, OVH)
|
|
|
|
\ "swift"
|
|
|
|
11 / Yandex Disk
|
|
|
|
\ "yandex"
|
|
|
|
Storage> 1
|
2015-10-03 15:23:12 +02:00
|
|
|
Amazon Application Client Id - leave blank normally.
|
2015-09-03 00:37:42 +02:00
|
|
|
client_id>
|
2015-10-03 15:23:12 +02:00
|
|
|
Amazon Application Client Secret - leave blank normally.
|
2015-09-03 00:37:42 +02:00
|
|
|
client_secret>
|
|
|
|
Remote config
|
2015-09-11 14:26:51 +02:00
|
|
|
If your browser doesn't open automatically go to the following link: http://127.0.0.1:53682/auth
|
|
|
|
Log in and authorize rclone for access
|
|
|
|
Waiting for code...
|
|
|
|
Got code
|
2015-09-03 00:37:42 +02:00
|
|
|
--------------------
|
|
|
|
[remote]
|
|
|
|
client_id =
|
|
|
|
client_secret =
|
|
|
|
token = {"access_token":"xxxxxxxxxxxxxxxxxxxxxxx","token_type":"bearer","refresh_token":"xxxxxxxxxxxxxxxxxx","expiry":"2015-09-06T16:07:39.658438471+01:00"}
|
|
|
|
--------------------
|
|
|
|
y) Yes this is OK
|
|
|
|
e) Edit this remote
|
|
|
|
d) Delete this remote
|
|
|
|
y/e/d> y
|
|
|
|
```
|
|
|
|
|
2016-01-07 16:20:32 +01:00
|
|
|
See the [remote setup docs](/remote_setup/) for how to set it up on a
|
|
|
|
machine with no Internet browser available.
|
|
|
|
|
2015-09-03 00:37:42 +02:00
|
|
|
Note that rclone runs a webserver on your local machine to collect the
|
2015-09-11 14:26:51 +02:00
|
|
|
token as returned from Amazon. This only runs from the moment it
|
|
|
|
opens your browser to the moment you get back the verification
|
2015-09-03 00:37:42 +02:00
|
|
|
code. This is on `http://127.0.0.1:53682/` and this it may require
|
|
|
|
you to unblock it temporarily if you are running a host firewall.
|
|
|
|
|
|
|
|
Once configured you can then use `rclone` like this,
|
|
|
|
|
|
|
|
List directories in top level of your Amazon cloud drive
|
|
|
|
|
|
|
|
rclone lsd remote:
|
|
|
|
|
|
|
|
List all the files in your Amazon cloud drive
|
|
|
|
|
|
|
|
rclone ls remote:
|
|
|
|
|
|
|
|
To copy a local directory to an Amazon cloud drive directory called backup
|
|
|
|
|
|
|
|
rclone copy /home/source remote:backup
|
|
|
|
|
|
|
|
### Modified time and MD5SUMs ###
|
|
|
|
|
|
|
|
Amazon cloud drive doesn't allow modification times to be changed via
|
|
|
|
the API so these won't be accurate or used for syncing.
|
|
|
|
|
|
|
|
It does store MD5SUMs so for a more accurate sync, you can use the
|
|
|
|
`--checksum` flag.
|
|
|
|
|
|
|
|
### Deleting files ###
|
|
|
|
|
|
|
|
Any files you delete with rclone will end up in the trash. Amazon
|
|
|
|
don't provide an API to permanently delete files, nor to empty the
|
|
|
|
trash, so you will have to do that with one of Amazon's apps or via
|
|
|
|
the Amazon cloud drive website.
|
|
|
|
|
2016-01-30 19:08:44 +01:00
|
|
|
### Specific options ###
|
|
|
|
|
|
|
|
Here are the command line options specific to this cloud storage
|
|
|
|
system.
|
|
|
|
|
|
|
|
#### --acd-templink-threshold=SIZE ####
|
|
|
|
|
|
|
|
Files this size or more will be downloaded via their `tempLink`. This
|
|
|
|
is to work around a problem with Amazon Cloud Drive which blocks
|
|
|
|
downloads of files bigger than about 10GB. The default for this is
|
|
|
|
9GB which shouldn't need to be changed.
|
|
|
|
|
|
|
|
To download files above this threshold, rclone requests a `tempLink`
|
|
|
|
which downloads the file through a temporary URL directly from the
|
|
|
|
underlying S3 storage.
|
|
|
|
|
2015-09-03 00:37:42 +02:00
|
|
|
### Limitations ###
|
|
|
|
|
2015-11-14 12:20:04 +01:00
|
|
|
Note that Amazon cloud drive is case insensitive so you can't have a
|
2015-09-03 00:37:42 +02:00
|
|
|
file called "Hello.doc" and one called "hello.doc".
|
|
|
|
|
|
|
|
Amazon cloud drive has rate limiting so you may notice errors in the
|
|
|
|
sync (429 errors). rclone will automatically retry the sync up to 3
|
|
|
|
times by default (see `--retries` flag) which should hopefully work
|
|
|
|
around this problem.
|
2016-01-04 13:23:33 +01:00
|
|
|
|
|
|
|
Amazon cloud drive has an internal limit of file sizes that can be
|
|
|
|
uploaded to the service. This limit is not officially published,
|
|
|
|
but all files larger than this will fail.
|
|
|
|
|
|
|
|
At the time of writing (Jan 2016) is in the area of 50GB per file.
|
|
|
|
This means that larger files are likely to fail.
|
|
|
|
|
|
|
|
Unfortunatly there is no way for rclone to see that this failure is
|
|
|
|
because of file size, so it will retry the operation, as any other
|
|
|
|
failure. To avoid this problem, use `--max-size=50GB` option to limit
|
|
|
|
the maximum size of uploaded files.
|