2022-10-13 14:04:56 +02:00
|
|
|
|
---
|
|
|
|
|
title: "Oracle Object Storage"
|
|
|
|
|
description: "Rclone docs for Oracle Object Storage"
|
2022-11-17 18:56:24 +01:00
|
|
|
|
versionIntroduced: "v1.60.0"
|
2022-10-13 14:04:56 +02:00
|
|
|
|
---
|
|
|
|
|
|
2022-10-20 21:24:19 +02:00
|
|
|
|
# {{< icon "fa fa-cloud" >}} Oracle Object Storage
|
2022-10-13 14:04:56 +02:00
|
|
|
|
|
|
|
|
|
[Oracle Object Storage Overview](https://docs.oracle.com/en-us/iaas/Content/Object/Concepts/objectstorageoverview.htm)
|
|
|
|
|
|
|
|
|
|
[Oracle Object Storage FAQ](https://www.oracle.com/cloud/storage/object-storage/faq/)
|
|
|
|
|
|
|
|
|
|
Paths are specified as `remote:bucket` (or `remote:` for the `lsd`
|
|
|
|
|
command.) You may put subdirectories in too, e.g. `remote:bucket/path/to/dir`.
|
|
|
|
|
|
|
|
|
|
## Configuration
|
|
|
|
|
|
|
|
|
|
Here is an example of making an oracle object storage configuration. `rclone config` walks you
|
|
|
|
|
through it.
|
|
|
|
|
|
|
|
|
|
Here is an example of how to make a remote called `remote`. First run:
|
|
|
|
|
|
|
|
|
|
rclone config
|
|
|
|
|
|
|
|
|
|
This will guide you through an interactive setup process:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
```
|
|
|
|
|
n) New remote
|
|
|
|
|
d) Delete remote
|
|
|
|
|
r) Rename remote
|
|
|
|
|
c) Copy remote
|
|
|
|
|
s) Set configuration password
|
|
|
|
|
q) Quit config
|
|
|
|
|
e/n/d/r/c/s/q> n
|
|
|
|
|
|
|
|
|
|
Enter name for new remote.
|
|
|
|
|
name> remote
|
|
|
|
|
|
|
|
|
|
Option Storage.
|
|
|
|
|
Type of storage to configure.
|
|
|
|
|
Choose a number from below, or type in your own value.
|
|
|
|
|
[snip]
|
|
|
|
|
XX / Oracle Cloud Infrastructure Object Storage
|
|
|
|
|
\ (oracleobjectstorage)
|
|
|
|
|
Storage> oracleobjectstorage
|
|
|
|
|
|
|
|
|
|
Option provider.
|
|
|
|
|
Choose your Auth Provider
|
|
|
|
|
Choose a number from below, or type in your own string value.
|
|
|
|
|
Press Enter for the default (env_auth).
|
|
|
|
|
1 / automatically pickup the credentials from runtime(env), first one to provide auth wins
|
|
|
|
|
\ (env_auth)
|
|
|
|
|
/ use an OCI user and an API key for authentication.
|
|
|
|
|
2 | you’ll need to put in a config file your tenancy OCID, user OCID, region, the path, fingerprint to an API key.
|
|
|
|
|
| https://docs.oracle.com/en-us/iaas/Content/API/Concepts/sdkconfig.htm
|
|
|
|
|
\ (user_principal_auth)
|
|
|
|
|
/ use instance principals to authorize an instance to make API calls.
|
|
|
|
|
3 | each instance has its own identity, and authenticates using the certificates that are read from instance metadata.
|
|
|
|
|
| https://docs.oracle.com/en-us/iaas/Content/Identity/Tasks/callingservicesfrominstances.htm
|
|
|
|
|
\ (instance_principal_auth)
|
|
|
|
|
4 / use resource principals to make API calls
|
|
|
|
|
\ (resource_principal_auth)
|
|
|
|
|
5 / no credentials needed, this is typically for reading public buckets
|
|
|
|
|
\ (no_auth)
|
|
|
|
|
provider> 2
|
|
|
|
|
|
|
|
|
|
Option namespace.
|
|
|
|
|
Object storage namespace
|
|
|
|
|
Enter a value.
|
|
|
|
|
namespace> idbamagbg734
|
|
|
|
|
|
|
|
|
|
Option compartment.
|
|
|
|
|
Object storage compartment OCID
|
|
|
|
|
Enter a value.
|
|
|
|
|
compartment> ocid1.compartment.oc1..aaaaaaaapufkxc7ame3sthry5i7ujrwfc7ejnthhu6bhanm5oqfjpyasjkba
|
|
|
|
|
|
|
|
|
|
Option region.
|
|
|
|
|
Object storage Region
|
|
|
|
|
Enter a value.
|
|
|
|
|
region> us-ashburn-1
|
|
|
|
|
|
|
|
|
|
Option endpoint.
|
|
|
|
|
Endpoint for Object storage API.
|
|
|
|
|
Leave blank to use the default endpoint for the region.
|
|
|
|
|
Enter a value. Press Enter to leave empty.
|
|
|
|
|
endpoint>
|
|
|
|
|
|
|
|
|
|
Option config_file.
|
|
|
|
|
Path to OCI config file
|
|
|
|
|
Choose a number from below, or type in your own string value.
|
|
|
|
|
Press Enter for the default (~/.oci/config).
|
|
|
|
|
1 / oci configuration file location
|
|
|
|
|
\ (~/.oci/config)
|
|
|
|
|
config_file> /etc/oci/dev.conf
|
|
|
|
|
|
|
|
|
|
Option config_profile.
|
|
|
|
|
Profile name inside OCI config file
|
|
|
|
|
Choose a number from below, or type in your own string value.
|
|
|
|
|
Press Enter for the default (Default).
|
|
|
|
|
1 / Use the default profile
|
|
|
|
|
\ (Default)
|
|
|
|
|
config_profile> Test
|
|
|
|
|
|
|
|
|
|
Edit advanced config?
|
|
|
|
|
y) Yes
|
|
|
|
|
n) No (default)
|
|
|
|
|
y/n> n
|
|
|
|
|
|
|
|
|
|
Configuration complete.
|
|
|
|
|
Options:
|
|
|
|
|
- type: oracleobjectstorage
|
|
|
|
|
- namespace: idbamagbg734
|
|
|
|
|
- compartment: ocid1.compartment.oc1..aaaaaaaapufkxc7ame3sthry5i7ujrwfc7ejnthhu6bhanm5oqfjpyasjkba
|
|
|
|
|
- region: us-ashburn-1
|
|
|
|
|
- provider: user_principal_auth
|
2022-10-20 21:24:19 +02:00
|
|
|
|
- config_file: /etc/oci/dev.conf
|
|
|
|
|
- config_profile: Test
|
2022-10-13 14:04:56 +02:00
|
|
|
|
Keep this "remote" remote?
|
|
|
|
|
y) Yes this is OK (default)
|
|
|
|
|
e) Edit this remote
|
|
|
|
|
d) Delete this remote
|
|
|
|
|
y/e/d> y
|
|
|
|
|
```
|
|
|
|
|
|
|
|
|
|
See all buckets
|
|
|
|
|
|
|
|
|
|
rclone lsd remote:
|
|
|
|
|
|
|
|
|
|
Create a new bucket
|
|
|
|
|
|
|
|
|
|
rclone mkdir remote:bucket
|
|
|
|
|
|
|
|
|
|
List the contents of a bucket
|
|
|
|
|
|
|
|
|
|
rclone ls remote:bucket
|
|
|
|
|
rclone ls remote:bucket --max-depth 1
|
|
|
|
|
|
|
|
|
|
### Modified time
|
|
|
|
|
|
|
|
|
|
The modified time is stored as metadata on the object as
|
|
|
|
|
`opc-meta-mtime` as floating point since the epoch, accurate to 1 ns.
|
|
|
|
|
|
|
|
|
|
If the modification time needs to be updated rclone will attempt to perform a server
|
|
|
|
|
side copy to update the modification if the object can be copied in a single part.
|
|
|
|
|
In the case the object is larger than 5Gb, the object will be uploaded rather than copied.
|
|
|
|
|
|
|
|
|
|
Note that reading this from the object takes an additional `HEAD` request as the metadata
|
|
|
|
|
isn't returned in object listings.
|
|
|
|
|
|
|
|
|
|
### Multipart uploads
|
|
|
|
|
|
|
|
|
|
rclone supports multipart uploads with OOS which means that it can
|
|
|
|
|
upload files bigger than 5 GiB.
|
|
|
|
|
|
|
|
|
|
Note that files uploaded *both* with multipart upload *and* through
|
|
|
|
|
crypt remotes do not have MD5 sums.
|
|
|
|
|
|
|
|
|
|
rclone switches from single part uploads to multipart uploads at the
|
|
|
|
|
point specified by `--oos-upload-cutoff`. This can be a maximum of 5 GiB
|
|
|
|
|
and a minimum of 0 (ie always upload multipart files).
|
|
|
|
|
|
|
|
|
|
The chunk sizes used in the multipart upload are specified by
|
|
|
|
|
`--oos-chunk-size` and the number of chunks uploaded concurrently is
|
|
|
|
|
specified by `--oos-upload-concurrency`.
|
|
|
|
|
|
|
|
|
|
Multipart uploads will use `--transfers` * `--oos-upload-concurrency` *
|
|
|
|
|
`--oos-chunk-size` extra memory. Single part uploads to not use extra
|
|
|
|
|
memory.
|
|
|
|
|
|
|
|
|
|
Single part transfers can be faster than multipart transfers or slower
|
|
|
|
|
depending on your latency from oos - the more latency, the more likely
|
|
|
|
|
single part transfers will be faster.
|
|
|
|
|
|
|
|
|
|
Increasing `--oos-upload-concurrency` will increase throughput (8 would
|
|
|
|
|
be a sensible value) and increasing `--oos-chunk-size` also increases
|
|
|
|
|
throughput (16M would be sensible). Increasing either of these will
|
|
|
|
|
use more memory. The default values are high enough to gain most of
|
|
|
|
|
the possible performance without using too much memory.
|
|
|
|
|
|
|
|
|
|
{{< rem autogenerated options start" - DO NOT EDIT - instead edit fs.RegInfo in backend/oracleobjectstorage/oracleobjectstorage.go then run make backenddocs" >}}
|
|
|
|
|
### Standard options
|
|
|
|
|
|
|
|
|
|
Here are the Standard options specific to oracleobjectstorage (Oracle Cloud Infrastructure Object Storage).
|
|
|
|
|
|
|
|
|
|
#### --oos-provider
|
|
|
|
|
|
|
|
|
|
Choose your Auth Provider
|
|
|
|
|
|
|
|
|
|
Properties:
|
|
|
|
|
|
|
|
|
|
- Config: provider
|
|
|
|
|
- Env Var: RCLONE_OOS_PROVIDER
|
|
|
|
|
- Type: string
|
|
|
|
|
- Default: "env_auth"
|
|
|
|
|
- Examples:
|
|
|
|
|
- "env_auth"
|
|
|
|
|
- automatically pickup the credentials from runtime(env), first one to provide auth wins
|
|
|
|
|
- "user_principal_auth"
|
|
|
|
|
- use an OCI user and an API key for authentication.
|
|
|
|
|
- you’ll need to put in a config file your tenancy OCID, user OCID, region, the path, fingerprint to an API key.
|
|
|
|
|
- https://docs.oracle.com/en-us/iaas/Content/API/Concepts/sdkconfig.htm
|
|
|
|
|
- "instance_principal_auth"
|
|
|
|
|
- use instance principals to authorize an instance to make API calls.
|
|
|
|
|
- each instance has its own identity, and authenticates using the certificates that are read from instance metadata.
|
|
|
|
|
- https://docs.oracle.com/en-us/iaas/Content/Identity/Tasks/callingservicesfrominstances.htm
|
|
|
|
|
- "resource_principal_auth"
|
|
|
|
|
- use resource principals to make API calls
|
|
|
|
|
- "no_auth"
|
|
|
|
|
- no credentials needed, this is typically for reading public buckets
|
|
|
|
|
|
|
|
|
|
#### --oos-namespace
|
|
|
|
|
|
|
|
|
|
Object storage namespace
|
|
|
|
|
|
|
|
|
|
Properties:
|
|
|
|
|
|
|
|
|
|
- Config: namespace
|
|
|
|
|
- Env Var: RCLONE_OOS_NAMESPACE
|
|
|
|
|
- Type: string
|
|
|
|
|
- Required: true
|
|
|
|
|
|
|
|
|
|
#### --oos-compartment
|
|
|
|
|
|
|
|
|
|
Object storage compartment OCID
|
|
|
|
|
|
|
|
|
|
Properties:
|
|
|
|
|
|
|
|
|
|
- Config: compartment
|
|
|
|
|
- Env Var: RCLONE_OOS_COMPARTMENT
|
|
|
|
|
- Provider: !no_auth
|
|
|
|
|
- Type: string
|
|
|
|
|
- Required: true
|
|
|
|
|
|
|
|
|
|
#### --oos-region
|
|
|
|
|
|
|
|
|
|
Object storage Region
|
|
|
|
|
|
|
|
|
|
Properties:
|
|
|
|
|
|
|
|
|
|
- Config: region
|
|
|
|
|
- Env Var: RCLONE_OOS_REGION
|
|
|
|
|
- Type: string
|
|
|
|
|
- Required: true
|
|
|
|
|
|
|
|
|
|
#### --oos-endpoint
|
|
|
|
|
|
|
|
|
|
Endpoint for Object storage API.
|
|
|
|
|
|
|
|
|
|
Leave blank to use the default endpoint for the region.
|
|
|
|
|
|
|
|
|
|
Properties:
|
|
|
|
|
|
|
|
|
|
- Config: endpoint
|
|
|
|
|
- Env Var: RCLONE_OOS_ENDPOINT
|
|
|
|
|
- Type: string
|
|
|
|
|
- Required: false
|
|
|
|
|
|
|
|
|
|
#### --oos-config-file
|
|
|
|
|
|
|
|
|
|
Path to OCI config file
|
|
|
|
|
|
|
|
|
|
Properties:
|
|
|
|
|
|
|
|
|
|
- Config: config_file
|
|
|
|
|
- Env Var: RCLONE_OOS_CONFIG_FILE
|
|
|
|
|
- Provider: user_principal_auth
|
|
|
|
|
- Type: string
|
|
|
|
|
- Default: "~/.oci/config"
|
|
|
|
|
- Examples:
|
|
|
|
|
- "~/.oci/config"
|
|
|
|
|
- oci configuration file location
|
|
|
|
|
|
|
|
|
|
#### --oos-config-profile
|
|
|
|
|
|
2022-10-17 08:19:23 +02:00
|
|
|
|
Profile name inside the oci config file
|
2022-10-13 14:04:56 +02:00
|
|
|
|
|
|
|
|
|
Properties:
|
|
|
|
|
|
|
|
|
|
- Config: config_profile
|
|
|
|
|
- Env Var: RCLONE_OOS_CONFIG_PROFILE
|
|
|
|
|
- Provider: user_principal_auth
|
|
|
|
|
- Type: string
|
|
|
|
|
- Default: "Default"
|
|
|
|
|
- Examples:
|
|
|
|
|
- "Default"
|
|
|
|
|
- Use the default profile
|
|
|
|
|
|
|
|
|
|
### Advanced options
|
|
|
|
|
|
|
|
|
|
Here are the Advanced options specific to oracleobjectstorage (Oracle Cloud Infrastructure Object Storage).
|
|
|
|
|
|
|
|
|
|
#### --oos-upload-cutoff
|
|
|
|
|
|
|
|
|
|
Cutoff for switching to chunked upload.
|
|
|
|
|
|
|
|
|
|
Any files larger than this will be uploaded in chunks of chunk_size.
|
|
|
|
|
The minimum is 0 and the maximum is 5 GiB.
|
|
|
|
|
|
|
|
|
|
Properties:
|
|
|
|
|
|
|
|
|
|
- Config: upload_cutoff
|
|
|
|
|
- Env Var: RCLONE_OOS_UPLOAD_CUTOFF
|
|
|
|
|
- Type: SizeSuffix
|
|
|
|
|
- Default: 200Mi
|
|
|
|
|
|
|
|
|
|
#### --oos-chunk-size
|
|
|
|
|
|
|
|
|
|
Chunk size to use for uploading.
|
|
|
|
|
|
|
|
|
|
When uploading files larger than upload_cutoff or files with unknown
|
|
|
|
|
size (e.g. from "rclone rcat" or uploaded with "rclone mount" or google
|
|
|
|
|
photos or google docs) they will be uploaded as multipart uploads
|
|
|
|
|
using this chunk size.
|
|
|
|
|
|
|
|
|
|
Note that "upload_concurrency" chunks of this size are buffered
|
|
|
|
|
in memory per transfer.
|
|
|
|
|
|
|
|
|
|
If you are transferring large files over high-speed links and you have
|
|
|
|
|
enough memory, then increasing this will speed up the transfers.
|
|
|
|
|
|
|
|
|
|
Rclone will automatically increase the chunk size when uploading a
|
|
|
|
|
large file of known size to stay below the 10,000 chunks limit.
|
|
|
|
|
|
|
|
|
|
Files of unknown size are uploaded with the configured
|
|
|
|
|
chunk_size. Since the default chunk size is 5 MiB and there can be at
|
|
|
|
|
most 10,000 chunks, this means that by default the maximum size of
|
|
|
|
|
a file you can stream upload is 48 GiB. If you wish to stream upload
|
|
|
|
|
larger files then you will need to increase chunk_size.
|
|
|
|
|
|
|
|
|
|
Increasing the chunk size decreases the accuracy of the progress
|
|
|
|
|
statistics displayed with "-P" flag.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Properties:
|
|
|
|
|
|
|
|
|
|
- Config: chunk_size
|
|
|
|
|
- Env Var: RCLONE_OOS_CHUNK_SIZE
|
|
|
|
|
- Type: SizeSuffix
|
|
|
|
|
- Default: 5Mi
|
|
|
|
|
|
|
|
|
|
#### --oos-upload-concurrency
|
|
|
|
|
|
|
|
|
|
Concurrency for multipart uploads.
|
|
|
|
|
|
|
|
|
|
This is the number of chunks of the same file that are uploaded
|
|
|
|
|
concurrently.
|
|
|
|
|
|
|
|
|
|
If you are uploading small numbers of large files over high-speed links
|
|
|
|
|
and these uploads do not fully utilize your bandwidth, then increasing
|
|
|
|
|
this may help to speed up the transfers.
|
|
|
|
|
|
|
|
|
|
Properties:
|
|
|
|
|
|
|
|
|
|
- Config: upload_concurrency
|
|
|
|
|
- Env Var: RCLONE_OOS_UPLOAD_CONCURRENCY
|
|
|
|
|
- Type: int
|
|
|
|
|
- Default: 10
|
|
|
|
|
|
|
|
|
|
#### --oos-copy-cutoff
|
|
|
|
|
|
|
|
|
|
Cutoff for switching to multipart copy.
|
|
|
|
|
|
|
|
|
|
Any files larger than this that need to be server-side copied will be
|
|
|
|
|
copied in chunks of this size.
|
|
|
|
|
|
|
|
|
|
The minimum is 0 and the maximum is 5 GiB.
|
|
|
|
|
|
|
|
|
|
Properties:
|
|
|
|
|
|
|
|
|
|
- Config: copy_cutoff
|
|
|
|
|
- Env Var: RCLONE_OOS_COPY_CUTOFF
|
|
|
|
|
- Type: SizeSuffix
|
|
|
|
|
- Default: 4.656Gi
|
|
|
|
|
|
|
|
|
|
#### --oos-copy-timeout
|
|
|
|
|
|
|
|
|
|
Timeout for copy.
|
|
|
|
|
|
|
|
|
|
Copy is an asynchronous operation, specify timeout to wait for copy to succeed
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Properties:
|
|
|
|
|
|
|
|
|
|
- Config: copy_timeout
|
|
|
|
|
- Env Var: RCLONE_OOS_COPY_TIMEOUT
|
|
|
|
|
- Type: Duration
|
|
|
|
|
- Default: 1m0s
|
|
|
|
|
|
|
|
|
|
#### --oos-disable-checksum
|
|
|
|
|
|
|
|
|
|
Don't store MD5 checksum with object metadata.
|
|
|
|
|
|
|
|
|
|
Normally rclone will calculate the MD5 checksum of the input before
|
|
|
|
|
uploading it so it can add it to metadata on the object. This is great
|
|
|
|
|
for data integrity checking but can cause long delays for large files
|
|
|
|
|
to start uploading.
|
|
|
|
|
|
|
|
|
|
Properties:
|
|
|
|
|
|
|
|
|
|
- Config: disable_checksum
|
|
|
|
|
- Env Var: RCLONE_OOS_DISABLE_CHECKSUM
|
|
|
|
|
- Type: bool
|
|
|
|
|
- Default: false
|
|
|
|
|
|
|
|
|
|
#### --oos-encoding
|
|
|
|
|
|
|
|
|
|
The encoding for the backend.
|
|
|
|
|
|
|
|
|
|
See the [encoding section in the overview](/overview/#encoding) for more info.
|
|
|
|
|
|
|
|
|
|
Properties:
|
|
|
|
|
|
|
|
|
|
- Config: encoding
|
|
|
|
|
- Env Var: RCLONE_OOS_ENCODING
|
|
|
|
|
- Type: MultiEncoder
|
|
|
|
|
- Default: Slash,InvalidUtf8,Dot
|
|
|
|
|
|
|
|
|
|
#### --oos-leave-parts-on-error
|
|
|
|
|
|
|
|
|
|
If true avoid calling abort upload on a failure, leaving all successfully uploaded parts on S3 for manual recovery.
|
|
|
|
|
|
|
|
|
|
It should be set to true for resuming uploads across different sessions.
|
|
|
|
|
|
|
|
|
|
WARNING: Storing parts of an incomplete multipart upload counts towards space usage on object storage and will add
|
|
|
|
|
additional costs if not cleaned up.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Properties:
|
|
|
|
|
|
|
|
|
|
- Config: leave_parts_on_error
|
|
|
|
|
- Env Var: RCLONE_OOS_LEAVE_PARTS_ON_ERROR
|
|
|
|
|
- Type: bool
|
|
|
|
|
- Default: false
|
|
|
|
|
|
2022-10-17 08:19:23 +02:00
|
|
|
|
#### --oos-no-check-bucket
|
|
|
|
|
|
|
|
|
|
If set, don't attempt to check the bucket exists or create it.
|
|
|
|
|
|
|
|
|
|
This can be useful when trying to minimise the number of transactions
|
|
|
|
|
rclone does if you know the bucket exists already.
|
|
|
|
|
|
|
|
|
|
It can also be needed if the user you are using does not have bucket
|
|
|
|
|
creation permissions.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Properties:
|
|
|
|
|
|
|
|
|
|
- Config: no_check_bucket
|
|
|
|
|
- Env Var: RCLONE_OOS_NO_CHECK_BUCKET
|
|
|
|
|
- Type: bool
|
|
|
|
|
- Default: false
|
|
|
|
|
|
2022-10-13 14:04:56 +02:00
|
|
|
|
## Backend commands
|
|
|
|
|
|
|
|
|
|
Here are the commands specific to the oracleobjectstorage backend.
|
|
|
|
|
|
|
|
|
|
Run them with
|
|
|
|
|
|
|
|
|
|
rclone backend COMMAND remote:
|
|
|
|
|
|
|
|
|
|
The help below will explain what arguments each command takes.
|
|
|
|
|
|
|
|
|
|
See the [backend](/commands/rclone_backend/) command for more
|
|
|
|
|
info on how to pass options and arguments.
|
|
|
|
|
|
|
|
|
|
These can be run on a running backend using the rc command
|
|
|
|
|
[backend/command](/rc/#backend-command).
|
|
|
|
|
|
|
|
|
|
### rename
|
|
|
|
|
|
|
|
|
|
change the name of an object
|
|
|
|
|
|
|
|
|
|
rclone backend rename remote: [options] [<arguments>+]
|
|
|
|
|
|
|
|
|
|
This command can be used to rename a object.
|
|
|
|
|
|
|
|
|
|
Usage Examples:
|
|
|
|
|
|
|
|
|
|
rclone backend rename oos:bucket relative-object-path-under-bucket object-new-name
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
### list-multipart-uploads
|
|
|
|
|
|
|
|
|
|
List the unfinished multipart uploads
|
|
|
|
|
|
|
|
|
|
rclone backend list-multipart-uploads remote: [options] [<arguments>+]
|
|
|
|
|
|
|
|
|
|
This command lists the unfinished multipart uploads in JSON format.
|
|
|
|
|
|
|
|
|
|
rclone backend list-multipart-uploads oos:bucket/path/to/object
|
|
|
|
|
|
|
|
|
|
It returns a dictionary of buckets with values as lists of unfinished
|
|
|
|
|
multipart uploads.
|
|
|
|
|
|
|
|
|
|
You can call it with no bucket in which case it lists all bucket, with
|
|
|
|
|
a bucket or with a bucket and path.
|
|
|
|
|
|
|
|
|
|
{
|
|
|
|
|
"test-bucket": [
|
|
|
|
|
{
|
|
|
|
|
"namespace": "test-namespace",
|
|
|
|
|
"bucket": "test-bucket",
|
|
|
|
|
"object": "600m.bin",
|
|
|
|
|
"uploadId": "51dd8114-52a4-b2f2-c42f-5291f05eb3c8",
|
|
|
|
|
"timeCreated": "2022-07-29T06:21:16.595Z",
|
|
|
|
|
"storageTier": "Standard"
|
|
|
|
|
}
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
### cleanup
|
|
|
|
|
|
|
|
|
|
Remove unfinished multipart uploads.
|
|
|
|
|
|
|
|
|
|
rclone backend cleanup remote: [options] [<arguments>+]
|
|
|
|
|
|
|
|
|
|
This command removes unfinished multipart uploads of age greater than
|
|
|
|
|
max-age which defaults to 24 hours.
|
|
|
|
|
|
|
|
|
|
Note that you can use -i/--dry-run with this command to see what it
|
|
|
|
|
would do.
|
|
|
|
|
|
|
|
|
|
rclone backend cleanup oos:bucket/path/to/object
|
|
|
|
|
rclone backend cleanup -o max-age=7w oos:bucket/path/to/object
|
|
|
|
|
|
|
|
|
|
Durations are parsed as per the rest of rclone, 2h, 7d, 7w etc.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Options:
|
|
|
|
|
|
|
|
|
|
- "max-age": Max age of upload to delete
|
|
|
|
|
|
|
|
|
|
{{< rem autogenerated options stop >}}
|