mirror of
https://github.com/rclone/rclone.git
synced 2024-11-07 09:04:52 +01:00
docs: fix typos found by codespell in docs and code comments
This commit is contained in:
parent
50b4a2398e
commit
3d473eb54e
@ -2119,7 +2119,7 @@ func (w *azChunkWriter) WriteChunk(ctx context.Context, chunkNumber int, reader
|
|||||||
return currentChunkSize, err
|
return currentChunkSize, err
|
||||||
}
|
}
|
||||||
|
|
||||||
// Abort the multpart upload.
|
// Abort the multipart upload.
|
||||||
//
|
//
|
||||||
// FIXME it would be nice to delete uncommitted blocks.
|
// FIXME it would be nice to delete uncommitted blocks.
|
||||||
//
|
//
|
||||||
|
@ -154,7 +154,7 @@ func init() {
|
|||||||
Default: "",
|
Default: "",
|
||||||
Help: `Impersonate this user ID when using a service account.
|
Help: `Impersonate this user ID when using a service account.
|
||||||
|
|
||||||
Settng this flag allows rclone, when using a JWT service account, to
|
Setting this flag allows rclone, when using a JWT service account, to
|
||||||
act on behalf of another user by setting the as-user header.
|
act on behalf of another user by setting the as-user header.
|
||||||
|
|
||||||
The user ID is the Box identifier for a user. User IDs can found for
|
The user ID is the Box identifier for a user. User IDs can found for
|
||||||
@ -1212,7 +1212,7 @@ func (f *Fs) ChangeNotify(ctx context.Context, notifyFunc func(string, fs.EntryT
|
|||||||
}
|
}
|
||||||
|
|
||||||
// box can send duplicate Event IDs. Use this map to track and filter
|
// box can send duplicate Event IDs. Use this map to track and filter
|
||||||
// the ones we've alread processed.
|
// the ones we've already processed.
|
||||||
processedEventIDs := make(map[string]time.Time)
|
processedEventIDs := make(map[string]time.Time)
|
||||||
|
|
||||||
var ticker *time.Ticker
|
var ticker *time.Ticker
|
||||||
|
@ -206,7 +206,7 @@ func NewFs(ctx context.Context, name, root string, m configmap.Mapper) (fs.Fs, e
|
|||||||
}
|
}
|
||||||
ci := fs.GetConfig(ctx)
|
ci := fs.GetConfig(ctx)
|
||||||
|
|
||||||
// cache *mega.Mega on username so we can re-use and share
|
// cache *mega.Mega on username so we can reuse and share
|
||||||
// them between remotes. They are expensive to make as they
|
// them between remotes. They are expensive to make as they
|
||||||
// contain all the objects and sharing the objects makes the
|
// contain all the objects and sharing the objects makes the
|
||||||
// move code easier as we don't have to worry about mixing
|
// move code easier as we don't have to worry about mixing
|
||||||
|
@ -1215,7 +1215,7 @@ func (f *Fs) upload(ctx context.Context, in io.Reader, leaf, dirID, sha1Str stri
|
|||||||
return nil, fmt.Errorf("failed to upload: %w", err)
|
return nil, fmt.Errorf("failed to upload: %w", err)
|
||||||
}
|
}
|
||||||
// refresh uploaded file info
|
// refresh uploaded file info
|
||||||
// Compared to `newfile.File` this upgrades several feilds...
|
// Compared to `newfile.File` this upgrades several fields...
|
||||||
// audit, links, modified_time, phase, revision, and web_content_link
|
// audit, links, modified_time, phase, revision, and web_content_link
|
||||||
return f.getFile(ctx, newfile.File.ID)
|
return f.getFile(ctx, newfile.File.ID)
|
||||||
}
|
}
|
||||||
|
@ -5611,7 +5611,7 @@ func (w *s3ChunkWriter) WriteChunk(ctx context.Context, chunkNumber int, reader
|
|||||||
return currentChunkSize, err
|
return currentChunkSize, err
|
||||||
}
|
}
|
||||||
|
|
||||||
// Abort the multpart upload
|
// Abort the multipart upload
|
||||||
func (w *s3ChunkWriter) Abort(ctx context.Context) error {
|
func (w *s3ChunkWriter) Abort(ctx context.Context) error {
|
||||||
err := w.f.pacer.Call(func() (bool, error) {
|
err := w.f.pacer.Call(func() (bool, error) {
|
||||||
_, err := w.f.c.AbortMultipartUploadWithContext(context.Background(), &s3.AbortMultipartUploadInput{
|
_, err := w.f.c.AbortMultipartUploadWithContext(context.Background(), &s3.AbortMultipartUploadInput{
|
||||||
|
@ -1014,7 +1014,7 @@ func (f *Fs) keyboardInteractiveReponse(user, instruction string, questions []st
|
|||||||
// save it so on reconnection we give back the previous string.
|
// save it so on reconnection we give back the previous string.
|
||||||
// This removes the ability to let the user correct a mistaken entry,
|
// This removes the ability to let the user correct a mistaken entry,
|
||||||
// but means that reconnects are transparent.
|
// but means that reconnects are transparent.
|
||||||
// We'll re-use config.Pass for this, 'cos we know it's not been
|
// We'll reuse config.Pass for this, 'cos we know it's not been
|
||||||
// specified.
|
// specified.
|
||||||
func (f *Fs) getPass() (string, error) {
|
func (f *Fs) getPass() (string, error) {
|
||||||
for f.savedpswd == "" {
|
for f.savedpswd == "" {
|
||||||
@ -1602,7 +1602,7 @@ func (f *Fs) About(ctx context.Context) (*fs.Usage, error) {
|
|||||||
fs.Debugf(f, "About path %q", aboutPath)
|
fs.Debugf(f, "About path %q", aboutPath)
|
||||||
vfsStats, err = c.sftpClient.StatVFS(aboutPath)
|
vfsStats, err = c.sftpClient.StatVFS(aboutPath)
|
||||||
}
|
}
|
||||||
f.putSftpConnection(&c, err) // Return to pool asap, if running shell command below it will be re-used
|
f.putSftpConnection(&c, err) // Return to pool asap, if running shell command below it will be reused
|
||||||
if vfsStats != nil {
|
if vfsStats != nil {
|
||||||
total := vfsStats.TotalSpace()
|
total := vfsStats.TotalSpace()
|
||||||
free := vfsStats.FreeSpace()
|
free := vfsStats.FreeSpace()
|
||||||
@ -2044,7 +2044,7 @@ func (o *Object) Update(ctx context.Context, in io.Reader, src fs.ObjectInfo, op
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("Update: %w", err)
|
return fmt.Errorf("Update: %w", err)
|
||||||
}
|
}
|
||||||
// Hang on to the connection for the whole upload so it doesn't get re-used while we are uploading
|
// Hang on to the connection for the whole upload so it doesn't get reused while we are uploading
|
||||||
file, err := c.sftpClient.OpenFile(o.path(), os.O_WRONLY|os.O_CREATE|os.O_TRUNC)
|
file, err := c.sftpClient.OpenFile(o.path(), os.O_WRONLY|os.O_CREATE|os.O_TRUNC)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
o.fs.putSftpConnection(&c, err)
|
o.fs.putSftpConnection(&c, err)
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
// Package cmd implements the rclone command
|
// Package cmd implements the rclone command
|
||||||
//
|
//
|
||||||
// It is in a sub package so it's internals can be re-used elsewhere
|
// It is in a sub package so it's internals can be reused elsewhere
|
||||||
package cmd
|
package cmd
|
||||||
|
|
||||||
// FIXME only attach the remote flags when using a remote???
|
// FIXME only attach the remote flags when using a remote???
|
||||||
|
@ -83,7 +83,7 @@ func mountOptions(fsys *FS, f fs.Fs, opt *mountlib.Options) (mountOpts *fuse.Mou
|
|||||||
// (128 kiB on Linux) and cannot be larger than MaxWrite.
|
// (128 kiB on Linux) and cannot be larger than MaxWrite.
|
||||||
//
|
//
|
||||||
// MaxReadAhead only affects buffered reads (=non-direct-io), but even then, the
|
// MaxReadAhead only affects buffered reads (=non-direct-io), but even then, the
|
||||||
// kernel can and does send larger reads to satisfy read reqests from applications
|
// kernel can and does send larger reads to satisfy read requests from applications
|
||||||
// (up to MaxWrite or VM_READAHEAD_PAGES=128 kiB, whichever is less).
|
// (up to MaxWrite or VM_READAHEAD_PAGES=128 kiB, whichever is less).
|
||||||
MaxReadAhead int
|
MaxReadAhead int
|
||||||
|
|
||||||
|
@ -442,7 +442,7 @@ Properties:
|
|||||||
|
|
||||||
Impersonate this user ID when using a service account.
|
Impersonate this user ID when using a service account.
|
||||||
|
|
||||||
Settng this flag allows rclone, when using a JWT service account, to
|
Setting this flag allows rclone, when using a JWT service account, to
|
||||||
act on behalf of another user by setting the as-user header.
|
act on behalf of another user by setting the as-user header.
|
||||||
|
|
||||||
The user ID is the Box identifier for a user. User IDs can found for
|
The user ID is the Box identifier for a user. User IDs can found for
|
||||||
|
@ -105,14 +105,14 @@ description: "Rclone Changelog"
|
|||||||
* Fix 425 "TLS session of data connection not resumed" errors (Nick Craig-Wood)
|
* Fix 425 "TLS session of data connection not resumed" errors (Nick Craig-Wood)
|
||||||
* Hdfs
|
* Hdfs
|
||||||
* Retry "replication in progress" errors when uploading (Nick Craig-Wood)
|
* Retry "replication in progress" errors when uploading (Nick Craig-Wood)
|
||||||
* Fix uploading to the wrong object on Update with overriden remote name (Nick Craig-Wood)
|
* Fix uploading to the wrong object on Update with overridden remote name (Nick Craig-Wood)
|
||||||
* HTTP
|
* HTTP
|
||||||
* CORS should not be sent if not set (yuudi)
|
* CORS should not be sent if not set (yuudi)
|
||||||
* Fix webdav OPTIONS response (yuudi)
|
* Fix webdav OPTIONS response (yuudi)
|
||||||
* Opendrive
|
* Opendrive
|
||||||
* Fix List on a just deleted and remade directory (Nick Craig-Wood)
|
* Fix List on a just deleted and remade directory (Nick Craig-Wood)
|
||||||
* Oracleobjectstorage
|
* Oracleobjectstorage
|
||||||
* Use rclone's rate limiter in mutipart transfers (Manoj Ghosh)
|
* Use rclone's rate limiter in multipart transfers (Manoj Ghosh)
|
||||||
* Implement `OpenChunkWriter` and multi-thread uploads (Manoj Ghosh)
|
* Implement `OpenChunkWriter` and multi-thread uploads (Manoj Ghosh)
|
||||||
* S3
|
* S3
|
||||||
* Refactor multipart upload to use `OpenChunkWriter` and `ChunkWriter` (Vitor Gomes)
|
* Refactor multipart upload to use `OpenChunkWriter` and `ChunkWriter` (Vitor Gomes)
|
||||||
@ -285,14 +285,14 @@ description: "Rclone Changelog"
|
|||||||
* Fix quickxorhash on 32 bit architectures (Nick Craig-Wood)
|
* Fix quickxorhash on 32 bit architectures (Nick Craig-Wood)
|
||||||
* Report any list errors during `rclone cleanup` (albertony)
|
* Report any list errors during `rclone cleanup` (albertony)
|
||||||
* Putio
|
* Putio
|
||||||
* Fix uploading to the wrong object on Update with overriden remote name (Nick Craig-Wood)
|
* Fix uploading to the wrong object on Update with overridden remote name (Nick Craig-Wood)
|
||||||
* Fix modification times not being preserved for server side copy and move (Nick Craig-Wood)
|
* Fix modification times not being preserved for server side copy and move (Nick Craig-Wood)
|
||||||
* Fix server side copy failures (400 errors) (Nick Craig-Wood)
|
* Fix server side copy failures (400 errors) (Nick Craig-Wood)
|
||||||
* S3
|
* S3
|
||||||
* Empty directory markers (Jānis Bebrītis, Nick Craig-Wood)
|
* Empty directory markers (Jānis Bebrītis, Nick Craig-Wood)
|
||||||
* Update Scaleway storage classes (Brian Starkey)
|
* Update Scaleway storage classes (Brian Starkey)
|
||||||
* Fix `--s3-versions` on individual objects (Nick Craig-Wood)
|
* Fix `--s3-versions` on individual objects (Nick Craig-Wood)
|
||||||
* Fix hang on aborting multpart upload with iDrive e2 (Nick Craig-Wood)
|
* Fix hang on aborting multipart upload with iDrive e2 (Nick Craig-Wood)
|
||||||
* Fix missing "tier" metadata (Nick Craig-Wood)
|
* Fix missing "tier" metadata (Nick Craig-Wood)
|
||||||
* Fix V3sign: add missing subresource delete (cc)
|
* Fix V3sign: add missing subresource delete (cc)
|
||||||
* Fix Arvancloud Domain and region changes and alphabetise the provider (Ehsan Tadayon)
|
* Fix Arvancloud Domain and region changes and alphabetise the provider (Ehsan Tadayon)
|
||||||
@ -309,7 +309,7 @@ description: "Rclone Changelog"
|
|||||||
* Code cleanup to avoid overwriting ctx before first use (fixes issue reported by the staticcheck linter) (albertony)
|
* Code cleanup to avoid overwriting ctx before first use (fixes issue reported by the staticcheck linter) (albertony)
|
||||||
* Storj
|
* Storj
|
||||||
* Fix "uplink: too many requests" errors when uploading to the same file (Nick Craig-Wood)
|
* Fix "uplink: too many requests" errors when uploading to the same file (Nick Craig-Wood)
|
||||||
* Fix uploading to the wrong object on Update with overriden remote name (Nick Craig-Wood)
|
* Fix uploading to the wrong object on Update with overridden remote name (Nick Craig-Wood)
|
||||||
* Swift
|
* Swift
|
||||||
* Ignore 404 error when deleting an object (Nick Craig-Wood)
|
* Ignore 404 error when deleting an object (Nick Craig-Wood)
|
||||||
* Union
|
* Union
|
||||||
@ -3938,7 +3938,7 @@ Point release to fix hubic and azureblob backends.
|
|||||||
* Revert to copy when moving file across file system boundaries
|
* Revert to copy when moving file across file system boundaries
|
||||||
* `--skip-links` to suppress symlink warnings (thanks Zhiming Wang)
|
* `--skip-links` to suppress symlink warnings (thanks Zhiming Wang)
|
||||||
* Mount
|
* Mount
|
||||||
* Re-use `rcat` internals to support uploads from all remotes
|
* Reuse `rcat` internals to support uploads from all remotes
|
||||||
* Dropbox
|
* Dropbox
|
||||||
* Fix "entry doesn't belong in directory" error
|
* Fix "entry doesn't belong in directory" error
|
||||||
* Stop using deprecated API methods
|
* Stop using deprecated API methods
|
||||||
|
@ -712,7 +712,7 @@ has a header and is divided into chunks.
|
|||||||
The initial nonce is generated from the operating systems crypto
|
The initial nonce is generated from the operating systems crypto
|
||||||
strong random number generator. The nonce is incremented for each
|
strong random number generator. The nonce is incremented for each
|
||||||
chunk read making sure each nonce is unique for each block written.
|
chunk read making sure each nonce is unique for each block written.
|
||||||
The chance of a nonce being re-used is minuscule. If you wrote an
|
The chance of a nonce being reused is minuscule. If you wrote an
|
||||||
exabyte of data (10¹⁸ bytes) you would have a probability of
|
exabyte of data (10¹⁸ bytes) you would have a probability of
|
||||||
approximately 2×10⁻³² of re-using a nonce.
|
approximately 2×10⁻³² of re-using a nonce.
|
||||||
|
|
||||||
|
@ -315,7 +315,7 @@ Make sure you have [Snapd installed](https://snapcraft.io/docs/installing-snapd)
|
|||||||
```bash
|
```bash
|
||||||
$ sudo snap install rclone
|
$ sudo snap install rclone
|
||||||
```
|
```
|
||||||
Due to the strict confinement of Snap, rclone snap cannot acess real /home/$USER/.config/rclone directory, default config path is as below.
|
Due to the strict confinement of Snap, rclone snap cannot access real /home/$USER/.config/rclone directory, default config path is as below.
|
||||||
|
|
||||||
- Default config directory:
|
- Default config directory:
|
||||||
- /home/$USER/snap/rclone/current/.config/rclone
|
- /home/$USER/snap/rclone/current/.config/rclone
|
||||||
@ -575,7 +575,7 @@ It requires .NET Framework, but it is preinstalled on newer versions of Windows,
|
|||||||
also provides alternative standalone distributions which includes necessary runtime (.NET 5).
|
also provides alternative standalone distributions which includes necessary runtime (.NET 5).
|
||||||
WinSW is a command-line only utility, where you have to manually create an XML file with
|
WinSW is a command-line only utility, where you have to manually create an XML file with
|
||||||
service configuration. This may be a drawback for some, but it can also be an advantage
|
service configuration. This may be a drawback for some, but it can also be an advantage
|
||||||
as it is easy to back up and re-use the configuration
|
as it is easy to back up and reuse the configuration
|
||||||
settings, without having go through manual steps in a GUI. One thing to note is that
|
settings, without having go through manual steps in a GUI. One thing to note is that
|
||||||
by default it does not restart the service on error, one have to explicit enable this
|
by default it does not restart the service on error, one have to explicit enable this
|
||||||
in the configuration file (via the "onfailure" parameter).
|
in the configuration file (via the "onfailure" parameter).
|
||||||
|
@ -209,7 +209,7 @@ rclone mount \
|
|||||||
# its exact meaning will depend on the backend. For HTTP based backends it is an HTTP PUT/GET/POST/etc and its response
|
# its exact meaning will depend on the backend. For HTTP based backends it is an HTTP PUT/GET/POST/etc and its response
|
||||||
--cache-dir /tmp/rclone/cache # Directory rclone will use for caching.
|
--cache-dir /tmp/rclone/cache # Directory rclone will use for caching.
|
||||||
--dir-cache-time 5m \ # Time to cache directory entries for (default 5m0s)
|
--dir-cache-time 5m \ # Time to cache directory entries for (default 5m0s)
|
||||||
--vfs-cache-mode writes \ # Cache mode off|minimal|writes|full (default off), writes gives the maximum compatiblity like a local disk
|
--vfs-cache-mode writes \ # Cache mode off|minimal|writes|full (default off), writes gives the maximum compatibility like a local disk
|
||||||
--vfs-cache-max-age 20m \ # Max age of objects in the cache (default 1h0m0s)
|
--vfs-cache-max-age 20m \ # Max age of objects in the cache (default 1h0m0s)
|
||||||
--vfs-cache-max-size 10G \ # Max total size of objects in the cache (default off)
|
--vfs-cache-max-size 10G \ # Max total size of objects in the cache (default off)
|
||||||
--vfs-cache-poll-interval 1m \ # Interval to poll the cache for stale objects (default 1m0s)
|
--vfs-cache-poll-interval 1m \ # Interval to poll the cache for stale objects (default 1m0s)
|
||||||
@ -372,7 +372,7 @@ Install NFS Utils
|
|||||||
sudo yum install -y nfs-utils
|
sudo yum install -y nfs-utils
|
||||||
```
|
```
|
||||||
|
|
||||||
Export the desired directory via NFS Server in the same machine where rclone has mounted to, ensure NFS serivce has
|
Export the desired directory via NFS Server in the same machine where rclone has mounted to, ensure NFS service has
|
||||||
desired permissions to read the directory. If it runs as root, then it will have permissions for sure, but if it runs
|
desired permissions to read the directory. If it runs as root, then it will have permissions for sure, but if it runs
|
||||||
as separate user then ensure that user has necessary desired privileges.
|
as separate user then ensure that user has necessary desired privileges.
|
||||||
```shell
|
```shell
|
||||||
|
@ -152,7 +152,7 @@ func TestMemoryObject(t *testing.T) {
|
|||||||
err = o.Update(context.Background(), newContent, src)
|
err = o.Update(context.Background(), newContent, src)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
checkContent(o, newStr)
|
checkContent(o, newStr)
|
||||||
assert.Equal(t, "Rutaba", string(content)) // check we didn't re-use the buffer
|
assert.Equal(t, "Rutaba", string(content)) // check we didn't reuse the buffer
|
||||||
|
|
||||||
// now try streaming
|
// now try streaming
|
||||||
newStr = "ABCDEFGHIJKLMNOPQRSTUVWXYZ"
|
newStr = "ABCDEFGHIJKLMNOPQRSTUVWXYZ"
|
||||||
|
@ -318,7 +318,7 @@ func TestRcSetTier(t *testing.T) {
|
|||||||
r.CheckRemoteItems(t, file1)
|
r.CheckRemoteItems(t, file1)
|
||||||
|
|
||||||
// Because we don't know what the current tier options here are, let's
|
// Because we don't know what the current tier options here are, let's
|
||||||
// just get the current tier, and re-use that
|
// just get the current tier, and reuse that
|
||||||
o, err := r.Fremote.NewObject(ctx, file1.Path)
|
o, err := r.Fremote.NewObject(ctx, file1.Path)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
trr, ok := o.(fs.GetTierer)
|
trr, ok := o.(fs.GetTierer)
|
||||||
@ -345,7 +345,7 @@ func TestRcSetTierFile(t *testing.T) {
|
|||||||
r.CheckRemoteItems(t, file1)
|
r.CheckRemoteItems(t, file1)
|
||||||
|
|
||||||
// Because we don't know what the current tier options here are, let's
|
// Because we don't know what the current tier options here are, let's
|
||||||
// just get the current tier, and re-use that
|
// just get the current tier, and reuse that
|
||||||
o, err := r.Fremote.NewObject(ctx, file1.Path)
|
o, err := r.Fremote.NewObject(ctx, file1.Path)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
trr, ok := o.(fs.GetTierer)
|
trr, ok := o.(fs.GetTierer)
|
||||||
|
@ -345,7 +345,7 @@ func (dls *Downloaders) _ensureDownloader(r ranges.Range) (err error) {
|
|||||||
start, offset := dl.getRange()
|
start, offset := dl.getRange()
|
||||||
|
|
||||||
// The downloader's offset to offset+window is the gap
|
// The downloader's offset to offset+window is the gap
|
||||||
// in which we would like to re-use this
|
// in which we would like to reuse this
|
||||||
// downloader. The downloader will never reach before
|
// downloader. The downloader will never reach before
|
||||||
// start and offset+windows is too far away - we'd
|
// start and offset+windows is too far away - we'd
|
||||||
// rather start another downloader.
|
// rather start another downloader.
|
||||||
|
Loading…
Reference in New Issue
Block a user