From b1d4de69c247dfc3689f87dda69b48d99ba10dc2 Mon Sep 17 00:00:00 2001 From: Dimitri Papadopoulos <3234522+DimitriPapadopoulos@users.noreply.github.com> Date: Wed, 15 Jan 2025 14:01:06 +0100 Subject: [PATCH] docs: fix typos found by codespell in docs and code comments --- backend/azurefiles/azurefiles.go | 2 +- backend/iclouddrive/iclouddrive.go | 2 +- backend/imagekit/client/media.go | 2 +- backend/pcloud/pcloud.go | 2 +- backend/s3/s3.go | 4 ++-- bin/not-in-stable.go | 10 +++++----- cmd/bisync/compare.go | 2 +- cmd/bisync/listing.go | 2 +- cmd/bisync/testdata/test_resync/golden/test.log | 2 +- cmd/bisync/testdata/test_resync/scenario.txt | 4 ++-- cmd/config/config.go | 4 ++-- cmd/copyurl/copyurl.go | 2 +- cmd/serve/nfs/filesystem.go | 2 +- cmd/serve/nfs/nfs.go | 2 +- cmd/serve/s3/backend.go | 4 ++-- cmd/serve/s3/list.go | 2 +- docs/content/bisync.md | 12 ++++++------ docs/content/crypt.md | 2 +- docs/content/docs.md | 2 +- docs/content/googlephotos.md | 2 +- docs/content/rc.md | 2 +- docs/content/s3.md | 2 +- fs/fshttp/http_test.go | 2 +- fs/log.go | 2 +- fs/object/object.go | 2 +- fs/object/object_test.go | 2 +- fstest/fstests/fstests.go | 2 +- vfs/rc.go | 2 +- vfs/vfs.go | 2 +- vfs/vfscache/item_test.go | 2 +- 30 files changed, 43 insertions(+), 43 deletions(-) diff --git a/backend/azurefiles/azurefiles.go b/backend/azurefiles/azurefiles.go index acff5be19..7d9a53075 100644 --- a/backend/azurefiles/azurefiles.go +++ b/backend/azurefiles/azurefiles.go @@ -899,7 +899,7 @@ func (o *Object) getMetadata(ctx context.Context) error { // Hash returns the MD5 of an object returning a lowercase hex string // -// May make a network request becaue the [fs.List] method does not +// May make a network request because the [fs.List] method does not // return MD5 hashes for DirEntry func (o *Object) Hash(ctx context.Context, ty hash.Type) (string, error) { if ty != hash.MD5 { diff --git a/backend/iclouddrive/iclouddrive.go b/backend/iclouddrive/iclouddrive.go index 591d81718..7228db060 100644 --- a/backend/iclouddrive/iclouddrive.go +++ b/backend/iclouddrive/iclouddrive.go @@ -445,7 +445,7 @@ func (f *Fs) Copy(ctx context.Context, src fs.Object, remote string) (fs.Object, } // build request - // cant use normal rename as file needs to be "activated" first + // can't use normal rename as file needs to be "activated" first r := api.NewUpdateFileInfo() r.DocumentID = doc.DocumentID diff --git a/backend/imagekit/client/media.go b/backend/imagekit/client/media.go index d495b713b..1d4ee030b 100644 --- a/backend/imagekit/client/media.go +++ b/backend/imagekit/client/media.go @@ -75,7 +75,7 @@ type MoveFolderParam struct { DestinationPath string `validate:"nonzero" json:"destinationPath"` } -// JobIDResponse respresents response struct with JobID for folder operations +// JobIDResponse represents response struct with JobID for folder operations type JobIDResponse struct { JobID string `json:"jobId"` } diff --git a/backend/pcloud/pcloud.go b/backend/pcloud/pcloud.go index f198c5563..763784645 100644 --- a/backend/pcloud/pcloud.go +++ b/backend/pcloud/pcloud.go @@ -424,7 +424,7 @@ func (f *Fs) newSingleConnClient(ctx context.Context) (*rest.Client, error) { }) // Set our own http client in the context ctx = oauthutil.Context(ctx, baseClient) - // create a new oauth client, re-use the token source + // create a new oauth client, reuse the token source oAuthClient := oauth2.NewClient(ctx, f.ts) return rest.NewClient(oAuthClient).SetRoot("https://" + f.opt.Hostname), nil } diff --git a/backend/s3/s3.go b/backend/s3/s3.go index a0eda19ab..d487d9404 100644 --- a/backend/s3/s3.go +++ b/backend/s3/s3.go @@ -3344,7 +3344,7 @@ func setQuirks(opt *Options) { listObjectsV2 = true // Always use ListObjectsV2 instead of ListObjects virtualHostStyle = true // Use bucket.provider.com instead of putting the bucket in the URL urlEncodeListings = true // URL encode the listings to help with control characters - useMultipartEtag = true // Set if Etags for multpart uploads are compatible with AWS + useMultipartEtag = true // Set if Etags for multipart uploads are compatible with AWS useAcceptEncodingGzip = true // Set Accept-Encoding: gzip mightGzip = true // assume all providers might use content encoding gzip until proven otherwise useAlreadyExists = true // Set if provider returns AlreadyOwnedByYou or no error if you try to remake your own bucket @@ -6057,7 +6057,7 @@ func (f *Fs) OpenChunkWriter(ctx context.Context, remote string, src fs.ObjectIn if mOut == nil { err = fserrors.RetryErrorf("internal error: no info from multipart upload") } else if mOut.UploadId == nil { - err = fserrors.RetryErrorf("internal error: no UploadId in multpart upload: %#v", *mOut) + err = fserrors.RetryErrorf("internal error: no UploadId in multipart upload: %#v", *mOut) } } return f.shouldRetry(ctx, err) diff --git a/bin/not-in-stable.go b/bin/not-in-stable.go index 57a025760..5f988af82 100644 --- a/bin/not-in-stable.go +++ b/bin/not-in-stable.go @@ -29,7 +29,7 @@ func readCommits(from, to string) (logMap map[string]string, logs []string) { cmd := exec.Command("git", "log", "--oneline", from+".."+to) out, err := cmd.Output() if err != nil { - log.Fatalf("failed to run git log %s: %v", from+".."+to, err) //nolint:gocritic // Don't include gocritic when running golangci-lint to avoid ruleguard suggesting fs. intead of log. + log.Fatalf("failed to run git log %s: %v", from+".."+to, err) //nolint:gocritic // Don't include gocritic when running golangci-lint to avoid ruleguard suggesting fs. instead of log. } logMap = map[string]string{} logs = []string{} @@ -39,7 +39,7 @@ func readCommits(from, to string) (logMap map[string]string, logs []string) { } match := logRe.FindSubmatch(line) if match == nil { - log.Fatalf("failed to parse line: %q", line) //nolint:gocritic // Don't include gocritic when running golangci-lint to avoid ruleguard suggesting fs. intead of log. + log.Fatalf("failed to parse line: %q", line) //nolint:gocritic // Don't include gocritic when running golangci-lint to avoid ruleguard suggesting fs. instead of log. } var hash, logMessage = string(match[1]), string(match[2]) logMap[logMessage] = hash @@ -52,12 +52,12 @@ func main() { flag.Parse() args := flag.Args() if len(args) != 0 { - log.Fatalf("Syntax: %s", os.Args[0]) //nolint:gocritic // Don't include gocritic when running golangci-lint to avoid ruleguard suggesting fs. intead of log. + log.Fatalf("Syntax: %s", os.Args[0]) //nolint:gocritic // Don't include gocritic when running golangci-lint to avoid ruleguard suggesting fs. instead of log. } // v1.54.0 versionBytes, err := os.ReadFile("VERSION") if err != nil { - log.Fatalf("Failed to read version: %v", err) //nolint:gocritic // Don't include gocritic when running golangci-lint to avoid ruleguard suggesting fs. intead of log. + log.Fatalf("Failed to read version: %v", err) //nolint:gocritic // Don't include gocritic when running golangci-lint to avoid ruleguard suggesting fs. instead of log. } if versionBytes[0] == 'v' { versionBytes = versionBytes[1:] @@ -65,7 +65,7 @@ func main() { versionBytes = bytes.TrimSpace(versionBytes) semver := semver.New(string(versionBytes)) stable := fmt.Sprintf("v%d.%d", semver.Major, semver.Minor-1) - log.Printf("Finding commits in %v not in stable %s", semver, stable) //nolint:gocritic // Don't include gocritic when running golangci-lint to avoid ruleguard suggesting fs. intead of log. + log.Printf("Finding commits in %v not in stable %s", semver, stable) //nolint:gocritic // Don't include gocritic when running golangci-lint to avoid ruleguard suggesting fs. instead of log. masterMap, masterLogs := readCommits(stable+".0", "master") stableMap, _ := readCommits(stable+".0", stable+"-stable") for _, logMessage := range masterLogs { diff --git a/cmd/bisync/compare.go b/cmd/bisync/compare.go index 4dea6bedc..73ced57af 100644 --- a/cmd/bisync/compare.go +++ b/cmd/bisync/compare.go @@ -218,7 +218,7 @@ func (b *bisyncRun) setFromCompareFlag(ctx context.Context) error { if b.opt.CompareFlag == "" { return nil } - var CompareFlag CompareOpt // for exlcusions + var CompareFlag CompareOpt // for exclusions opts := strings.Split(b.opt.CompareFlag, ",") for _, opt := range opts { switch strings.ToLower(strings.TrimSpace(opt)) { diff --git a/cmd/bisync/listing.go b/cmd/bisync/listing.go index 13179b26f..7451a5cbb 100644 --- a/cmd/bisync/listing.go +++ b/cmd/bisync/listing.go @@ -394,7 +394,7 @@ func parseHash(str string) (string, string, error) { return "", "", fmt.Errorf("invalid hash %q", str) } -// checkListing verifies that listing is not empty (unless resynching) +// checkListing verifies that listing is not empty (unless resyncing) func (b *bisyncRun) checkListing(ls *fileList, listing, msg string) error { if b.opt.Resync || !ls.empty() { return nil diff --git a/cmd/bisync/testdata/test_resync/golden/test.log b/cmd/bisync/testdata/test_resync/golden/test.log index f9488a672..13a3bddcb 100644 --- a/cmd/bisync/testdata/test_resync/golden/test.log +++ b/cmd/bisync/testdata/test_resync/golden/test.log @@ -23,7 +23,7 @@ INFO : Validating listings for Path1 "{path1/}" vs Path2 "{path2/}" INFO : Bisync successful (05) : move-listings empty-path1 -(06) : test 2. resync with empty path2, resulting in synching all content to path2. +(06) : test 2. resync with empty path2, resulting in syncing all content to path2. (07) : purge-children {path2/} (08) : bisync resync INFO : Setting --ignore-listing-checksum as neither --checksum nor --compare checksum are set. diff --git a/cmd/bisync/testdata/test_resync/scenario.txt b/cmd/bisync/testdata/test_resync/scenario.txt index 078254646..4a0e86475 100644 --- a/cmd/bisync/testdata/test_resync/scenario.txt +++ b/cmd/bisync/testdata/test_resync/scenario.txt @@ -1,6 +1,6 @@ test resync # 1. Resync with empty Path1, resulting in copying all content FROM Path2 -# 2. Resync with empty Path2, resulting in synching all content TO Path2 +# 2. Resync with empty Path2, resulting in syncing all content TO Path2 # 3. Exercise all of the various file difference scenarios during a resync: # File Path1 Path2 Expected action Who wins # - file1.txt Exists Missing Sync Path1 >Path2 Path1 @@ -17,7 +17,7 @@ purge-children {path1/} bisync resync move-listings empty-path1 -test 2. resync with empty path2, resulting in synching all content to path2. +test 2. resync with empty path2, resulting in syncing all content to path2. purge-children {path2/} bisync resync move-listings empty-path2 diff --git a/cmd/config/config.go b/cmd/config/config.go index 4aefe62c0..90630a9bc 100644 --- a/cmd/config/config.go +++ b/cmd/config/config.go @@ -549,12 +549,12 @@ password to re-encrypt the config. When |--password-command| is called to change the password then the environment variable |RCLONE_PASSWORD_CHANGE=1| will be set. So if -changing passwords programatically you can use the environment +changing passwords programmatically you can use the environment variable to distinguish which password you must supply. Alternatively you can remove the password first (with |rclone config encryption remove|), then set it again with this command which may be -easier if you don't mind the unecrypted config file being on the disk +easier if you don't mind the unencrypted config file being on the disk briefly. `, "|", "`"), RunE: func(command *cobra.Command, args []string) error { diff --git a/cmd/copyurl/copyurl.go b/cmd/copyurl/copyurl.go index 2a615de28..c5f46cbc7 100644 --- a/cmd/copyurl/copyurl.go +++ b/cmd/copyurl/copyurl.go @@ -54,7 +54,7 @@ destination if there is one with the same name. Setting |--stdout| or making the output file name |-| will cause the output to be written to standard output. -### Troublshooting +### Troubleshooting If you can't get |rclone copyurl| to work then here are some things you can try: diff --git a/cmd/serve/nfs/filesystem.go b/cmd/serve/nfs/filesystem.go index 15e786c49..fe30326cb 100644 --- a/cmd/serve/nfs/filesystem.go +++ b/cmd/serve/nfs/filesystem.go @@ -194,7 +194,7 @@ func (f *FS) Chown(name string, uid, gid int) (err error) { return file.Chown(uid, gid) } -// Chtimes changes the acces time and modified time +// Chtimes changes the access time and modified time func (f *FS) Chtimes(name string, atime time.Time, mtime time.Time) (err error) { defer log.Trace(name, "atime=%v, mtime=%v", atime, mtime)("err=%v", &err) return f.vfs.Chtimes(name, atime, mtime) diff --git a/cmd/serve/nfs/nfs.go b/cmd/serve/nfs/nfs.go index acbe943f5..ad94c5a49 100644 --- a/cmd/serve/nfs/nfs.go +++ b/cmd/serve/nfs/nfs.go @@ -145,7 +145,7 @@ that it uses an on disk cache, but the cache entries are held as symlinks. Rclone will use the handle of the underlying file as the NFS handle which improves performance. This sort of cache can't be backed up and restored as the underlying handles will change. This is Linux -only. It requres running rclone as root or with |CAP_DAC_READ_SEARCH|. +only. It requires running rclone as root or with |CAP_DAC_READ_SEARCH|. You can run rclone with this extra permission by doing this to the rclone binary |sudo setcap cap_dac_read_search+ep /path/to/rclone|. diff --git a/cmd/serve/s3/backend.go b/cmd/serve/s3/backend.go index 42671b11e..33ff6387b 100644 --- a/cmd/serve/s3/backend.go +++ b/cmd/serve/s3/backend.go @@ -158,7 +158,7 @@ func (b *s3Backend) HeadObject(ctx context.Context, bucketName, objectName strin }, nil } -// GetObject fetchs the object from the filesystem. +// GetObject fetches the object from the filesystem. func (b *s3Backend) GetObject(ctx context.Context, bucketName, objectName string, rangeRequest *gofakes3.ObjectRangeRequest) (obj *gofakes3.Object, err error) { _vfs, err := b.s.getVFS(ctx) if err != nil { @@ -400,7 +400,7 @@ func (b *s3Backend) deleteObject(ctx context.Context, bucketName, objectName str } fp := path.Join(bucketName, objectName) - // S3 does not report an error when attemping to delete a key that does not exist, so + // S3 does not report an error when attempting to delete a key that does not exist, so // we need to skip IsNotExist errors. if err := _vfs.Remove(fp); err != nil && !os.IsNotExist(err) { return err diff --git a/cmd/serve/s3/list.go b/cmd/serve/s3/list.go index 554d152ab..9b0a2e11e 100644 --- a/cmd/serve/s3/list.go +++ b/cmd/serve/s3/list.go @@ -19,7 +19,7 @@ func (b *s3Backend) entryListR(_vfs *vfs.VFS, bucket, fdPath, name string, addPr for _, entry := range dirEntries { object := entry.Name() - // workround for control-chars detect + // workaround for control-chars detect objectPath := path.Join(fdPath, object) if !strings.HasPrefix(object, name) { diff --git a/docs/content/bisync.md b/docs/content/bisync.md index 94a51fd9e..458f92eab 100644 --- a/docs/content/bisync.md +++ b/docs/content/bisync.md @@ -487,7 +487,7 @@ See the [bisync filters](#filtering) section and generic [--filter-from](/filtering/#filter-from-read-filtering-patterns-from-a-file) documentation. An [example filters file](#example-filters-file) contains filters for -non-allowed files for synching with Dropbox. +non-allowed files for syncing with Dropbox. If you make changes to your filters file then bisync requires a run with `--resync`. This is a safety feature, which prevents existing files @@ -664,7 +664,7 @@ Using `--check-sync=false` will disable it and may significantly reduce the sync run times for very large numbers of files. The check may be run manually with `--check-sync=only`. It runs only the -integrity check and terminates without actually synching. +integrity check and terminates without actually syncing. Note that currently, `--check-sync` **only checks listing snapshots and NOT the actual files on the remotes.** Note also that the listing snapshots will not @@ -1141,7 +1141,7 @@ The `--include*`, `--exclude*`, and `--filter` flags are also supported. ### How to filter directories -Filtering portions of the directory tree is a critical feature for synching. +Filtering portions of the directory tree is a critical feature for syncing. Examples of directory trees (always beneath the Path1/Path2 root level) you may want to exclude from your sync: @@ -1250,7 +1250,7 @@ quashed by adding `--quiet` to the bisync command line. ## Example exclude-style filters files for use with Dropbox {#exclude-filters} -- Dropbox disallows synching the listed temporary and configuration/data files. +- Dropbox disallows syncing the listed temporary and configuration/data files. The `- ` filters exclude these files where ever they may occur in the sync tree. Consider adding similar exclusions for file types you don't need to sync, such as core dump and software build files. @@ -1584,7 +1584,7 @@ test command flags can be equally prefixed by a single `-` or double dash. - `go test . -case basic -remote local -remote2 local` runs the `test_basic` test case using only the local filesystem, - synching one local directory with another local directory. + syncing one local directory with another local directory. Test script output is to the console, while commands within scenario.txt have their output sent to the `.../workdir/test.log` file, which is finally compared to the golden copy. @@ -1860,4 +1860,4 @@ causing bisync to consider more files than necessary due to overbroad filters du * Added [new `--ignore-listing-checksum` flag](https://forum.rclone.org/t/bisync-bugs-and-feature-requests/37636#:~:text=6.%20%2D%2Dignore%2Dchecksum%20should%20be%20split%20into%20two%20flags%20for%20separate%20purposes) to distinguish from `--ignore-checksum` * [Performance improvements](https://forum.rclone.org/t/bisync-bugs-and-feature-requests/37636#:~:text=6.%20Deletes%20take%20several%20times%20longer%20than%20copies) for large remotes -* Documentation and testing improvements \ No newline at end of file +* Documentation and testing improvements diff --git a/docs/content/crypt.md b/docs/content/crypt.md index 2ea2bf7d5..db3edcc74 100644 --- a/docs/content/crypt.md +++ b/docs/content/crypt.md @@ -741,7 +741,7 @@ strong random number generator. The nonce is incremented for each chunk read making sure each nonce is unique for each block written. The chance of a nonce being reused is minuscule. If you wrote an exabyte of data (10¹⁸ bytes) you would have a probability of -approximately 2×10⁻³² of re-using a nonce. +approximately 2×10⁻³² of reusing a nonce. #### Chunk diff --git a/docs/content/docs.md b/docs/content/docs.md index dccb5c5f7..8b221f130 100644 --- a/docs/content/docs.md +++ b/docs/content/docs.md @@ -2930,7 +2930,7 @@ so they take exactly the same form. The options set by environment variables can be seen with the `-vv` flag, e.g. `rclone version -vv`. Options that can appear multiple times (type `stringArray`) are -treated slighly differently as environment variables can only be +treated slightly differently as environment variables can only be defined once. In order to allow a simple mechanism for adding one or many items, the input is treated as a [CSV encoded](https://godoc.org/encoding/csv) string. For example diff --git a/docs/content/googlephotos.md b/docs/content/googlephotos.md index 26faedcc6..c3549d158 100644 --- a/docs/content/googlephotos.md +++ b/docs/content/googlephotos.md @@ -384,7 +384,7 @@ Use the gphotosdl proxy for downloading the full resolution images The Google API will deliver images and video which aren't full resolution, and/or have EXIF data missing. -However if you ue the gphotosdl proxy tnen you can download original, +However if you use the gphotosdl proxy then you can download original, unchanged images. This runs a headless browser in the background. diff --git a/docs/content/rc.md b/docs/content/rc.md index 541093bb9..4f54bb170 100644 --- a/docs/content/rc.md +++ b/docs/content/rc.md @@ -2068,7 +2068,7 @@ the `--vfs-cache-mode` is off, it will return an empty result. ], } -The `expiry` time is the time until the file is elegible for being +The `expiry` time is the time until the file is eligible for being uploaded in floating point seconds. This may go negative. As rclone only transfers `--transfers` files at once, only the lowest `--transfers` expiry times will have `uploading` as `true`. So there diff --git a/docs/content/s3.md b/docs/content/s3.md index fa1fbe2e5..a71549e94 100644 --- a/docs/content/s3.md +++ b/docs/content/s3.md @@ -750,7 +750,7 @@ Notes on above: that `USER_NAME` has been created. 2. The Resource entry must include both resource ARNs, as one implies the bucket and the other implies the bucket's objects. -3. When using [s3-no-check-bucket](#s3-no-check-bucket) and the bucket already exsits, the `"arn:aws:s3:::BUCKET_NAME"` doesn't have to be included. +3. When using [s3-no-check-bucket](#s3-no-check-bucket) and the bucket already exists, the `"arn:aws:s3:::BUCKET_NAME"` doesn't have to be included. For reference, [here's an Ansible script](https://gist.github.com/ebridges/ebfc9042dd7c756cd101cfa807b7ae2b) that will generate one or more buckets that will work with `rclone sync`. diff --git a/fs/fshttp/http_test.go b/fs/fshttp/http_test.go index 2c10e84a2..18f191e73 100644 --- a/fs/fshttp/http_test.go +++ b/fs/fshttp/http_test.go @@ -133,7 +133,7 @@ func TestCertificates(t *testing.T) { assert.Fail(t, "Certificate expired", "Certificate expires at %s, current time is %s", cert[0].NotAfter.Sub(startTime), time.Since(startTime)) } - // Write some test data to fullfil the request + // Write some test data to fulfill the request w.Header().Set("Content-Type", "text/plain") _, _ = fmt.Fprintln(w, "test data") })) diff --git a/fs/log.go b/fs/log.go index 4b3abf611..5088d1600 100644 --- a/fs/log.go +++ b/fs/log.go @@ -95,7 +95,7 @@ func LogValueHide(key string, value interface{}) LogValueItem { return LogValueItem{key: key, value: value, render: false} } -// String returns the representation of value. If render is fals this +// String returns the representation of value. If render is false this // is an empty string so LogValueItem entries won't show in the // textual representation of logs. func (j LogValueItem) String() string { diff --git a/fs/object/object.go b/fs/object/object.go index a8346edee..40aee8dec 100644 --- a/fs/object/object.go +++ b/fs/object/object.go @@ -297,7 +297,7 @@ func (o *MemoryObject) Open(ctx context.Context, options ...fs.OpenOption) (io.R // Update in to the object with the modTime given of the given size // -// This re-uses the internal buffer if at all possible. +// This reuses the internal buffer if at all possible. func (o *MemoryObject) Update(ctx context.Context, in io.Reader, src fs.ObjectInfo, options ...fs.OpenOption) (err error) { size := src.Size() if size == 0 { diff --git a/fs/object/object_test.go b/fs/object/object_test.go index 64bb36140..2a4236272 100644 --- a/fs/object/object_test.go +++ b/fs/object/object_test.go @@ -142,7 +142,7 @@ func TestMemoryObject(t *testing.T) { assert.NoError(t, err) checkContent(o, "Rutabaga") assert.Equal(t, newNow, o.ModTime(context.Background())) - assert.Equal(t, "Rutaba", string(content)) // check we re-used the buffer + assert.Equal(t, "Rutaba", string(content)) // check we reused the buffer // not within the buffer newStr := "0123456789" diff --git a/fstest/fstests/fstests.go b/fstest/fstests/fstests.go index 6024177c6..2369e7c2d 100644 --- a/fstest/fstests/fstests.go +++ b/fstest/fstests/fstests.go @@ -2391,7 +2391,7 @@ func Run(t *testing.T, opt *Opt) { var itemCopy = item itemCopy.Path += ".copy" - // Set copy cutoff to mininum value so we make chunks + // Set copy cutoff to minimum value so we make chunks origCutoff, err := do.SetCopyCutoff(minChunkSize) require.NoError(t, err) defer func() { diff --git a/vfs/rc.go b/vfs/rc.go index 5c74e9f91..c2ad3caca 100644 --- a/vfs/rc.go +++ b/vfs/rc.go @@ -464,7 +464,7 @@ the |--vfs-cache-mode| is off, it will return an empty result. ], } -The |expiry| time is the time until the file is elegible for being +The |expiry| time is the time until the file is eligible for being uploaded in floating point seconds. This may go negative. As rclone only transfers |--transfers| files at once, only the lowest |--transfers| expiry times will have |uploading| as |true|. So there diff --git a/vfs/vfs.go b/vfs/vfs.go index aa6fe15eb..68b17f483 100644 --- a/vfs/vfs.go +++ b/vfs/vfs.go @@ -216,7 +216,7 @@ func New(f fs.Fs, opt *vfscommon.Options) *VFS { configName := fs.ConfigString(f) for _, activeVFS := range active[configName] { if vfs.Opt == activeVFS.Opt { - fs.Debugf(f, "Re-using VFS from active cache") + fs.Debugf(f, "Reusing VFS from active cache") activeVFS.inUse.Add(1) return activeVFS } diff --git a/vfs/vfscache/item_test.go b/vfs/vfscache/item_test.go index 9fa2f3dbe..fa5840e01 100644 --- a/vfs/vfscache/item_test.go +++ b/vfs/vfscache/item_test.go @@ -428,7 +428,7 @@ func TestItemReloadCacheStale(t *testing.T) { assert.Equal(t, int64(110), fi.Size()) // Write to the file to make it dirty - // This checks we aren't re-using stale data + // This checks we aren't reusing stale data n, err := item.WriteAt([]byte("HELLO"), 0) require.NoError(t, err) assert.Equal(t, 5, n)