mirror of
https://github.com/rclone/rclone.git
synced 2024-12-15 11:41:43 +01:00
40111ba5e1
This changes the OpenWriterAt implementation to make client/fd handling atomic. This PR stabilizes the situation of bigger files and multi-threaded uploads. The root cause boils down to the old "fun" property of pclouds fileops API: sessions are bound to TCP connections. This forces us to use a http client with only a single connection underneath. With large files, we reuse the same connection for each chunk. If that connection interrupts (e.g. because we are talking through the internet), all chunks will fail. The probability for latter one increases with larger files. As the point of the whole multi-threaded feature was to speed-up large files in the first place, this change pulls the client creation (and hence connection handling) into each chunk. This should stabilize the situation, as each chunk (and retry) gets its own connection.
262 lines
7.6 KiB
Go
262 lines
7.6 KiB
Go
package pcloud
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"crypto/sha1"
|
|
"encoding/hex"
|
|
"fmt"
|
|
"net/url"
|
|
"strconv"
|
|
"time"
|
|
|
|
"github.com/rclone/rclone/backend/pcloud/api"
|
|
"github.com/rclone/rclone/fs"
|
|
"github.com/rclone/rclone/lib/rest"
|
|
)
|
|
|
|
// writerAt implements fs.WriterAtCloser, adding the OpenWrtierAt feature to pcloud.
|
|
type writerAt struct {
|
|
ctx context.Context
|
|
fs *Fs
|
|
size int64
|
|
remote string
|
|
fileID int64
|
|
}
|
|
|
|
// Close implements WriterAt.Close.
|
|
func (c *writerAt) Close() error {
|
|
// Avoiding race conditions: Depending on the tcp connection, there might be
|
|
// caching issues when checking the size immediately after write.
|
|
// Hence we try avoiding them by checking the resulting size on a different connection.
|
|
if c.size < 0 {
|
|
// Without knowing the size, we cannot do size checks.
|
|
// Falling back to a sleep of 1s for sake of hope.
|
|
time.Sleep(1 * time.Second)
|
|
return nil
|
|
}
|
|
sizeOk := false
|
|
sizeLastSeen := int64(0)
|
|
for retry := 0; retry < 5; retry++ {
|
|
fs.Debugf(c.remote, "checking file size: try %d/5", retry)
|
|
obj, err := c.fs.NewObject(c.ctx, c.remote)
|
|
if err != nil {
|
|
return fmt.Errorf("get uploaded obj: %w", err)
|
|
}
|
|
sizeLastSeen = obj.Size()
|
|
if obj.Size() == c.size {
|
|
sizeOk = true
|
|
break
|
|
}
|
|
time.Sleep(1 * time.Second)
|
|
}
|
|
|
|
if !sizeOk {
|
|
return fmt.Errorf("incorrect size after upload: got %d, want %d", sizeLastSeen, c.size)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// WriteAt implements fs.WriteAt.
|
|
func (c *writerAt) WriteAt(buffer []byte, offset int64) (n int, err error) {
|
|
contentLength := len(buffer)
|
|
|
|
inSHA1Bytes := sha1.Sum(buffer)
|
|
inSHA1 := hex.EncodeToString(inSHA1Bytes[:])
|
|
|
|
client, err := c.fs.newSingleConnClient(c.ctx)
|
|
if err != nil {
|
|
return 0, fmt.Errorf("create client: %w", err)
|
|
}
|
|
|
|
openResult, err := fileOpen(c.ctx, client, c.fs, c.fileID)
|
|
if err != nil {
|
|
return 0, fmt.Errorf("open file: %w", err)
|
|
}
|
|
|
|
// get target hash
|
|
outChecksum, err := fileChecksum(c.ctx, client, c.fs.pacer, openResult.FileDescriptor, offset, int64(contentLength))
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
outSHA1 := outChecksum.SHA1
|
|
|
|
if outSHA1 == "" || inSHA1 == "" {
|
|
return 0, fmt.Errorf("expect both hashes to be filled: src: %q, target: %q", inSHA1, outSHA1)
|
|
}
|
|
|
|
// check hash of buffer, skip if fits
|
|
if inSHA1 == outSHA1 {
|
|
return contentLength, nil
|
|
}
|
|
|
|
// upload buffer with offset if necessary
|
|
if _, err := filePWrite(c.ctx, client, c.fs.pacer, openResult.FileDescriptor, offset, buffer); err != nil {
|
|
return 0, err
|
|
}
|
|
|
|
// close fd
|
|
if _, err := fileClose(c.ctx, client, c.fs.pacer, openResult.FileDescriptor); err != nil {
|
|
return contentLength, fmt.Errorf("close fd: %w", err)
|
|
}
|
|
|
|
return contentLength, nil
|
|
}
|
|
|
|
// Call pcloud file_open using folderid and name with O_CREAT and O_WRITE flags, see [API Doc.]
|
|
// [API Doc]: https://docs.pcloud.com/methods/fileops/file_open.html
|
|
func fileOpenNew(ctx context.Context, c *rest.Client, srcFs *Fs, directoryID, filename string) (*api.FileOpenResponse, error) {
|
|
opts := rest.Opts{
|
|
Method: "PUT",
|
|
Path: "/file_open",
|
|
Parameters: url.Values{},
|
|
TransferEncoding: []string{"identity"}, // pcloud doesn't like chunked encoding
|
|
ExtraHeaders: map[string]string{
|
|
"Connection": "keep-alive",
|
|
},
|
|
}
|
|
filename = srcFs.opt.Enc.FromStandardName(filename)
|
|
opts.Parameters.Set("name", filename)
|
|
opts.Parameters.Set("folderid", dirIDtoNumber(directoryID))
|
|
opts.Parameters.Set("flags", "0x0042") // O_CREAT, O_WRITE
|
|
|
|
result := &api.FileOpenResponse{}
|
|
err := srcFs.pacer.CallNoRetry(func() (bool, error) {
|
|
resp, err := c.CallJSON(ctx, &opts, nil, result)
|
|
err = result.Error.Update(err)
|
|
return shouldRetry(ctx, resp, err)
|
|
})
|
|
if err != nil {
|
|
return nil, fmt.Errorf("open new file descriptor: %w", err)
|
|
}
|
|
return result, nil
|
|
}
|
|
|
|
// Call pcloud file_open using fileid with O_WRITE flags, see [API Doc.]
|
|
// [API Doc]: https://docs.pcloud.com/methods/fileops/file_open.html
|
|
func fileOpen(ctx context.Context, c *rest.Client, srcFs *Fs, fileID int64) (*api.FileOpenResponse, error) {
|
|
opts := rest.Opts{
|
|
Method: "PUT",
|
|
Path: "/file_open",
|
|
Parameters: url.Values{},
|
|
TransferEncoding: []string{"identity"}, // pcloud doesn't like chunked encoding
|
|
ExtraHeaders: map[string]string{
|
|
"Connection": "keep-alive",
|
|
},
|
|
}
|
|
opts.Parameters.Set("fileid", strconv.FormatInt(fileID, 10))
|
|
opts.Parameters.Set("flags", "0x0002") // O_WRITE
|
|
|
|
result := &api.FileOpenResponse{}
|
|
err := srcFs.pacer.CallNoRetry(func() (bool, error) {
|
|
resp, err := c.CallJSON(ctx, &opts, nil, result)
|
|
err = result.Error.Update(err)
|
|
return shouldRetry(ctx, resp, err)
|
|
})
|
|
if err != nil {
|
|
return nil, fmt.Errorf("open new file descriptor: %w", err)
|
|
}
|
|
return result, nil
|
|
}
|
|
|
|
// Call pcloud file_checksum, see [API Doc.]
|
|
// [API Doc]: https://docs.pcloud.com/methods/fileops/file_checksum.html
|
|
func fileChecksum(
|
|
ctx context.Context,
|
|
client *rest.Client,
|
|
pacer *fs.Pacer,
|
|
fd, offset, count int64,
|
|
) (*api.FileChecksumResponse, error) {
|
|
opts := rest.Opts{
|
|
Method: "PUT",
|
|
Path: "/file_checksum",
|
|
Parameters: url.Values{},
|
|
TransferEncoding: []string{"identity"}, // pcloud doesn't like chunked encoding
|
|
ExtraHeaders: map[string]string{
|
|
"Connection": "keep-alive",
|
|
},
|
|
}
|
|
opts.Parameters.Set("fd", strconv.FormatInt(fd, 10))
|
|
opts.Parameters.Set("offset", strconv.FormatInt(offset, 10))
|
|
opts.Parameters.Set("count", strconv.FormatInt(count, 10))
|
|
|
|
result := &api.FileChecksumResponse{}
|
|
err := pacer.CallNoRetry(func() (bool, error) {
|
|
resp, err := client.CallJSON(ctx, &opts, nil, result)
|
|
err = result.Error.Update(err)
|
|
return shouldRetry(ctx, resp, err)
|
|
})
|
|
if err != nil {
|
|
return nil, fmt.Errorf("checksum of fd %d with offset %d and size %d: %w", fd, offset, count, err)
|
|
}
|
|
return result, nil
|
|
}
|
|
|
|
// Call pcloud file_pwrite, see [API Doc.]
|
|
// [API Doc]: https://docs.pcloud.com/methods/fileops/file_pwrite.html
|
|
func filePWrite(
|
|
ctx context.Context,
|
|
client *rest.Client,
|
|
pacer *fs.Pacer,
|
|
fd int64,
|
|
offset int64,
|
|
buf []byte,
|
|
) (*api.FilePWriteResponse, error) {
|
|
contentLength := int64(len(buf))
|
|
opts := rest.Opts{
|
|
Method: "PUT",
|
|
Path: "/file_pwrite",
|
|
Body: bytes.NewReader(buf),
|
|
ContentLength: &contentLength,
|
|
Parameters: url.Values{},
|
|
TransferEncoding: []string{"identity"}, // pcloud doesn't like chunked encoding
|
|
Close: false,
|
|
ExtraHeaders: map[string]string{
|
|
"Connection": "keep-alive",
|
|
},
|
|
}
|
|
opts.Parameters.Set("fd", strconv.FormatInt(fd, 10))
|
|
opts.Parameters.Set("offset", strconv.FormatInt(offset, 10))
|
|
|
|
result := &api.FilePWriteResponse{}
|
|
err := pacer.CallNoRetry(func() (bool, error) {
|
|
resp, err := client.CallJSON(ctx, &opts, nil, result)
|
|
err = result.Error.Update(err)
|
|
return shouldRetry(ctx, resp, err)
|
|
})
|
|
if err != nil {
|
|
return nil, fmt.Errorf("write %d bytes to fd %d with offset %d: %w", contentLength, fd, offset, err)
|
|
}
|
|
return result, nil
|
|
}
|
|
|
|
// Call pcloud file_close, see [API Doc.]
|
|
// [API Doc]: https://docs.pcloud.com/methods/fileops/file_close.html
|
|
func fileClose(
|
|
ctx context.Context,
|
|
client *rest.Client,
|
|
pacer *fs.Pacer,
|
|
fd int64,
|
|
) (*api.FileCloseResponse, error) {
|
|
opts := rest.Opts{
|
|
Method: "PUT",
|
|
Path: "/file_close",
|
|
Parameters: url.Values{},
|
|
TransferEncoding: []string{"identity"}, // pcloud doesn't like chunked encoding
|
|
Close: true,
|
|
}
|
|
opts.Parameters.Set("fd", strconv.FormatInt(fd, 10))
|
|
|
|
result := &api.FileCloseResponse{}
|
|
err := pacer.CallNoRetry(func() (bool, error) {
|
|
resp, err := client.CallJSON(ctx, &opts, nil, result)
|
|
err = result.Error.Update(err)
|
|
return shouldRetry(ctx, resp, err)
|
|
})
|
|
if err != nil {
|
|
return nil, fmt.Errorf("close file descriptor: %w", err)
|
|
}
|
|
return result, nil
|
|
}
|