mirror of
https://github.com/rclone/rclone.git
synced 2024-12-04 06:16:39 +01:00
e7bd392a69
This fixes an important bug with listing that affects users with more than 500 objects in a listing operation.
281 lines
5.9 KiB
Go
281 lines
5.9 KiB
Go
// Copyright (C) 2019 Storj Labs, Inc.
|
|
// See LICENSE for copying information
|
|
|
|
package sync2
|
|
|
|
import (
|
|
"fmt"
|
|
"io"
|
|
"io/ioutil"
|
|
"math"
|
|
"sync"
|
|
)
|
|
|
|
// pipe is a io.Reader/io.Writer pipe backed by ReadAtWriteAtCloser.
|
|
type pipe struct {
|
|
noCopy noCopy // nolint: structcheck
|
|
|
|
buffer ReadAtWriteAtCloser
|
|
|
|
mu sync.Mutex
|
|
nodata sync.Cond
|
|
read int64
|
|
write int64
|
|
limit int64
|
|
|
|
writerDone bool
|
|
writerErr error
|
|
|
|
readerDone bool
|
|
readerErr error
|
|
}
|
|
|
|
// NewPipeFile returns a pipe that uses file-system to offload memory.
|
|
func NewPipeFile(tempdir string) (PipeReader, PipeWriter, error) {
|
|
tempfile, err := ioutil.TempFile(tempdir, "filepipe")
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
|
|
handles := int64(2)
|
|
pipe := &pipe{
|
|
buffer: &offsetFile{
|
|
file: tempfile,
|
|
open: &handles,
|
|
},
|
|
limit: math.MaxInt64,
|
|
}
|
|
pipe.nodata.L = &pipe.mu
|
|
|
|
return pipeReader{pipe}, pipeWriter{pipe}, nil
|
|
}
|
|
|
|
// NewPipeMemory returns a pipe that uses an in-memory buffer.
|
|
func NewPipeMemory(pipeSize int64) (PipeReader, PipeWriter, error) {
|
|
pipe := &pipe{
|
|
buffer: make(memory, pipeSize),
|
|
limit: pipeSize,
|
|
}
|
|
pipe.nodata.L = &pipe.mu
|
|
return pipeReader{pipe}, pipeWriter{pipe}, nil
|
|
}
|
|
|
|
type pipeReader struct{ pipe *pipe }
|
|
type pipeWriter struct{ pipe *pipe }
|
|
|
|
// Close implements io.Reader Close.
|
|
func (reader pipeReader) Close() error { return reader.CloseWithError(nil) }
|
|
|
|
// Close implements io.Writer Close.
|
|
func (writer pipeWriter) Close() error { return writer.CloseWithError(nil) }
|
|
|
|
// CloseWithError implements closing with error.
|
|
func (reader pipeReader) CloseWithError(err error) error {
|
|
if err == nil {
|
|
err = io.ErrClosedPipe
|
|
}
|
|
|
|
pipe := reader.pipe
|
|
pipe.mu.Lock()
|
|
if pipe.readerDone {
|
|
pipe.mu.Unlock()
|
|
return io.ErrClosedPipe
|
|
}
|
|
pipe.readerDone = true
|
|
pipe.readerErr = err
|
|
pipe.mu.Unlock()
|
|
|
|
return pipe.buffer.Close()
|
|
}
|
|
|
|
// CloseWithError implements closing with error.
|
|
func (writer pipeWriter) CloseWithError(err error) error {
|
|
if err == nil {
|
|
err = io.EOF
|
|
}
|
|
|
|
pipe := writer.pipe
|
|
pipe.mu.Lock()
|
|
if pipe.writerDone {
|
|
pipe.mu.Unlock()
|
|
return io.ErrClosedPipe
|
|
}
|
|
pipe.writerDone = true
|
|
pipe.writerErr = err
|
|
pipe.nodata.Broadcast()
|
|
pipe.mu.Unlock()
|
|
|
|
return pipe.buffer.Close()
|
|
}
|
|
|
|
// Write writes to the pipe returning io.ErrClosedPipe when pipeSize is reached.
|
|
func (writer pipeWriter) Write(data []byte) (n int, err error) {
|
|
pipe := writer.pipe
|
|
pipe.mu.Lock()
|
|
|
|
// has the reader finished?
|
|
if pipe.readerDone {
|
|
pipe.mu.Unlock()
|
|
return 0, pipe.readerErr
|
|
}
|
|
|
|
// have we closed already
|
|
if pipe.writerDone {
|
|
pipe.mu.Unlock()
|
|
return 0, io.ErrClosedPipe
|
|
}
|
|
|
|
// check how much do they want to write
|
|
canWrite := pipe.limit - pipe.write
|
|
|
|
// no more room to write
|
|
if canWrite == 0 {
|
|
pipe.mu.Unlock()
|
|
return 0, io.ErrClosedPipe
|
|
}
|
|
|
|
// figure out how much to write
|
|
toWrite := int64(len(data))
|
|
if toWrite > canWrite {
|
|
toWrite = canWrite
|
|
}
|
|
|
|
writeAt := pipe.write
|
|
pipe.mu.Unlock()
|
|
|
|
// write data to buffer
|
|
writeAmount, err := pipe.buffer.WriteAt(data[:toWrite], writeAt)
|
|
|
|
pipe.mu.Lock()
|
|
// update writing head
|
|
pipe.write += int64(writeAmount)
|
|
// wake up reader
|
|
pipe.nodata.Broadcast()
|
|
// check whether we have finished
|
|
done := pipe.write >= pipe.limit
|
|
pipe.mu.Unlock()
|
|
|
|
if err == nil && done {
|
|
err = io.ErrClosedPipe
|
|
}
|
|
return writeAmount, err
|
|
}
|
|
|
|
// Read reads from the pipe returning io.EOF when writer is closed or pipeSize is reached.
|
|
func (reader pipeReader) Read(data []byte) (n int, err error) {
|
|
pipe := reader.pipe
|
|
pipe.mu.Lock()
|
|
// wait until we have something to read
|
|
for pipe.read >= pipe.write {
|
|
// has the writer finished?
|
|
if pipe.writerDone {
|
|
pipe.mu.Unlock()
|
|
return 0, pipe.writerErr
|
|
}
|
|
|
|
// have we closed already
|
|
if pipe.readerDone {
|
|
pipe.mu.Unlock()
|
|
return 0, io.ErrClosedPipe
|
|
}
|
|
|
|
// have we run out of the limit
|
|
if pipe.read >= pipe.limit {
|
|
pipe.mu.Unlock()
|
|
return 0, io.EOF
|
|
}
|
|
|
|
// ok, lets wait
|
|
pipe.nodata.Wait()
|
|
}
|
|
|
|
// how much there's available for reading
|
|
canRead := pipe.write - pipe.read
|
|
// how much do they want to read?
|
|
toRead := int64(len(data))
|
|
if toRead > canRead {
|
|
toRead = canRead
|
|
}
|
|
readAt := pipe.read
|
|
pipe.mu.Unlock()
|
|
|
|
// read data
|
|
readAmount, err := pipe.buffer.ReadAt(data[:toRead], readAt)
|
|
|
|
pipe.mu.Lock()
|
|
// update info on how much we have read
|
|
pipe.read += int64(readAmount)
|
|
done := pipe.read >= pipe.limit
|
|
pipe.mu.Unlock()
|
|
|
|
if err == nil && done {
|
|
err = io.EOF
|
|
}
|
|
return readAmount, err
|
|
}
|
|
|
|
// MultiPipe is a multipipe backed by a single file.
|
|
type MultiPipe struct {
|
|
pipes []pipe
|
|
}
|
|
|
|
// NewMultiPipeFile returns a new MultiPipe that is created in tempdir
|
|
// if tempdir == "" the fill will be created it into os.TempDir.
|
|
func NewMultiPipeFile(tempdir string, pipeCount, pipeSize int64) (*MultiPipe, error) {
|
|
tempfile, err := ioutil.TempFile(tempdir, "multifilepipe")
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
err = tempfile.Truncate(pipeCount * pipeSize)
|
|
if err != nil {
|
|
closeErr := tempfile.Close()
|
|
if closeErr != nil {
|
|
return nil, fmt.Errorf("%v/%v", err, closeErr)
|
|
}
|
|
return nil, err
|
|
}
|
|
|
|
multipipe := &MultiPipe{
|
|
pipes: make([]pipe, pipeCount),
|
|
}
|
|
|
|
handles := 2 * pipeCount
|
|
for i := range multipipe.pipes {
|
|
pipe := &multipipe.pipes[i]
|
|
pipe.buffer = offsetFile{
|
|
file: tempfile,
|
|
offset: int64(i) * pipeSize,
|
|
open: &handles,
|
|
}
|
|
pipe.limit = pipeSize
|
|
pipe.nodata.L = &pipe.mu
|
|
}
|
|
|
|
return multipipe, nil
|
|
}
|
|
|
|
// NewMultiPipeMemory returns a new MultiPipe that is using a memory buffer.
|
|
func NewMultiPipeMemory(pipeCount, pipeSize int64) (*MultiPipe, error) {
|
|
buffer := make(memory, pipeCount*pipeSize)
|
|
|
|
multipipe := &MultiPipe{
|
|
pipes: make([]pipe, pipeCount),
|
|
}
|
|
|
|
for i := range multipipe.pipes {
|
|
pipe := &multipipe.pipes[i]
|
|
pipe.buffer = buffer[i*int(pipeSize) : (i+1)*int(pipeSize)]
|
|
pipe.limit = pipeSize
|
|
pipe.nodata.L = &pipe.mu
|
|
}
|
|
|
|
return multipipe, nil
|
|
}
|
|
|
|
// Pipe returns the two ends of a block stream pipe.
|
|
func (multipipe *MultiPipe) Pipe(index int) (PipeReader, PipeWriter) {
|
|
pipe := &multipipe.pipes[index]
|
|
return pipeReader{pipe}, pipeWriter{pipe}
|
|
}
|