// This keeps track of the files which need to be written back package vfscache import ( "container/heap" "context" "sync" "time" "github.com/rclone/rclone/fs" "github.com/rclone/rclone/fs/fserrors" "github.com/rclone/rclone/vfs/vfscommon" ) const ( maxUploadDelay = 5 * time.Minute // max delay betwen upload attempts ) // putFn is the interface that item provides to store the data type putFn func(context.Context) error // writeBack keeps track of the items which need to be written back to the disk at some point type writeBack struct { ctx context.Context mu sync.Mutex items writeBackItems // priority queue of *writeBackItem - writeBackItems are in here while awaiting transfer only lookup map[*Item]*writeBackItem // for getting a *writeBackItem from a *Item - writeBackItems are in here until cancelled opt *vfscommon.Options // VFS options timer *time.Timer // next scheduled time for the uploader expiry time.Time // time the next item exires or IsZero uploads int // number of uploads in progress id uint64 // id of the last writeBackItem created } // make a new writeBack // // cancel the context to stop the background goroutine func newWriteBack(ctx context.Context, opt *vfscommon.Options) *writeBack { wb := &writeBack{ ctx: ctx, items: writeBackItems{}, lookup: make(map[*Item]*writeBackItem), opt: opt, } heap.Init(&wb.items) return wb } // writeBackItem stores an Item awaiting writeback // // These are stored on the items heap when awaiting transfer but // removed from the items heap when transferring. They remain in the // lookup map until cancelled. // // writeBack.mu must be held to manipulate this type writeBackItem struct { name string // name of the item so we don't have to read it from item id uint64 // id of the item index int // index into the priority queue for update item *Item // Item that needs writeback expiry time.Time // When this expires we will write it back uploading bool // True if item is being processed by upload() method onHeap bool // true if this item is on the items heap cancel context.CancelFunc // To cancel the upload with done chan struct{} // closed when the cancellation completes putFn putFn // To write the object data tries int // number of times we have tried to upload delay time.Duration // delay between upload attempts } // A writeBackItems implements a priority queue by implementing // heap.Interface and holds writeBackItems. type writeBackItems []*writeBackItem func (ws writeBackItems) Len() int { return len(ws) } func (ws writeBackItems) Less(i, j int) bool { a, b := ws[i], ws[j] // If times are equal then use ID to disambiguate if a.expiry.Equal(b.expiry) { return a.id < b.id } return a.expiry.Before(b.expiry) } func (ws writeBackItems) Swap(i, j int) { ws[i], ws[j] = ws[j], ws[i] ws[i].index = i ws[j].index = j } func (ws *writeBackItems) Push(x interface{}) { n := len(*ws) item := x.(*writeBackItem) item.index = n *ws = append(*ws, item) } func (ws *writeBackItems) Pop() interface{} { old := *ws n := len(old) item := old[n-1] old[n-1] = nil // avoid memory leak item.index = -1 // for safety *ws = old[0 : n-1] return item } // update modifies the expiry of an Item in the queue. // // call with lock held func (ws *writeBackItems) _update(item *writeBackItem, expiry time.Time) { item.expiry = expiry heap.Fix(ws, item.index) } // return a new expiry time based from now until the WriteBack timeout // // call with lock held func (wb *writeBack) _newExpiry() time.Time { expiry := time.Now() if wb.opt.WriteBack > 0 { expiry = expiry.Add(wb.opt.WriteBack) } // expiry = expiry.Round(time.Millisecond) return expiry } // make a new writeBackItem // // call with the lock held func (wb *writeBack) _newItem(item *Item, name string) *writeBackItem { wb.id++ wbItem := &writeBackItem{ name: name, item: item, expiry: wb._newExpiry(), delay: wb.opt.WriteBack, id: wb.id, } wb._addItem(wbItem) wb._pushItem(wbItem) return wbItem } // add a writeBackItem to the lookup map // // call with the lock held func (wb *writeBack) _addItem(wbItem *writeBackItem) { wb.lookup[wbItem.item] = wbItem } // delete a writeBackItem from the lookup map // // call with the lock held func (wb *writeBack) _delItem(wbItem *writeBackItem) { delete(wb.lookup, wbItem.item) } // pop a writeBackItem from the items heap // // call with the lock held func (wb *writeBack) _popItem() (wbItem *writeBackItem) { wbItem = heap.Pop(&wb.items).(*writeBackItem) wbItem.onHeap = false return wbItem } // push a writeBackItem onto the items heap // // call with the lock held func (wb *writeBack) _pushItem(wbItem *writeBackItem) { if !wbItem.onHeap { heap.Push(&wb.items, wbItem) wbItem.onHeap = true } } // remove a writeBackItem from the items heap // // call with the lock held func (wb *writeBack) _removeItem(wbItem *writeBackItem) { if wbItem.onHeap { heap.Remove(&wb.items, wbItem.index) wbItem.onHeap = false } } // peek the oldest writeBackItem - may be nil // // call with the lock held func (wb *writeBack) _peekItem() (wbItem *writeBackItem) { if len(wb.items) == 0 { return nil } return wb.items[0] } // stop the timer which runs the expiries func (wb *writeBack) _stopTimer() { if wb.expiry.IsZero() { return } wb.expiry = time.Time{} fs.Debugf(nil, "resetTimer STOP") if wb.timer != nil { wb.timer.Stop() wb.timer = nil } } // reset the timer which runs the expiries func (wb *writeBack) _resetTimer() { wbItem := wb._peekItem() if wbItem == nil { wb._stopTimer() } else { if wb.expiry.Equal(wbItem.expiry) { return } wb.expiry = wbItem.expiry dt := time.Until(wbItem.expiry) if dt < 0 { dt = 0 } fs.Debugf(nil, "resetTimer dt=%v", dt) if wb.timer != nil { wb.timer.Stop() } wb.timer = time.AfterFunc(dt, func() { wb.processItems(wb.ctx) }) } } // add adds an item to the writeback queue or resets its timer if it // is already there. // // if modified is false then it it doesn't a pending upload func (wb *writeBack) add(item *Item, name string, modified bool, putFn putFn) *writeBackItem { wb.mu.Lock() defer wb.mu.Unlock() wbItem, ok := wb.lookup[item] if !ok { wbItem = wb._newItem(item, name) } else { if wbItem.uploading && modified { // We are uploading already so cancel the upload wb._cancelUpload(wbItem) } // Kick the timer on wb.items._update(wbItem, wb._newExpiry()) } wbItem.putFn = putFn wb._resetTimer() return wbItem } // Call when a file is removed. This cancels a writeback if there is // one and doesn't return the item to the queue. func (wb *writeBack) remove(item *Item) (found bool) { wb.mu.Lock() defer wb.mu.Unlock() wbItem, found := wb.lookup[item] if found { fs.Debugf(wbItem.name, "vfs cache: cancelling writeback (uploading %v) %p item %p", wbItem.uploading, wbItem, wbItem.item) if wbItem.uploading { // We are uploading already so cancel the upload wb._cancelUpload(wbItem) } // Remove the item from the heap wb._removeItem(wbItem) // Remove the item from the lookup map wb._delItem(wbItem) } wb._resetTimer() return found } // upload the item - called as a goroutine // // uploading will have been incremented here already func (wb *writeBack) upload(ctx context.Context, wbItem *writeBackItem) { wb.mu.Lock() defer wb.mu.Unlock() putFn := wbItem.putFn wbItem.tries++ wb.mu.Unlock() err := putFn(ctx) wb.mu.Lock() wbItem.cancel() // cancel context to release resources since store done //fs.Debugf(wbItem.name, "uploading = false %p item %p", wbItem, wbItem.item) wbItem.uploading = false wb.uploads-- if err != nil { // FIXME should this have a max number of transfer attempts? wbItem.delay *= 2 if wbItem.delay > maxUploadDelay { wbItem.delay = maxUploadDelay } if _, uerr := fserrors.Cause(err); uerr == context.Canceled { fs.Infof(wbItem.name, "vfs cache: upload canceled sucessfully") // Upload was cancelled so reset timer wbItem.delay = wb.opt.WriteBack } else { fs.Errorf(wbItem.name, "vfs cache: failed to upload try #%d, will retry in %v: %v", wbItem.tries, wbItem.delay, err) } // push the item back on the queue for retry wb._pushItem(wbItem) wb.items._update(wbItem, time.Now().Add(wbItem.delay)) } else { fs.Infof(wbItem.name, "vfs cache: upload succeeded try #%d", wbItem.tries) // show that we are done with the item wb._delItem(wbItem) } wb._resetTimer() close(wbItem.done) } // cancel the upload - the item should be on the heap after this returns // // call with lock held func (wb *writeBack) _cancelUpload(wbItem *writeBackItem) { if !wbItem.uploading { return } fs.Infof(wbItem.name, "vfs cache: cancelling upload") if wbItem.cancel != nil { // Cancel the upload - this may or may not be effective wbItem.cancel() // wait for the uploader to finish // // we need to wait without the lock otherwise the // background part will never run. wb.mu.Unlock() <-wbItem.done wb.mu.Lock() } // uploading items are not on the heap so add them back wb._pushItem(wbItem) fs.Infof(wbItem.name, "vfs cache: cancelled upload") } // cancelUpload cancels the upload of the item if there is one in progress // // it returns true if there was an upload in progress func (wb *writeBack) cancelUpload(item *Item) bool { wb.mu.Lock() defer wb.mu.Unlock() wbItem, ok := wb.lookup[item] if !ok || !wbItem.uploading { return false } wb._cancelUpload(wbItem) return true } // this uploads as many items as possible func (wb *writeBack) processItems(ctx context.Context) { wb.mu.Lock() defer wb.mu.Unlock() if wb.ctx.Err() != nil { return } resetTimer := true for wbItem := wb._peekItem(); wbItem != nil && time.Until(wbItem.expiry) <= 0; wbItem = wb._peekItem() { // If reached transfer limit don't restart the timer if wb.uploads >= fs.Config.Transfers { fs.Debugf(wbItem.name, "vfs cache: delaying writeback as --transfers exceeded") resetTimer = false break } // Pop the item, mark as uploading and start the uploader wbItem = wb._popItem() //fs.Debugf(wbItem.name, "uploading = true %p item %p", wbItem, wbItem.item) wbItem.uploading = true wb.uploads++ newCtx, cancel := context.WithCancel(ctx) wbItem.cancel = cancel wbItem.done = make(chan struct{}) go wb.upload(newCtx, wbItem) } if resetTimer { wb._resetTimer() } else { wb._stopTimer() } } // return the number of uploads in progress func (wb *writeBack) getStats() (uploadsInProgress, uploadsQueued int) { wb.mu.Lock() defer wb.mu.Unlock() return wb.uploads, len(wb.items) }