// Package ulozto provides an interface to the Uloz.to storage system. package ulozto import ( "bytes" "context" "encoding/base64" "encoding/gob" "encoding/hex" "fmt" "io" "net/http" "net/url" "path" "strconv" "strings" "time" "github.com/pkg/errors" "github.com/rclone/rclone/backend/ulozto/api" "github.com/rclone/rclone/fs" "github.com/rclone/rclone/fs/config" "github.com/rclone/rclone/fs/config/configmap" "github.com/rclone/rclone/fs/config/configstruct" "github.com/rclone/rclone/fs/fserrors" "github.com/rclone/rclone/fs/fshttp" "github.com/rclone/rclone/fs/hash" "github.com/rclone/rclone/lib/dircache" "github.com/rclone/rclone/lib/encoder" "github.com/rclone/rclone/lib/pacer" "github.com/rclone/rclone/lib/rest" ) // TODO Uloz.to only supports file names of 255 characters or less and silently truncates names that are longer. const ( minSleep = 10 * time.Millisecond maxSleep = 2 * time.Second decayConstant = 2 // bigger for slower decay, exponential rootURL = "https://apis.uloz.to" // TODO temporary limitation, remove with chunked upload impl maxFileSizeBytes = 2500 * 1024 * 1024 ) // Options defines the configuration for this backend type Options struct { AppToken string `config:"app_token"` Username string `config:"username"` Password string `config:"password"` RootFolderSlug string `config:"root_folder_slug"` Enc encoder.MultiEncoder `config:"encoding"` ListPageSize int `config:"list_page_size"` } func init() { fs.Register(&fs.RegInfo{ Name: "ulozto", Description: "Uloz.to", NewFs: NewFs, Options: []fs.Option{ { Name: "app_token", Default: "", Help: "The application token identifying the app. An app API key can be either found in the API " + "doc https://uloz.to/upload-resumable-api-beta or obtained from customer service.", Sensitive: true, }, { Name: "username", Default: "", Help: "The username of the principal to operate as.", Sensitive: true, }, { Name: "password", Default: "", Help: "The password for the user.", Sensitive: true, }, { Name: "root_folder_slug", Help: "If set, rclone will use this folder as the root folder for all operations. For example, " + "if the slug identifies 'foo/bar/', 'ulozto:baz' is equivalent to 'ulozto:foo/bar/baz' without " + "any root slug set.", Default: "", Advanced: true, Sensitive: true, }, { Name: "list_page_size", Default: 500, Help: "The size of a single page for list commands. 1-500", Advanced: true, }, { Name: config.ConfigEncoding, Help: config.ConfigEncodingHelp, Advanced: true, Default: encoder.Display | encoder.EncodeInvalidUtf8 | encoder.EncodeBackSlash, }, }}) } // Fs represents a remote uloz.to storage type Fs struct { name string // name of this remote root string // the path we are working on opt Options // parsed options features *fs.Features // optional features rest *rest.Client // REST client with authentication headers set, used to communicate with API endpoints cdn *rest.Client // REST client without authentication headers set, used for CDN payload upload/download dirCache *dircache.DirCache // Map of directory path to directory id pacer *fs.Pacer // pacer for API calls } // NewFs constructs a Fs from the path, container:path func NewFs(ctx context.Context, name, root string, m configmap.Mapper) (fs.Fs, error) { // Parse config into Options struct opt := new(Options) err := configstruct.Set(m, opt) if err != nil { return nil, err } client := fshttp.NewClient(ctx) f := &Fs{ name: name, root: root, opt: *opt, cdn: rest.NewClient(client), rest: rest.NewClient(client).SetRoot(rootURL), pacer: fs.NewPacer(ctx, pacer.NewDefault(pacer.MinSleep(minSleep), pacer.MaxSleep(maxSleep), pacer.DecayConstant(decayConstant))), } f.features = (&fs.Features{ DuplicateFiles: true, CanHaveEmptyDirectories: true, }).Fill(ctx, f) f.rest.SetErrorHandler(errorHandler) f.rest.SetHeader("X-Auth-Token", f.opt.AppToken) auth, err := f.authenticate(ctx) if err != nil { return f, err } var rootSlug string if opt.RootFolderSlug == "" { rootSlug = auth.Session.User.RootFolderSlug } else { rootSlug = opt.RootFolderSlug } f.dirCache = dircache.New(root, rootSlug, f) err = f.dirCache.FindRoot(ctx, false) if errors.Is(err, fs.ErrorDirNotFound) { // All good, we'll create the folder later on. return f, nil } if errors.Is(err, fs.ErrorIsFile) { rootFolder, _ := dircache.SplitPath(root) f.root = rootFolder f.dirCache = dircache.New(rootFolder, rootSlug, f) err = f.dirCache.FindRoot(ctx, false) if err != nil { return f, err } return f, fs.ErrorIsFile } return f, err } // errorHandler parses a non 2xx error response into an error func errorHandler(resp *http.Response) error { // Decode error response errResponse := new(api.Error) err := rest.DecodeJSON(resp, &errResponse) if err != nil { fs.Debugf(nil, "Couldn't decode error response: %v", err) } if errResponse.StatusCode == 0 { errResponse.StatusCode = resp.StatusCode } return errors.WithStack(errResponse) } // retryErrorCodes is a slice of error codes that we will retry var retryErrorCodes = []int{ 429, // Too Many Requests. // TODO: random 500s should be retried but the error code corresponds to a known issue with uploading large files, // leading to numerous (slow & resource consuming) retries. Don't retry them until the root cause is addressed. // 500, // Internal Server Error 502, // Bad Gateway 503, // Service Unavailable 504, // Gateway Timeout } // shouldRetry returns a boolean whether this resp and err should be retried. // It also returns the err for convenience. func (f *Fs) shouldRetry(ctx context.Context, resp *http.Response, err error, reauth bool) (bool, error) { if err == nil { return false, nil } if fserrors.ContextError(ctx, &err) { return false, err } var apiErr *api.Error if resp != nil && resp.StatusCode == 401 && errors.As(err, &apiErr) && apiErr.ErrorCode == 70001 { fs.Debugf(nil, "Should retry: %v", err) if reauth { _, err = f.authenticate(ctx) if err != nil { return false, err } } return true, err } return fserrors.ShouldRetry(err) || fserrors.ShouldRetryHTTP(resp, retryErrorCodes), err } func (f *Fs) authenticate(ctx context.Context) (response *api.AuthenticateResponse, err error) { // TODO only reauth once if the token expires // Remove the old user token f.rest.RemoveHeader("X-User-Token") opts := rest.Opts{ Method: "PUT", Path: "/v6/session", } authRequest := api.AuthenticateRequest{ Login: f.opt.Username, Password: f.opt.Password, } err = f.pacer.Call(func() (bool, error) { httpResp, err := f.rest.CallJSON(ctx, &opts, &authRequest, &response) return f.shouldRetry(ctx, httpResp, err, false) }) if err != nil { return nil, err } f.rest.SetHeader("X-User-Token", response.TokenID) return response, nil } // UploadSession represents a single Uloz.to upload session. // // Uloz.to supports uploading multiple files at once and committing them atomically. This functionality isn't being used // by the backend implementation and for simplicity, each session corresponds to a single file being uploaded. type UploadSession struct { Filesystem *Fs URL string PrivateSlug string ValidUntil time.Time } func (f *Fs) createUploadSession(ctx context.Context) (session *UploadSession, err error) { session = &UploadSession{ Filesystem: f, } err = session.renewUploadSession(ctx) if err != nil { return nil, err } return session, nil } func (session *UploadSession) renewUploadSession(ctx context.Context) error { opts := rest.Opts{ Method: "POST", Path: "/v5/upload/link", Parameters: url.Values{}, } createUploadURLReq := api.CreateUploadURLRequest{ UserLogin: session.Filesystem.opt.Username, Realm: "ulozto", } if session.PrivateSlug != "" { createUploadURLReq.ExistingSessionSlug = session.PrivateSlug } var err error var response api.CreateUploadURLResponse err = session.Filesystem.pacer.Call(func() (bool, error) { httpResp, err := session.Filesystem.rest.CallJSON(ctx, &opts, &createUploadURLReq, &response) return session.Filesystem.shouldRetry(ctx, httpResp, err, true) }) if err != nil { return err } session.PrivateSlug = response.PrivateSlug session.URL = response.UploadURL session.ValidUntil = response.ValidUntil return nil } func (f *Fs) uploadUnchecked(ctx context.Context, name, parentSlug string, info fs.ObjectInfo, payload io.Reader) (fs.Object, error) { session, err := f.createUploadSession(ctx) if err != nil { return nil, err } hashes := hash.NewHashSet(hash.MD5, hash.SHA256) hasher, err := hash.NewMultiHasherTypes(hashes) if err != nil { return nil, err } payload = io.TeeReader(payload, hasher) encodedName := f.opt.Enc.FromStandardName(name) opts := rest.Opts{ Method: "POST", Body: payload, // Not using Parameters as the session URL has parameters itself RootURL: session.URL + "&batch_file_id=1&is_porn=false", MultipartContentName: "file", MultipartFileName: encodedName, Parameters: url.Values{}, } var uploadResponse api.SendFilePayloadResponse err = f.pacer.CallNoRetry(func() (bool, error) { httpResp, err := f.cdn.CallJSON(ctx, &opts, nil, &uploadResponse) return f.shouldRetry(ctx, httpResp, err, true) }) if err != nil { return nil, err } sha256digest, err := hasher.Sum(hash.SHA256) if err != nil { return nil, err } md5digest, err := hasher.Sum(hash.MD5) if err != nil { return nil, err } if hex.EncodeToString(md5digest) != uploadResponse.Md5 { return nil, errors.New("MD5 digest mismatch") } metadata := DescriptionEncodedMetadata{ Md5Hash: md5digest, Sha256Hash: sha256digest, ModTimeEpochMicros: info.ModTime(ctx).UnixMicro(), } encodedMetadata, err := metadata.encode() if err != nil { return nil, err } // Successfully uploaded, now move the file where it belongs and commit it updateReq := api.BatchUpdateFilePropertiesRequest{ Name: encodedName, FolderSlug: parentSlug, Description: encodedMetadata, Slugs: []string{uploadResponse.Slug}, UploadTokens: map[string]string{uploadResponse.Slug: session.PrivateSlug + ":1"}, } var updateResponse []api.File opts = rest.Opts{ Method: "PATCH", Path: "/v8/file-list/private", Parameters: url.Values{}, } err = f.pacer.Call(func() (bool, error) { httpResp, err := session.Filesystem.rest.CallJSON(ctx, &opts, &updateReq, &updateResponse) return f.shouldRetry(ctx, httpResp, err, true) }) if err != nil { return nil, err } if len(updateResponse) != 1 { return nil, errors.New("unexpected number of files in the response") } opts = rest.Opts{ Method: "PATCH", Path: "/v8/upload-batch/private/" + session.PrivateSlug, Parameters: url.Values{}, } commitRequest := api.CommitUploadBatchRequest{ Status: "confirmed", OwnerLogin: f.opt.Username, } var commitResponse api.CommitUploadBatchResponse err = f.pacer.Call(func() (bool, error) { httpResp, err := session.Filesystem.rest.CallJSON(ctx, &opts, &commitRequest, &commitResponse) return f.shouldRetry(ctx, httpResp, err, true) }) if err != nil { return nil, err } file, err := f.newObjectWithInfo(ctx, info.Remote(), &updateResponse[0]) return file, err } // Put implements the mandatory method fs.Fs.Put. func (f *Fs) Put(ctx context.Context, in io.Reader, src fs.ObjectInfo, options ...fs.OpenOption) (fs.Object, error) { // TODO: workaround for uloz.to's bug. Remove when chunked upload support is implemented. if src.Size() > maxFileSizeBytes { return nil, errors.New("file size over the supported max threshold") } existingObj, err := f.NewObject(ctx, src.Remote()) switch { case err == nil: return existingObj, existingObj.Update(ctx, in, src, options...) case errors.Is(err, fs.ErrorObjectNotFound): // Not found so create it return f.PutUnchecked(ctx, in, src, options...) default: return nil, err } } // PutUnchecked implements the optional interface fs.PutUncheckeder. // // Uloz.to allows to have multiple files of the same name in the same folder. func (f *Fs) PutUnchecked(ctx context.Context, in io.Reader, src fs.ObjectInfo, options ...fs.OpenOption) (fs.Object, error) { filename, folderSlug, err := f.dirCache.FindPath(ctx, src.Remote(), true) if err != nil { return nil, err } return f.uploadUnchecked(ctx, filename, folderSlug, src, in) } // Mkdir implements the mandatory method fs.Fs.Mkdir. func (f *Fs) Mkdir(ctx context.Context, dir string) (err error) { _, err = f.dirCache.FindDir(ctx, dir, true) return err } func (f *Fs) isDirEmpty(ctx context.Context, slug string) (empty bool, err error) { folders, err := f.fetchListFolderPage(ctx, slug, "", 1, 0) if err != nil { return false, err } if len(folders) > 0 { return false, nil } files, err := f.fetchListFilePage(ctx, slug, "", 1, 0) if err != nil { return false, err } if len(files) > 0 { return false, nil } return true, nil } // Rmdir implements the mandatory method fs.Fs.Rmdir. func (f *Fs) Rmdir(ctx context.Context, dir string) error { slug, err := f.dirCache.FindDir(ctx, dir, false) if err != nil { return err } empty, err := f.isDirEmpty(ctx, slug) if err != nil { return err } if !empty { return fs.ErrorDirectoryNotEmpty } opts := rest.Opts{ Method: "DELETE", Path: "/v5/user/" + f.opt.Username + "/folder-list", } req := api.DeleteFoldersRequest{Slugs: []string{slug}} err = f.pacer.Call(func() (bool, error) { httpResp, err := f.rest.CallJSON(ctx, &opts, req, nil) return f.shouldRetry(ctx, httpResp, err, true) }) if err != nil { return err } f.dirCache.FlushDir(dir) return nil } // Name of the remote (as passed into NewFs) func (f *Fs) Name() string { return f.name } // Root of the remote (as passed into NewFs) func (f *Fs) Root() string { return f.root } // String converts this Fs to a string func (f *Fs) String() string { return fmt.Sprintf("uloz.to root '%s'", f.root) } // Features returns the optional features of this Fs func (f *Fs) Features() *fs.Features { return f.features } // Precision return the precision of this Fs func (f *Fs) Precision() time.Duration { return time.Microsecond } // Hashes implements fs.Fs.Hashes by returning the supported hash types of the filesystem. func (f *Fs) Hashes() hash.Set { return hash.NewHashSet(hash.SHA256, hash.MD5) } // DescriptionEncodedMetadata represents a set of metadata encoded as Uloz.to description. // // Uloz.to doesn't support setting metadata such as mtime but allows the user to set an arbitrary description field. // The content of this structure will be serialized and stored in the backend. // // The files themselves are immutable so there's no danger that the file changes, and we'll forget to update the hashes. // It is theoretically possible to rewrite the description to provide incorrect information for a file. However, in case // it's a real attack vector, a nefarious person already has write access to the repo, and the situation is above // rclone's pay grade already. type DescriptionEncodedMetadata struct { Md5Hash []byte // The MD5 hash of the file Sha256Hash []byte // The SHA256 hash of the file ModTimeEpochMicros int64 // The mtime of the file, as set by rclone } func (md *DescriptionEncodedMetadata) encode() (string, error) { b := bytes.Buffer{} e := gob.NewEncoder(&b) err := e.Encode(md) if err != nil { return "", err } // Version the encoded string from the beginning even though we don't need it yet. return "1;" + base64.StdEncoding.EncodeToString(b.Bytes()), nil } func decodeDescriptionMetadata(str string) (*DescriptionEncodedMetadata, error) { // The encoded data starts with a version number which is not a part iof the serialized object spl := strings.SplitN(str, ";", 2) if len(spl) < 2 || spl[0] != "1" { return nil, errors.New("can't decode, unknown encoded metadata version") } m := DescriptionEncodedMetadata{} by, err := base64.StdEncoding.DecodeString(spl[1]) if err != nil { return nil, err } b := bytes.Buffer{} b.Write(by) d := gob.NewDecoder(&b) err = d.Decode(&m) if err != nil { return nil, err } return &m, nil } // Object describes an uloz.to object. // // Valid objects will always have all fields but encodedMetadata set. type Object struct { fs *Fs // what this object is part of remote string // The remote path name string // The file name size int64 // size of the object slug string // ID of the object remoteFsMtime time.Time // The time the object was last modified in the remote fs. // Metadata not available natively and encoded in the description field. May not be present if the encoded metadata // is not present (e.g. if file wasn't uploaded by rclone) or invalid. encodedMetadata *DescriptionEncodedMetadata } // Storable implements the mandatory method fs.ObjectInfo.Storable func (o *Object) Storable() bool { return true } func (o *Object) updateFileProperties(ctx context.Context, req interface{}) (err error) { var resp *api.File opts := rest.Opts{ Method: "PATCH", Path: "/v8/file/" + o.slug + "/private", } err = o.fs.pacer.Call(func() (bool, error) { httpResp, err := o.fs.rest.CallJSON(ctx, &opts, &req, &resp) return o.fs.shouldRetry(ctx, httpResp, err, true) }) if err != nil { return err } return o.setMetaData(resp) } // SetModTime implements the mandatory method fs.Object.SetModTime func (o *Object) SetModTime(ctx context.Context, t time.Time) (err error) { var newMetadata DescriptionEncodedMetadata if o.encodedMetadata == nil { newMetadata = DescriptionEncodedMetadata{} } else { newMetadata = *o.encodedMetadata } newMetadata.ModTimeEpochMicros = t.UnixMicro() encoded, err := newMetadata.encode() if err != nil { return err } return o.updateFileProperties(ctx, api.UpdateDescriptionRequest{ Description: encoded, }) } // Open implements the mandatory method fs.Object.Open func (o *Object) Open(ctx context.Context, options ...fs.OpenOption) (rc io.ReadCloser, err error) { opts := rest.Opts{ Method: "POST", Path: "/v5/file/download-link/vipdata", } req := &api.GetDownloadLinkRequest{ Slug: o.slug, UserLogin: o.fs.opt.Username, // Has to be set but doesn't seem to be used server side. DeviceID: "foobar", } var resp *api.GetDownloadLinkResponse err = o.fs.pacer.Call(func() (bool, error) { httpResp, err := o.fs.rest.CallJSON(ctx, &opts, &req, &resp) return o.fs.shouldRetry(ctx, httpResp, err, true) }) if err != nil { return nil, err } opts = rest.Opts{ Method: "GET", RootURL: resp.Link, Options: options, } var httpResp *http.Response err = o.fs.pacer.Call(func() (bool, error) { httpResp, err = o.fs.cdn.Call(ctx, &opts) return o.fs.shouldRetry(ctx, httpResp, err, true) }) if err != nil { return nil, err } return httpResp.Body, err } func (o *Object) copyFrom(other *Object) { o.fs = other.fs o.remote = other.remote o.size = other.size o.slug = other.slug o.remoteFsMtime = other.remoteFsMtime o.encodedMetadata = other.encodedMetadata } // RenamingObjectInfoProxy is a delegating proxy for fs.ObjectInfo // with the option of specifying a different remote path. type RenamingObjectInfoProxy struct { delegate fs.ObjectInfo remote string } // Remote implements fs.ObjectInfo.Remote by delegating to the wrapped instance. func (s *RenamingObjectInfoProxy) String() string { return s.delegate.String() } // Remote implements fs.ObjectInfo.Remote by returning the specified remote path. func (s *RenamingObjectInfoProxy) Remote() string { return s.remote } // ModTime implements fs.ObjectInfo.ModTime by delegating to the wrapped instance. func (s *RenamingObjectInfoProxy) ModTime(ctx context.Context) time.Time { return s.delegate.ModTime(ctx) } // Size implements fs.ObjectInfo.Size by delegating to the wrapped instance. func (s *RenamingObjectInfoProxy) Size() int64 { return s.delegate.Size() } // Fs implements fs.ObjectInfo.Fs by delegating to the wrapped instance. func (s *RenamingObjectInfoProxy) Fs() fs.Info { return s.delegate.Fs() } // Hash implements fs.ObjectInfo.Hash by delegating to the wrapped instance. func (s *RenamingObjectInfoProxy) Hash(ctx context.Context, ty hash.Type) (string, error) { return s.delegate.Hash(ctx, ty) } // Storable implements fs.ObjectInfo.Storable by delegating to the wrapped instance. func (s *RenamingObjectInfoProxy) Storable() bool { return s.delegate.Storable() } // Update implements the mandatory method fs.Object.Update func (o *Object) Update(ctx context.Context, in io.Reader, src fs.ObjectInfo, options ...fs.OpenOption) error { // The backend allows to store multiple files with the same name, so simply upload the new file and remove the old // one afterwards. info := &RenamingObjectInfoProxy{ delegate: src, remote: o.Remote(), } newo, err := o.fs.PutUnchecked(ctx, in, info, options...) if err != nil { return err } err = o.Remove(ctx) if err != nil { return err } o.copyFrom(newo.(*Object)) return nil } // Remove implements the mandatory method fs.Object.Remove func (o *Object) Remove(ctx context.Context) error { for i := 0; i < 2; i++ { // First call moves the item to recycle bin, second deletes it for good var err error opts := rest.Opts{ Method: "DELETE", Path: "/v6/file/" + o.slug + "/private", } err = o.fs.pacer.Call(func() (bool, error) { httpResp, err := o.fs.rest.CallJSON(ctx, &opts, nil, nil) return o.fs.shouldRetry(ctx, httpResp, err, true) }) if err != nil { return err } } return nil } // ModTime implements the mandatory method fs.Object.ModTime func (o *Object) ModTime(ctx context.Context) time.Time { if o.encodedMetadata != nil { return time.UnixMicro(o.encodedMetadata.ModTimeEpochMicros) } // The time the object was last modified on the server - a handwavy guess, but we don't have any better return o.remoteFsMtime } // Fs implements the mandatory method fs.Object.Fs func (o *Object) Fs() fs.Info { return o.fs } // String returns the string representation of the remote object reference. func (o *Object) String() string { if o == nil { return "" } return o.remote } // Remote returns the remote path func (o *Object) Remote() string { return o.remote } // Size returns the size of an object in bytes func (o *Object) Size() int64 { return o.size } // Hash implements the mandatory method fs.Object.Hash. // // Supports SHA256 and MD5 hashes. func (o *Object) Hash(ctx context.Context, t hash.Type) (string, error) { if t != hash.MD5 && t != hash.SHA256 { return "", hash.ErrUnsupported } if o.encodedMetadata == nil { return "", nil } switch t { case hash.MD5: return hex.EncodeToString(o.encodedMetadata.Md5Hash), nil case hash.SHA256: return hex.EncodeToString(o.encodedMetadata.Sha256Hash), nil } panic("Should never get here") } // FindLeaf implements dircache.DirCacher.FindLeaf by successively walking through the folder hierarchy until // the desired folder is found, or there's nowhere to continue. func (f *Fs) FindLeaf(ctx context.Context, folderSlug, leaf string) (leafSlug string, found bool, err error) { folders, err := f.listFolders(ctx, folderSlug, leaf) if err != nil { if errors.Is(err, fs.ErrorDirNotFound) { return "", false, nil } return "", false, err } for _, folder := range folders { if folder.Name == leaf { return folder.Slug, true, nil } } // Uloz.to allows creation of multiple files / folders with the same name in the same parent folder. rclone always // expects folder paths to be unique (no other file or folder with the same name should exist). As a result we also // need to look at the files to return the correct error if necessary. files, err := f.listFiles(ctx, folderSlug, leaf) if err != nil { return "", false, err } for _, file := range files { if file.Name == leaf { return "", false, fs.ErrorIsFile } } // The parent folder exists but no file or folder with the given name was found in it. return "", false, nil } // CreateDir implements dircache.DirCacher.CreateDir by creating a folder with the given name under a folder identified // by parentSlug. func (f *Fs) CreateDir(ctx context.Context, parentSlug, leaf string) (newID string, err error) { var folder *api.Folder opts := rest.Opts{ Method: "POST", Path: "/v6/user/" + f.opt.Username + "/folder", Parameters: url.Values{}, } mkdir := api.CreateFolderRequest{ Name: f.opt.Enc.FromStandardName(leaf), ParentFolderSlug: parentSlug, } err = f.pacer.Call(func() (bool, error) { httpResp, err := f.rest.CallJSON(ctx, &opts, &mkdir, &folder) return f.shouldRetry(ctx, httpResp, err, true) }) if err != nil { return "", err } return folder.Slug, nil } func (f *Fs) newObjectWithInfo(ctx context.Context, remote string, info *api.File) (*Object, error) { o := &Object{ fs: f, remote: remote, } var err error if info == nil { info, err = f.readMetaDataForPath(ctx, remote) } if err != nil { return nil, err } err = o.setMetaData(info) if err != nil { return nil, err } return o, nil } // readMetaDataForPath reads the metadata from the path func (f *Fs) readMetaDataForPath(ctx context.Context, path string) (info *api.File, err error) { filename, folderSlug, err := f.dirCache.FindPath(ctx, path, false) if err != nil { if errors.Is(err, fs.ErrorDirNotFound) { return nil, fs.ErrorObjectNotFound } return nil, err } files, err := f.listFiles(ctx, folderSlug, filename) if err != nil { return nil, err } for _, file := range files { if file.Name == filename { return &file, nil } } folders, err := f.listFolders(ctx, folderSlug, filename) if err != nil { return nil, err } for _, file := range folders { if file.Name == filename { return nil, fs.ErrorIsDir } } return nil, fs.ErrorObjectNotFound } func (o *Object) setMetaData(info *api.File) (err error) { o.name = info.Name o.size = info.Filesize o.remoteFsMtime = info.LastUserModified o.encodedMetadata, err = decodeDescriptionMetadata(info.Description) if err != nil { fs.Debugf(o, "Couldn't decode metadata: %v", err) } o.slug = info.Slug return nil } // NewObject implements fs.Fs.NewObject. func (f *Fs) NewObject(ctx context.Context, remote string) (fs.Object, error) { return f.newObjectWithInfo(ctx, remote, nil) } // List implements fs.Fs.List by listing all files and folders in the given folder. func (f *Fs) List(ctx context.Context, dir string) (entries fs.DirEntries, err error) { folderSlug, err := f.dirCache.FindDir(ctx, dir, false) if err != nil { return nil, err } folders, err := f.listFolders(ctx, folderSlug, "") if err != nil { return nil, err } for _, folder := range folders { remote := path.Join(dir, folder.Name) f.dirCache.Put(remote, folder.Slug) entries = append(entries, fs.NewDir(remote, folder.LastUserModified)) } files, err := f.listFiles(ctx, folderSlug, "") if err != nil { return nil, err } for _, file := range files { remote := path.Join(dir, file.Name) remoteFile, err := f.newObjectWithInfo(ctx, remote, &file) if err != nil { return nil, err } entries = append(entries, remoteFile) } return entries, nil } func (f *Fs) fetchListFolderPage( ctx context.Context, folderSlug string, searchQuery string, limit int, offset int) (folders []api.Folder, err error) { opts := rest.Opts{ Method: "GET", Path: "/v9/user/" + f.opt.Username + "/folder/" + folderSlug + "/folder-list", Parameters: url.Values{}, } opts.Parameters.Set("status", "ok") opts.Parameters.Set("limit", strconv.Itoa(limit)) if offset > 0 { opts.Parameters.Set("offset", strconv.Itoa(offset)) } if searchQuery != "" { opts.Parameters.Set("search_query", f.opt.Enc.FromStandardName(searchQuery)) } var respBody *api.ListFoldersResponse err = f.pacer.Call(func() (bool, error) { httpResp, err := f.rest.CallJSON(ctx, &opts, nil, &respBody) return f.shouldRetry(ctx, httpResp, err, true) }) if err != nil { return nil, err } for i := range respBody.Subfolders { respBody.Subfolders[i].Name = f.opt.Enc.ToStandardName(respBody.Subfolders[i].Name) } return respBody.Subfolders, nil } func (f *Fs) listFolders( ctx context.Context, folderSlug string, searchQuery string) (folders []api.Folder, err error) { targetPageSize := f.opt.ListPageSize lastPageSize := targetPageSize offset := 0 for targetPageSize == lastPageSize { page, err := f.fetchListFolderPage(ctx, folderSlug, searchQuery, targetPageSize, offset) if err != nil { var apiErr *api.Error casted := errors.As(err, &apiErr) if casted && apiErr.ErrorCode == 30001 { return nil, fs.ErrorDirNotFound } return nil, err } lastPageSize = len(page) offset += lastPageSize folders = append(folders, page...) } return folders, nil } func (f *Fs) fetchListFilePage( ctx context.Context, folderSlug string, searchQuery string, limit int, offset int) (folders []api.File, err error) { opts := rest.Opts{ Method: "GET", Path: "/v8/user/" + f.opt.Username + "/folder/" + folderSlug + "/file-list", Parameters: url.Values{}, } opts.Parameters.Set("status", "ok") opts.Parameters.Set("limit", strconv.Itoa(limit)) if offset > 0 { opts.Parameters.Set("offset", strconv.Itoa(offset)) } if searchQuery != "" { opts.Parameters.Set("search_query", f.opt.Enc.FromStandardName(searchQuery)) } var respBody *api.ListFilesResponse err = f.pacer.Call(func() (bool, error) { httpResp, err := f.rest.CallJSON(ctx, &opts, nil, &respBody) return f.shouldRetry(ctx, httpResp, err, true) }) if err != nil { return nil, fmt.Errorf("couldn't list files: %w", err) } for i := range respBody.Items { respBody.Items[i].Name = f.opt.Enc.ToStandardName(respBody.Items[i].Name) } return respBody.Items, nil } func (f *Fs) listFiles( ctx context.Context, folderSlug string, searchQuery string) (folders []api.File, err error) { targetPageSize := f.opt.ListPageSize lastPageSize := targetPageSize offset := 0 for targetPageSize == lastPageSize { page, err := f.fetchListFilePage(ctx, folderSlug, searchQuery, targetPageSize, offset) if err != nil { return nil, err } lastPageSize = len(page) offset += lastPageSize folders = append(folders, page...) } return folders, nil } // DirCacheFlush implements the optional fs.DirCacheFlusher interface. func (f *Fs) DirCacheFlush() { f.dirCache.ResetRoot() } // Check the interfaces are satisfied var ( _ fs.Fs = (*Fs)(nil) _ dircache.DirCacher = (*Fs)(nil) _ fs.DirCacheFlusher = (*Fs)(nil) _ fs.PutUncheckeder = (*Fs)(nil) _ fs.Object = (*Object)(nil) _ fs.ObjectInfo = (*RenamingObjectInfoProxy)(nil) )