// Package filefabric provides an interface to Storage Made Easy's // Enterprise File Fabric storage system. package filefabric /* Docs: https://product-demo.smestorage.com/?p=apidoc Missing features: - M-Stream support - Oauth-like flow (soon being changed to oauth) // TestFileFabric maxFileLength = 14094 */ import ( "bytes" "context" "encoding/base64" "fmt" "io" "io/ioutil" "net/http" "net/url" "path" "strings" "sync" "sync/atomic" "time" "github.com/rclone/rclone/lib/atexit" "github.com/rclone/rclone/lib/encoder" "github.com/rclone/rclone/lib/random" "github.com/pkg/errors" "github.com/rclone/rclone/backend/filefabric/api" "github.com/rclone/rclone/fs" "github.com/rclone/rclone/fs/config" "github.com/rclone/rclone/fs/config/configmap" "github.com/rclone/rclone/fs/config/configstruct" "github.com/rclone/rclone/fs/fserrors" "github.com/rclone/rclone/fs/fshttp" "github.com/rclone/rclone/fs/hash" "github.com/rclone/rclone/fs/log" "github.com/rclone/rclone/lib/dircache" "github.com/rclone/rclone/lib/pacer" "github.com/rclone/rclone/lib/rest" ) const ( minSleep = 20 * time.Millisecond maxSleep = 10 * time.Second decayConstant = 2 // bigger for slower decay, exponential listChunks = 1000 // chunk size to read directory listings tokenLifeTime = 55 * time.Minute // 1 hour minus a bit of leeway defaultRootID = "" // default root ID emptyMimeType = "application/vnd.rclone.empty.file" ) // Register with Fs func init() { fs.Register(&fs.RegInfo{ Name: "filefabric", Description: "Enterprise File Fabric", NewFs: NewFs, Options: []fs.Option{{ Name: "url", Help: "URL of the Enterprise File Fabric to connect to.", Required: true, Examples: []fs.OptionExample{{ Value: "https://storagemadeeasy.com", Help: "Storage Made Easy US", }, { Value: "https://eu.storagemadeeasy.com", Help: "Storage Made Easy EU", }, { Value: "https://yourfabric.smestorage.com", Help: "Connect to your Enterprise File Fabric", }}, }, { Name: "root_folder_id", Help: `ID of the root folder. Leave blank normally. Fill in to make rclone start with directory of a given ID. `, }, { Name: "permanent_token", Help: `Permanent Authentication Token. A Permanent Authentication Token can be created in the Enterprise File Fabric, on the users Dashboard under Security, there is an entry you'll see called "My Authentication Tokens". Click the Manage button to create one. These tokens are normally valid for several years. For more info see: https://docs.storagemadeeasy.com/organisationcloud/api-tokens `, }, { Name: "token", Help: `Session Token. This is a session token which rclone caches in the config file. It is usually valid for 1 hour. Don't set this value - rclone will set it automatically. `, Advanced: true, }, { Name: "token_expiry", Help: `Token expiry time. Don't set this value - rclone will set it automatically. `, Advanced: true, }, { Name: "version", Help: `Version read from the file fabric. Don't set this value - rclone will set it automatically. `, Advanced: true, }, { Name: config.ConfigEncoding, Help: config.ConfigEncodingHelp, Advanced: true, Default: (encoder.Display | encoder.EncodeInvalidUtf8), }}, }) } // Options defines the configuration for this backend type Options struct { URL string `config:"url"` RootFolderID string `config:"root_folder_id"` PermanentToken string `config:"permanent_token"` Token string `config:"token"` TokenExpiry string `config:"token_expiry"` Version string `config:"version"` Enc encoder.MultiEncoder `config:"encoding"` } // Fs represents a remote filefabric type Fs struct { name string // name of this remote root string // the path we are working on opt Options // parsed options features *fs.Features // optional features m configmap.Mapper // to save config srv *rest.Client // the connection to the one drive server dirCache *dircache.DirCache // Map of directory path to directory id pacer *fs.Pacer // pacer for API calls tokenMu sync.Mutex // hold when reading the token token string // current access token tokenExpiry time.Time // time the current token expires tokenExpired int32 // read and written with atomic canCopyWithName bool // set if detected that can use fi_name in copy precision time.Duration // precision reported } // Object describes a filefabric object // // Will definitely have info but maybe not meta type Object struct { fs *Fs // what this object is part of remote string // The remote path hasMetaData bool // whether info below has been set size int64 // size of the object modTime time.Time // modification time of the object id string // ID of the object contentType string // ContentType of object } // ------------------------------------------------------------ // Name of the remote (as passed into NewFs) func (f *Fs) Name() string { return f.name } // Root of the remote (as passed into NewFs) func (f *Fs) Root() string { return f.root } // String converts this Fs to a string func (f *Fs) String() string { return fmt.Sprintf("filefabric root '%s'", f.root) } // Features returns the optional features of this Fs func (f *Fs) Features() *fs.Features { return f.features } // parsePath parses a filefabric 'url' func parsePath(path string) (root string) { root = strings.Trim(path, "/") return } // retryErrorCodes is a slice of error codes that we will retry var retryErrorCodes = []int{ 429, // Too Many Requests. 500, // Internal Server Error 502, // Bad Gateway 503, // Service Unavailable 504, // Gateway Timeout 509, // Bandwidth Limit Exceeded } // Retry any of these var retryStatusCodes = []struct { code string sleep time.Duration }{ { // Can not create folder now. We are not able to complete the // requested operation with such name. We are processing // delete in that folder. Please try again later or use // another name. (error_background) code: "error_background", sleep: 1 * time.Second, }, } // shouldRetry returns a boolean as to whether this resp and err // deserve to be retried. It returns the err as a convenience // try should be the number of the tries so far, counting up from 1 func (f *Fs) shouldRetry(ctx context.Context, resp *http.Response, err error, status api.OKError, try int) (bool, error) { if fserrors.ContextError(ctx, &err) { return false, err } if err != nil { return fserrors.ShouldRetry(err) || fserrors.ShouldRetryHTTP(resp, retryErrorCodes), err } if status != nil && !status.OK() { err = status // return the error from the RPC code := status.GetCode() if code == "login_token_expired" { atomic.AddInt32(&f.tokenExpired, 1) } else { for _, retryCode := range retryStatusCodes { if code == retryCode.code { if retryCode.sleep > 0 { // make this thread only sleep exponentially increasing extra time sleepTime := retryCode.sleep << (try - 1) fs.Debugf(f, "Sleeping for %v to wait for %q error to clear", sleepTime, retryCode.code) time.Sleep(sleepTime) } return true, err } } } } return false, err } // readMetaDataForPath reads the metadata from the path func (f *Fs) readMetaDataForPath(ctx context.Context, rootID string, path string) (info *api.Item, err error) { var resp api.FileResponse _, err = f.rpc(ctx, "checkPathExists", params{ "path": f.opt.Enc.FromStandardPath(path), "pid": rootID, }, &resp, nil) if err != nil { return nil, errors.Wrap(err, "failed to check path exists") } if resp.Exists != "y" { return nil, fs.ErrorObjectNotFound } return &resp.Item, nil /* // defer fs.Trace(f, "path=%q", path)("info=%+v, err=%v", &info, &err) leaf, directoryID, err := f.dirCache.FindPath(ctx, path, false) if err != nil { if err == fs.ErrorDirNotFound { return nil, fs.ErrorObjectNotFound } return nil, err } found, err := f.listAll(ctx, directoryID, false, true, func(item *api.Item) bool { if item.Name == leaf { info = item return true } return false }) if err != nil { return nil, err } if !found { return nil, fs.ErrorObjectNotFound } return info, nil */ } // Get the appliance info so we can set Version func (f *Fs) getApplianceInfo(ctx context.Context) error { var applianceInfo api.ApplianceInfo _, err := f.rpc(ctx, "getApplianceInfo", params{ "token": "*", }, &applianceInfo, nil) if err != nil { return errors.Wrap(err, "failed to read appliance version") } f.opt.Version = applianceInfo.SoftwareVersionLabel f.m.Set("version", f.opt.Version) return nil } // Gets the token or gets a new one if necessary func (f *Fs) getToken(ctx context.Context) (token string, err error) { f.tokenMu.Lock() var refreshed = false defer func() { if refreshed { atomic.StoreInt32(&f.tokenExpired, 0) } f.tokenMu.Unlock() }() expired := atomic.LoadInt32(&f.tokenExpired) != 0 if expired { fs.Debugf(f, "Token invalid - refreshing") } if f.token == "" { fs.Debugf(f, "Empty token - refreshing") expired = true } now := time.Now() if f.tokenExpiry.IsZero() || now.After(f.tokenExpiry) { fs.Debugf(f, "Token expired - refreshing") expired = true } if !expired { return f.token, nil } var info api.GetTokenByAuthTokenResponse _, err = f.rpc(ctx, "getTokenByAuthToken", params{ "token": "*", "authtoken": f.opt.PermanentToken, }, &info, nil) if err != nil { return "", errors.Wrap(err, "failed to get session token") } refreshed = true now = now.Add(tokenLifeTime) f.token = info.Token f.tokenExpiry = now f.m.Set("token", f.token) f.m.Set("token_expiry", now.Format(time.RFC3339)) // Read appliance info when we update the token err = f.getApplianceInfo(ctx) if err != nil { return "", err } f.setCapabilities() return f.token, nil } // params for rpc type params map[string]interface{} // rpc calls the rpc.php method of the SME file fabric // // This is an entry point to all the method calls // // If result is nil then resp.Body will need closing func (f *Fs) rpc(ctx context.Context, function string, p params, result api.OKError, options []fs.OpenOption) (resp *http.Response, err error) { defer log.Trace(f, "%s(%+v) options=%+v", function, p, options)("result=%+v, err=%v", &result, &err) // Get the token from params if present otherwise call getToken var token string if tokenI, ok := p["token"]; !ok { token, err = f.getToken(ctx) if err != nil { return resp, err } } else { token = tokenI.(string) } var data = url.Values{ "function": {function}, "token": {token}, "apiformat": {"json"}, } for k, v := range p { data.Set(k, fmt.Sprint(v)) } opts := rest.Opts{ Method: "POST", Path: "/api/rpc.php", ContentType: "application/x-www-form-urlencoded", Options: options, } try := 0 err = f.pacer.Call(func() (bool, error) { try++ // Refresh the body each retry opts.Body = strings.NewReader(data.Encode()) resp, err = f.srv.CallJSON(ctx, &opts, nil, result) return f.shouldRetry(ctx, resp, err, result, try) }) if err != nil { return resp, err } return resp, nil } // NewFs constructs an Fs from the path, container:path func NewFs(ctx context.Context, name, root string, m configmap.Mapper) (fs.Fs, error) { // Parse config into Options struct opt := new(Options) err := configstruct.Set(m, opt) if err != nil { return nil, err } opt.URL = strings.TrimSuffix(opt.URL, "/") if opt.URL == "" { return nil, errors.New("url must be set") } root = parsePath(root) client := fshttp.NewClient(ctx) f := &Fs{ name: name, root: root, opt: *opt, m: m, srv: rest.NewClient(client).SetRoot(opt.URL), pacer: fs.NewPacer(ctx, pacer.NewDefault(pacer.MinSleep(minSleep), pacer.MaxSleep(maxSleep), pacer.DecayConstant(decayConstant))), token: opt.Token, } f.features = (&fs.Features{ CaseInsensitive: true, CanHaveEmptyDirectories: true, ReadMimeType: true, WriteMimeType: true, }).Fill(ctx, f) if f.opt.Version == "" { err = f.getApplianceInfo(ctx) if err != nil { return nil, err } } f.setCapabilities() if opt.TokenExpiry != "" { tokenExpiry, err := time.Parse(time.RFC3339, opt.TokenExpiry) if err != nil { fs.Errorf(nil, "Failed to parse token_expiry option: %v", err) } else { f.tokenExpiry = tokenExpiry } } if opt.RootFolderID == "" { opt.RootFolderID = defaultRootID } f.dirCache = dircache.New(f.root, opt.RootFolderID, f) // Find out whether the root is a file or a directory or doesn't exist var errReturn error if f.root != "" { info, err := f.readMetaDataForPath(ctx, f.opt.RootFolderID, f.root) if err == nil && info != nil { if info.Type == api.ItemTypeFile { // Root is a file // Point the root to the parent directory f.root, _ = dircache.SplitPath(root) f.dirCache = dircache.New(f.root, opt.RootFolderID, f) errReturn = fs.ErrorIsFile // Cache the ID of the parent of the file as the root ID f.dirCache.Put(f.root, info.PID) } else if info.Type == api.ItemTypeFolder { // Root is a dir - cache its ID f.dirCache.Put(f.root, info.ID) } } else { // Root is not found so a directory } } return f, errReturn } // set the capabilities of this version of software func (f *Fs) setCapabilities() { version := f.opt.Version if version == "" { version = "0000.00" } if version >= "2006.02" { f.precision = time.Second f.canCopyWithName = true } else { // times can be altered this much on renames f.precision = 1 * time.Hour f.canCopyWithName = false } } // Return an Object from a path // // If it can't be found it returns the error fs.ErrorObjectNotFound. func (f *Fs) newObjectWithInfo(ctx context.Context, remote string, info *api.Item) (fs.Object, error) { o := &Object{ fs: f, remote: remote, } var err error if info != nil { // Set info err = o.setMetaData(info) } else { err = o.readMetaData(ctx) // reads info and meta, returning an error } if err != nil { return nil, err } return o, nil } // NewObject finds the Object at remote. If it can't be found // it returns the error fs.ErrorObjectNotFound. func (f *Fs) NewObject(ctx context.Context, remote string) (fs.Object, error) { return f.newObjectWithInfo(ctx, remote, nil) } // FindLeaf finds a directory of name leaf in the folder with ID pathID func (f *Fs) FindLeaf(ctx context.Context, pathID, leaf string) (pathIDOut string, found bool, err error) { // Find the leaf in pathID found, err = f.listAll(ctx, pathID, true, false, func(item *api.Item) bool { if strings.EqualFold(item.Name, leaf) { pathIDOut = item.ID return true } return false }) return pathIDOut, found, err } // CreateDir makes a directory with pathID as parent and name leaf func (f *Fs) CreateDir(ctx context.Context, pathID, leaf string) (newID string, err error) { //fs.Debugf(f, "CreateDir(%q, %q)\n", pathID, leaf) var info api.DoCreateNewFolderResponse _, err = f.rpc(ctx, "doCreateNewFolder", params{ "fi_pid": pathID, "fi_name": f.opt.Enc.FromStandardName(leaf), }, &info, nil) if err != nil { return "", errors.Wrap(err, "failed to create directory") } // fmt.Printf("...Id %q\n", *info.Id) return info.Item.ID, nil } // list the objects into the function supplied // // If directories is set it only sends directories // User function to process a File item from listAll // // Should return true to finish processing type listAllFn func(*api.Item) bool // Lists the directory required calling the user function on each item found // // If the user fn ever returns true then it early exits with found = true func (f *Fs) listAll(ctx context.Context, dirID string, directoriesOnly bool, filesOnly bool, fn listAllFn) (found bool, err error) { var ( p = params{ "fi_pid": dirID, "count": listChunks, "subfolders": "y", // Cut down the things that are returned "options": "filelist|" + api.ItemFields, } n = 0 ) OUTER: for { var info api.GetFolderContentsResponse _, err = f.rpc(ctx, "getFolderContents", p, &info, nil) if err != nil { return false, errors.Wrap(err, "failed to list directory") } for i := range info.Items { item := &info.Items[i] if item.Type == api.ItemTypeFolder { if filesOnly { continue } } else if item.Type == api.ItemTypeFile { if directoriesOnly { continue } } else { fs.Debugf(f, "Ignoring %q - unknown type %q", item.Name, item.Type) continue } if item.Trash { continue } item.Name = f.opt.Enc.ToStandardName(item.Name) if fn(item) { found = true break OUTER } } // if didn't get any items then exit if len(info.Items) == 0 { break } n += len(info.Items) if n >= info.Total { break } p["from"] = n } return found, nil } // List the objects and directories in dir into entries. The // entries can be returned in any order but should be for a // complete directory. // // dir should be "" to list the root, and should not have // trailing slashes. // // This should return ErrDirNotFound if the directory isn't // found. func (f *Fs) List(ctx context.Context, dir string) (entries fs.DirEntries, err error) { directoryID, err := f.dirCache.FindDir(ctx, dir, false) if err != nil { return nil, err } var iErr error _, err = f.listAll(ctx, directoryID, false, false, func(info *api.Item) bool { remote := path.Join(dir, info.Name) if info.Type == api.ItemTypeFolder { // cache the directory ID for later lookups f.dirCache.Put(remote, info.ID) d := fs.NewDir(remote, time.Time(info.Modified)).SetID(info.ID).SetItems(info.SubFolders) entries = append(entries, d) } else if info.Type == api.ItemTypeFile { o, err := f.newObjectWithInfo(ctx, remote, info) if err != nil { iErr = err return true } entries = append(entries, o) } return false }) if err != nil { return nil, err } if iErr != nil { return nil, iErr } return entries, nil } // Creates from the parameters passed in a half finished Object which // must have setMetaData called on it // // Returns the object, leaf, directoryID and error // // Used to create new objects func (f *Fs) createObject(ctx context.Context, remote string, modTime time.Time, size int64) (o *Object, leaf string, directoryID string, err error) { // Create the directory for the object if it doesn't exist leaf, directoryID, err = f.dirCache.FindPath(ctx, remote, true) if err != nil { return } // Temporary Object under construction o = &Object{ fs: f, remote: remote, } return o, leaf, directoryID, nil } // Put the object // // Copy the reader in to the new object which is returned // // The new object may have been created if an error is returned func (f *Fs) Put(ctx context.Context, in io.Reader, src fs.ObjectInfo, options ...fs.OpenOption) (fs.Object, error) { remote := src.Remote() size := src.Size() modTime := src.ModTime(ctx) o, _, _, err := f.createObject(ctx, remote, modTime, size) if err != nil { return nil, err } return o, o.Update(ctx, in, src, options...) } // Mkdir creates the container if it doesn't exist func (f *Fs) Mkdir(ctx context.Context, dir string) error { _, err := f.dirCache.FindDir(ctx, dir, true) return err } // deleteObject removes an object by ID func (f *Fs) deleteObject(ctx context.Context, id string) (err error) { var info api.DeleteResponse _, err = f.rpc(ctx, "doDeleteFile", params{ "fi_id": id, "completedeletion": "n", }, &info, nil) if err != nil { return errors.Wrap(err, "failed to delete file") } return nil } // purgeCheck removes the root directory, if check is set then it // refuses to do so if it has anything in func (f *Fs) purgeCheck(ctx context.Context, dir string, check bool) error { root := path.Join(f.root, dir) if root == "" { return errors.New("can't purge root directory") } dc := f.dirCache rootID, err := dc.FindDir(ctx, dir, false) if err != nil { return err } if check { found, err := f.listAll(ctx, rootID, false, false, func(item *api.Item) bool { fs.Debugf(dir, "Rmdir: contains file: %q", item.Name) return true }) if err != nil { return err } if found { return fs.ErrorDirectoryNotEmpty } } var info api.EmptyResponse _, err = f.rpc(ctx, "doDeleteFolder", params{ "fi_id": rootID, }, &info, nil) f.dirCache.FlushDir(dir) if err != nil { return errors.Wrap(err, "failed to remove directory") } return nil } // Rmdir deletes the root folder // // Returns an error if it isn't empty func (f *Fs) Rmdir(ctx context.Context, dir string) error { //fs.Debugf(f, "CreateDir(%q, %q)\n", pathID, leaf) return f.purgeCheck(ctx, dir, true) } // Precision return the precision of this Fs func (f *Fs) Precision() time.Duration { return f.precision } // Copy src to this remote using server side copy operations. // // This is stored with the remote path given // // It returns the destination Object and a possible error // // Will only be called if src.Fs().Name() == f.Name() // // If it isn't possible then return fs.ErrorCantCopy func (f *Fs) Copy(ctx context.Context, src fs.Object, remote string) (fs.Object, error) { srcObj, ok := src.(*Object) if !ok { fs.Debugf(src, "Can't copy - not same remote type") return nil, fs.ErrorCantCopy } err := srcObj.readMetaData(ctx) if err != nil { return nil, err } // Create temporary object dstObj, leaf, directoryID, err := f.createObject(ctx, remote, srcObj.modTime, srcObj.size) if err != nil { return nil, err } if !f.canCopyWithName && leaf != path.Base(srcObj.remote) { fs.Debugf(src, "Can't copy - can't change the name of files") return nil, fs.ErrorCantCopy } // Copy the object var info api.FileResponse p := params{ "fi_id": srcObj.id, "fi_pid": directoryID, "force": "y", "options": "allownoextension", // without this the filefabric adds extensions to files without } if f.canCopyWithName { p["fi_name"] = f.opt.Enc.FromStandardName(leaf) } _, err = f.rpc(ctx, "doCopyFile", p, &info, nil) if err != nil { return nil, errors.Wrap(err, "failed to copy file") } err = dstObj.setMetaData(&info.Item) if err != nil { return nil, err } return dstObj, nil } // Purge deletes all the files and the container // // Optional interface: Only implement this if you have a way of // deleting all the files quicker than just running Remove() on the // result of List() func (f *Fs) Purge(ctx context.Context, dir string) error { return f.purgeCheck(ctx, dir, false) } // Wait for the the background task to complete if necessary func (f *Fs) waitForBackgroundTask(ctx context.Context, taskID api.String) (err error) { if taskID == "" || taskID == "0" { // No task to wait for return nil } start := time.Now() sleepTime := time.Second for { var info api.TasksResponse _, err = f.rpc(ctx, "getUserBackgroundTasks", params{ "taskid": taskID, }, &info, nil) if err != nil { return errors.Wrapf(err, "failed to wait for task %s to complete", taskID) } if len(info.Tasks) == 0 { // task has finished break } if len(info.Tasks) > 1 { fs.Errorf(f, "Unexpected number of tasks returned %d", len(info.Tasks)) } task := info.Tasks[0] if task.BtStatus == "c" { // task completed break } dt := time.Since(start) fs.Debugf(f, "Waiting for task ID %s: %s: to completed for %v - waited %v already", task.BtID, task.BtTitle, sleepTime, dt) time.Sleep(sleepTime) } return nil } // Rename the leaf of a file or directory in a directory func (f *Fs) renameLeaf(ctx context.Context, isDir bool, id string, newLeaf string) (item *api.Item, err error) { var info api.FileResponse method := "doRenameFile" if isDir { method = "doRenameFolder" } _, err = f.rpc(ctx, method, params{ "fi_id": id, "fi_name": newLeaf, }, &info, nil) if err != nil { return nil, errors.Wrap(err, "failed to rename leaf") } err = f.waitForBackgroundTask(ctx, info.Status.TaskID) if err != nil { return nil, err } return &info.Item, nil } // move a file or folder // // This is complicated by the fact that there is an API to move files // between directories and a separate one to rename them. We try to // call the minimum number of API calls. func (f *Fs) move(ctx context.Context, isDir bool, id, oldLeaf, newLeaf, oldDirectoryID, newDirectoryID string) (item *api.Item, err error) { newLeaf = f.opt.Enc.FromStandardName(newLeaf) oldLeaf = f.opt.Enc.FromStandardName(oldLeaf) doRenameLeaf := oldLeaf != newLeaf doMove := oldDirectoryID != newDirectoryID // Now rename the leaf to a temporary name if we are moving to // another directory to make sure we don't overwrite something // in the destination directory by accident if doRenameLeaf && doMove { tmpLeaf := newLeaf + "." + random.String(8) item, err = f.renameLeaf(ctx, isDir, id, tmpLeaf) if err != nil { return nil, err } } // Move the object to a new directory (with the existing name) // if required if doMove { var info api.MoveFilesResponse method := "doMoveFiles" if isDir { method = "doMoveFolders" } _, err = f.rpc(ctx, method, params{ "fi_ids": id, "dir_id": newDirectoryID, }, &info, nil) if err != nil { return nil, errors.Wrap(err, "failed to move file to new directory") } item = &info.Item err = f.waitForBackgroundTask(ctx, info.Status.TaskID) if err != nil { return nil, err } } // Rename the leaf to its final name if required if doRenameLeaf { item, err = f.renameLeaf(ctx, isDir, id, newLeaf) if err != nil { return nil, err } } return item, nil } // Move src to this remote using server side move operations. // // This is stored with the remote path given // // It returns the destination Object and a possible error // // Will only be called if src.Fs().Name() == f.Name() // // If it isn't possible then return fs.ErrorCantMove func (f *Fs) Move(ctx context.Context, src fs.Object, remote string) (fs.Object, error) { srcObj, ok := src.(*Object) if !ok { fs.Debugf(src, "Can't move - not same remote type") return nil, fs.ErrorCantMove } // find the source directoryID srcLeaf, srcDirectoryID, err := srcObj.fs.dirCache.FindPath(ctx, srcObj.remote, false) if err != nil { return nil, err } // Create temporary object dstObj, dstLeaf, dstDirectoryID, err := f.createObject(ctx, remote, srcObj.modTime, srcObj.size) if err != nil { return nil, err } // Do the move item, err := f.move(ctx, false, srcObj.id, srcLeaf, dstLeaf, srcDirectoryID, dstDirectoryID) if err != nil { return nil, err } // Set the metadata from what was returned or read it fresh if item == nil { err = dstObj.readMetaData(ctx) if err != nil { return nil, err } } else { err = dstObj.setMetaData(item) if err != nil { return nil, err } } return dstObj, nil } // DirMove moves src, srcRemote to this remote at dstRemote // using server side move operations. // // Will only be called if src.Fs().Name() == f.Name() // // If it isn't possible then return fs.ErrorCantDirMove // // If destination exists then return fs.ErrorDirExists func (f *Fs) DirMove(ctx context.Context, src fs.Fs, srcRemote, dstRemote string) error { srcFs, ok := src.(*Fs) if !ok { fs.Debugf(srcFs, "Can't move directory - not same remote type") return fs.ErrorCantDirMove } srcID, srcDirectoryID, srcLeaf, dstDirectoryID, dstLeaf, err := f.dirCache.DirMove(ctx, srcFs.dirCache, srcFs.root, srcRemote, f.root, dstRemote) if err != nil { return err } // Do the move _, err = f.move(ctx, true, srcID, srcLeaf, dstLeaf, srcDirectoryID, dstDirectoryID) if err != nil { return err } srcFs.dirCache.FlushDir(srcRemote) return nil } // CleanUp empties the trash func (f *Fs) CleanUp(ctx context.Context) (err error) { var info api.EmptyResponse _, err = f.rpc(ctx, "emptyTrashInBackground", params{}, &info, nil) if err != nil { return errors.Wrap(err, "failed to empty trash") } return nil } // DirCacheFlush resets the directory cache - used in testing as an // optional interface func (f *Fs) DirCacheFlush() { f.dirCache.ResetRoot() } // Hashes returns the supported hash sets. func (f *Fs) Hashes() hash.Set { return hash.Set(hash.None) } // ------------------------------------------------------------ // Fs returns the parent Fs func (o *Object) Fs() fs.Info { return o.fs } // Return a string version func (o *Object) String() string { if o == nil { return "" } return o.remote } // Remote returns the remote path func (o *Object) Remote() string { return o.remote } // Hash of the object in the requested format as a lowercase hex string func (o *Object) Hash(ctx context.Context, t hash.Type) (string, error) { return "", hash.ErrUnsupported } // Size returns the size of an object in bytes func (o *Object) Size() int64 { err := o.readMetaData(context.TODO()) if err != nil { fs.Logf(o, "Failed to read metadata: %v", err) return 0 } if o.contentType == emptyMimeType { return 0 } return o.size } // setMetaData sets the metadata from info func (o *Object) setMetaData(info *api.Item) (err error) { if info.Type != api.ItemTypeFile { return fs.ErrorIsDir } o.hasMetaData = true o.size = info.Size o.modTime = time.Time(info.Modified) if !time.Time(info.LocalTime).IsZero() { o.modTime = time.Time(info.LocalTime) } o.id = info.ID o.contentType = info.ContentType return nil } // readMetaData gets the metadata if it hasn't already been fetched // // it also sets the info func (o *Object) readMetaData(ctx context.Context) (err error) { if o.hasMetaData { return nil } rootID, err := o.fs.dirCache.RootID(ctx, false) if err != nil { if err == fs.ErrorDirNotFound { err = fs.ErrorObjectNotFound } return err } info, err := o.fs.readMetaDataForPath(ctx, rootID, o.remote) if err != nil { if apiErr, ok := err.(*api.Status); ok { if apiErr.Code == "not_found" || apiErr.Code == "trashed" { return fs.ErrorObjectNotFound } } return err } return o.setMetaData(info) } // ModTime returns the modification time of the object // // // It attempts to read the objects mtime and if that isn't present the // LastModified returned in the http headers func (o *Object) ModTime(ctx context.Context) time.Time { err := o.readMetaData(ctx) if err != nil { fs.Logf(o, "Failed to read metadata: %v", err) return time.Now() } return o.modTime } // modifyFile updates file metadata // // keyValues should be key, value pairs func (o *Object) modifyFile(ctx context.Context, keyValues [][2]string) error { var info api.FileResponse var data strings.Builder for _, keyValue := range keyValues { data.WriteString(keyValue[0]) data.WriteRune('=') data.WriteString(keyValue[1]) data.WriteRune('\n') } _, err := o.fs.rpc(ctx, "doModifyFile", params{ "fi_id": o.id, "data": data.String(), }, &info, nil) if err != nil { return errors.Wrap(err, "failed to update metadata") } return o.setMetaData(&info.Item) } // SetModTime sets the modification time of the local fs object func (o *Object) SetModTime(ctx context.Context, modTime time.Time) error { return o.modifyFile(ctx, [][2]string{ {"fi_localtime", api.Time(modTime).String()}, }) } // Storable returns a boolean showing whether this object storable func (o *Object) Storable() bool { return true } // Open an object for read func (o *Object) Open(ctx context.Context, options ...fs.OpenOption) (in io.ReadCloser, err error) { if o.id == "" { return nil, errors.New("can't download - no id") } if o.contentType == emptyMimeType { return ioutil.NopCloser(bytes.NewReader([]byte{})), nil } fs.FixRangeOption(options, o.size) resp, err := o.fs.rpc(ctx, "getFile", params{ "fi_id": o.id, }, nil, options) if err != nil { return nil, err } return resp.Body, nil } // Update the object with the contents of the io.Reader, modTime and size // // If existing is set then it updates the object rather than creating a new one // // The new object may have been created if an error is returned func (o *Object) Update(ctx context.Context, in io.Reader, src fs.ObjectInfo, options ...fs.OpenOption) (err error) { modTime := src.ModTime(ctx) remote := o.remote size := src.Size() // Can't upload 0 length files - these upload as a single // space. // if size == 0 { // return fs.ErrorCantUploadEmptyFiles // } // Create the directory for the object if it doesn't exist leaf, directoryID, err := o.fs.dirCache.FindPath(ctx, remote, true) if err != nil { return err } // Initialise the upload var upload api.DoInitUploadResponse timestamp := api.Time(modTime).String() encodedLeaf := o.fs.opt.Enc.FromStandardName(leaf) base64EncodedLeaf := base64.StdEncoding.EncodeToString([]byte(encodedLeaf)) contentType := fs.MimeType(ctx, src) if size == 0 { contentType = emptyMimeType } p := params{ "fi_name": encodedLeaf, "fi_pid": directoryID, "fi_filename": encodedLeaf, "fi_localtime": timestamp, "fi_modified": timestamp, "fi_contenttype": contentType, "responsetype": "json", // make the upload.cgi return JSON "directuploadsupport": "n", // FIXME should we support this? // "chunkifbig": "n", // FIXME multipart? } // Set the size if known if size >= 0 { p["fi_size"] = size } _, err = o.fs.rpc(ctx, "doInitUpload", p, &upload, nil) if err != nil { return errors.Wrap(err, "failed to initialize upload") } // Cancel the upload if aborted or it fails finalized := false defer atexit.OnError(&err, func() { if finalized { return } fs.Debugf(o, "Cancelling upload %s", upload.UploadCode) var cancel api.EmptyResponse _, fErr := o.fs.rpc(ctx, "doAbortUpload", params{ "uploadcode": upload.UploadCode, }, &cancel, nil) if fErr != nil { fs.Errorf(o, "failed to cancel upload: %v", fErr) } })() // Post the file with the upload code var uploader api.UploaderResponse opts := rest.Opts{ //Method: "POST", Method: "PUT", Path: "/cgi-bin/uploader/uploader1.cgi/" + base64EncodedLeaf + "?" + upload.UploadCode, Body: in, ContentType: contentType, // MultipartParams: url.Values{}, // MultipartContentName: "file", // MultipartFileName: "datafile", } // Set the size if known if size >= 0 { var contentLength = size opts.ContentLength = &contentLength // NB CallJSON scribbles on this which is naughty } try := 0 err = o.fs.pacer.CallNoRetry(func() (bool, error) { try++ resp, err := o.fs.srv.CallJSON(ctx, &opts, nil, &uploader) return o.fs.shouldRetry(ctx, resp, err, nil, try) }) if err != nil { return errors.Wrap(err, "failed to upload") } if uploader.Success != "y" { return errors.Errorf("upload failed") } if size > 0 && uploader.FileSize != size { return errors.Errorf("upload failed: size mismatch: want %d got %d", size, uploader.FileSize) } // Now finalize the file var finalize api.DoCompleteUploadResponse p = params{ "uploadcode": upload.UploadCode, "remotetime": timestamp, "fi_size": uploader.FileSize, } _, err = o.fs.rpc(ctx, "doCompleteUpload", p, &finalize, nil) if err != nil { return errors.Wrap(err, "failed to finalize upload") } finalized = true err = o.setMetaData(&finalize.File) if err != nil { return err } // Make sure content type is correct if o.contentType != contentType { fs.Debugf(o, "Correcting mime type from %q to %q", o.contentType, contentType) return o.modifyFile(ctx, [][2]string{ {"fi_contenttype", contentType}, }) } return nil } // Remove an object func (o *Object) Remove(ctx context.Context) error { return o.fs.deleteObject(ctx, o.id) } // ID returns the ID of the Object if known, or "" if not func (o *Object) ID() string { return o.id } // MimeType returns the content type of the Object if // known, or "" if not func (o *Object) MimeType(ctx context.Context) string { return o.contentType } // Check the interfaces are satisfied var ( _ fs.Fs = (*Fs)(nil) _ fs.Purger = (*Fs)(nil) _ fs.Copier = (*Fs)(nil) _ fs.Mover = (*Fs)(nil) _ fs.DirMover = (*Fs)(nil) _ fs.DirCacheFlusher = (*Fs)(nil) _ fs.CleanUpper = (*Fs)(nil) _ fs.Object = (*Object)(nil) _ fs.IDer = (*Object)(nil) _ fs.MimeTyper = (*Object)(nil) )