// Amazon Cloud Drive interface package amazonclouddrive /* FIXME make searching for directory in id and file in id more efficient - use the name: search parameter - remember the escaping rules - use Folder GetNode and GetFile FIXME make the default for no files and no dirs be (FILE & FOLDER) so we ignore assets completely! */ import ( "fmt" "io" "log" "math/rand" "net/http" "regexp" "strings" "sync" "time" "github.com/ncw/go-acd" "github.com/ncw/rclone/dircache" "github.com/ncw/rclone/fs" "github.com/ncw/rclone/oauthutil" "golang.org/x/oauth2" ) const ( rcloneClientID = "amzn1.application-oa2-client.6bf18d2d1f5b485c94c8988bb03ad0e7" rcloneClientSecret = "k8/NyszKm5vEkZXAwsbGkd6C3NrbjIqMg4qEhIeF14Szub2wur+/teS3ubXgsLe9//+tr/qoqK+lq6mg8vWkoA==" bindAddress = "127.0.0.1:53682" redirectURL = "http://" + bindAddress + "/" folderKind = "FOLDER" fileKind = "FILE" assetKind = "ASSET" statusAvailable = "AVAILABLE" timeFormat = time.RFC3339 // 2014-03-07T22:31:12.173Z minBackoff = 1 * time.Second maxBackoff = 256 * time.Second ) // Globals var ( // Description of how to auth for this app acdConfig = &oauth2.Config{ Scopes: []string{"clouddrive:read_all", "clouddrive:write"}, Endpoint: oauth2.Endpoint{ AuthURL: "https://www.amazon.com/ap/oa", TokenURL: "https://api.amazon.com/auth/o2/token", }, ClientID: rcloneClientID, ClientSecret: fs.Reveal(rcloneClientSecret), RedirectURL: redirectURL, } FIXME = fmt.Errorf("FIXME not implemented") ) // Register with Fs func init() { fs.Register(&fs.FsInfo{ Name: "amazon cloud drive", NewFs: NewFs, Config: func(name string) { err := oauthutil.ConfigWithWebserver(name, acdConfig, bindAddress) if err != nil { log.Fatalf("Failed to configure token: %v", err) } }, Options: []fs.Option{{ Name: "client_id", Help: "Amazon Application Client Id - leave blank to use rclone's.", }, { Name: "client_secret", Help: "Amazon Application Client Secret - leave blank to use rclone's.", }}, }) } // FsAcd represents a remote acd server type FsAcd struct { name string // name of this remote c *acd.Client // the connection to the acd server root string // the path we are working on dirCache *dircache.DirCache // Map of directory path to directory id backoffLock sync.Mutex backoff time.Duration // current backoff } // FsObjectAcd describes a acd object // // Will definitely have info but maybe not meta type FsObjectAcd struct { acd *FsAcd // what this object is part of remote string // The remote path info *acd.Node // Info from the acd object if known } // ------------------------------------------------------------ // The name of the remote (as passed into NewFs) func (f *FsAcd) Name() string { return f.name } // The root of the remote (as passed into NewFs) func (f *FsAcd) Root() string { return f.root } // String converts this FsAcd to a string func (f *FsAcd) String() string { return fmt.Sprintf("Amazon cloud drive root '%s'", f.root) } // Pattern to match a acd path var matcher = regexp.MustCompile(`^([^/]*)(.*)$`) // parsePath parses an acd 'url' func parsePath(path string) (root string) { root = strings.Trim(path, "/") return } // NewFs contstructs an FsAcd from the path, container:path func NewFs(name, root string) (fs.Fs, error) { root = parsePath(root) oAuthClient, err := oauthutil.NewClient(name, acdConfig) if err != nil { log.Fatalf("Failed to configure amazon cloud drive: %v", err) } c := acd.NewClient(oAuthClient) c.UserAgent = fs.UserAgent f := &FsAcd{ name: name, root: root, c: c, } // Update endpoints _, _, err = f.c.Account.GetEndpoints() if err != nil { return nil, fmt.Errorf("Failed to get endpoints: %v", err) } // Get rootID rootInfo, _, err := f.c.Nodes.GetRoot() if err != nil || rootInfo.Id == nil { return nil, fmt.Errorf("Failed to get root: %v", err) } f.dirCache = dircache.New(root, *rootInfo.Id, f) // Find the current root err = f.dirCache.FindRoot(false) if err != nil { // Assume it is a file newRoot, remote := dircache.SplitPath(root) newF := *f newF.dirCache = dircache.New(newRoot, *rootInfo.Id, &newF) newF.root = newRoot // Make new Fs which is the parent err = newF.dirCache.FindRoot(false) if err != nil { // No root so return old f return f, nil } obj := newF.newFsObjectWithInfo(remote, nil) if obj == nil { // File doesn't exist so return old f return f, nil } // return a Fs Limited to this object return fs.NewLimited(&newF, obj), nil } return f, nil } // Return an FsObject from a path // // May return nil if an error occurred func (f *FsAcd) newFsObjectWithInfo(remote string, info *acd.Node) fs.Object { o := &FsObjectAcd{ acd: f, remote: remote, } if info != nil { // Set info but not meta o.info = info } else { err := o.readMetaData() // reads info and meta, returning an error if err != nil { // logged already FsDebug("Failed to read info: %s", err) return nil } } return o } // Return an FsObject from a path // // May return nil if an error occurred func (f *FsAcd) NewFsObject(remote string) fs.Object { return f.newFsObjectWithInfo(remote, nil) } // errChk checks the response and if it is a 429 error returns a // RetryError, otherwise it returns just a plain error // // It also implements the backoff strategy func (f *FsAcd) errChk(resp *http.Response, err error) error { if err != nil && resp != nil && resp.StatusCode == 429 { // Update backoff f.backoffLock.Lock() backoff := f.backoff if backoff == 0 { backoff = minBackoff } else { backoff *= 2 if backoff > maxBackoff { backoff = maxBackoff } } f.backoff = backoff f.backoffLock.Unlock() // Sleep for the backoff time sleepTime := time.Duration(rand.Int63n(int64(backoff))) fs.Debug(f, "Retry error: backoff is now %v, sleeping for %v", backoff, sleepTime) time.Sleep(sleepTime) return fs.RetryError(err) } // Reset backoff on success if err == nil { f.backoffLock.Lock() f.backoff = 0 f.backoffLock.Unlock() } return err } // FindLeaf finds a directory of name leaf in the folder with ID pathId func (f *FsAcd) FindLeaf(pathId, leaf string) (pathIdOut string, found bool, err error) { //fs.Debug(f, "FindLeaf(%q, %q)", pathId, leaf) folder := acd.FolderFromId(pathId, f.c.Nodes) subFolder, _, err := folder.GetFolder(leaf) // err = f.errChk(resp, err) if err != nil { if err == acd.ErrorNodeNotFound { //fs.Debug(f, "...Not found") return "", false, nil } //fs.Debug(f, "...Error %v", err) return "", false, err } if subFolder.Status != nil && *subFolder.Status != statusAvailable { fs.Debug(f, "Ignoring folder %q in state %q", *subFolder.Status) time.Sleep(1 * time.Second) // FIXME wait for problem to go away! return "", false, nil } //fs.Debug(f, "...Found(%q, %v)", *subFolder.Id, leaf) return *subFolder.Id, true, nil } // CreateDir makes a directory with pathId as parent and name leaf func (f *FsAcd) CreateDir(pathId, leaf string) (newId string, err error) { //fs.Debug(f, "CreateDir(%q, %q)", pathId, leaf) folder := acd.FolderFromId(pathId, f.c.Nodes) info, _, err := folder.CreateFolder(leaf) // err = f.errChk(resp, err) if err != nil { fs.Debug(f, "...Error %v", err) return "", err } //fs.Debug(f, "...Id %q", *info.Id) return *info.Id, nil } // list the objects into the function supplied // // If directories is set it only sends directories // User function to process a File item from listAll // // Should return true to finish processing type listAllFn func(*acd.Node) bool // Lists the directory required calling the user function on each item found // // If the user fn ever returns true then it early exits with found = true func (f *FsAcd) listAll(dirId string, title string, directoriesOnly bool, filesOnly bool, fn listAllFn) (found bool, err error) { query := "parents:" + dirId if directoriesOnly { query += " AND kind:" + folderKind } else if filesOnly { query += " AND kind:" + fileKind } else { // FIXME none of these work //query += " AND kind:(" + fileKind + " OR " + folderKind + ")" //query += " AND (kind:" + fileKind + " OR kind:" + folderKind + ")" } opts := acd.NodeListOptions{ Filters: query, } var nodes []*acd.Node //var resp *http.Response OUTER: for { nodes, _, err = f.c.Nodes.GetNodes(&opts) // err = f.errChk(resp, err) if err != nil { fs.Stats.Error() fs.ErrorLog(f, "Couldn't list files: %v", err) break } if nodes == nil { break } for _, node := range nodes { if node.Name != nil && node.Id != nil && node.Kind != nil && node.Status != nil { // Ignore nodes if not AVAILABLE if *node.Status != statusAvailable { continue } if fn(node) { found = true break OUTER } } } } return } // Path should be directory path either "" or "path/" // // List the directory using a recursive list from the root // // This fetches the minimum amount of stuff but does more API calls // which makes it slow func (f *FsAcd) listDirRecursive(dirId string, path string, out fs.ObjectsChan) error { var subError error // Make the API request _, err := f.listAll(dirId, "", false, false, func(node *acd.Node) bool { // Recurse on directories // FIXME should do this in parallel // use a wg to sync then collect error switch *node.Kind { case folderKind: subError = f.listDirRecursive(*node.Id, path+*node.Name+"/", out) if subError != nil { return true } case fileKind: if fs := f.newFsObjectWithInfo(path+*node.Name, node); fs != nil { out <- fs } default: // ignore ASSET etc } return false }) if err != nil { return err } if subError != nil { return subError } return nil } // Walk the path returning a channel of FsObjects func (f *FsAcd) List() fs.ObjectsChan { out := make(fs.ObjectsChan, fs.Config.Checkers) go func() { defer close(out) err := f.dirCache.FindRoot(false) if err != nil { fs.Stats.Error() fs.ErrorLog(f, "Couldn't find root: %s", err) } else { err = f.listDirRecursive(f.dirCache.RootID(), "", out) if err != nil { fs.Stats.Error() fs.ErrorLog(f, "List failed: %s", err) } } }() return out } // Lists the containers func (f *FsAcd) ListDir() fs.DirChan { out := make(fs.DirChan, fs.Config.Checkers) go func() { defer close(out) err := f.dirCache.FindRoot(false) if err != nil { fs.Stats.Error() fs.ErrorLog(f, "Couldn't find root: %s", err) } else { _, err := f.listAll(f.dirCache.RootID(), "", true, false, func(item *acd.Node) bool { dir := &fs.Dir{ Name: *item.Name, Bytes: -1, Count: -1, } dir.When, _ = time.Parse(timeFormat, *item.ModifiedDate) out <- dir return false }) if err != nil { fs.Stats.Error() fs.ErrorLog(f, "ListDir failed: %s", err) } } }() return out } // Put the object into the container // // Copy the reader in to the new object which is returned // // The new object may have been created if an error is returned func (f *FsAcd) Put(in io.Reader, remote string, modTime time.Time, size int64) (fs.Object, error) { // Temporary FsObject under construction o := &FsObjectAcd{ acd: f, remote: remote, } leaf, directoryID, err := f.dirCache.FindPath(remote, true) if err != nil { return nil, err } folder := acd.FolderFromId(directoryID, o.acd.c.Nodes) var info *acd.File var resp *http.Response if size != 0 { info, resp, err = folder.Put(in, leaf) } else { info, resp, err = folder.PutSized(in, size, leaf) } err = f.errChk(resp, err) if err != nil { return nil, err } o.info = info.Node return o, nil } // Mkdir creates the container if it doesn't exist func (f *FsAcd) Mkdir() error { return f.dirCache.FindRoot(true) } // purgeCheck remotes the root directory, if check is set then it // refuses to do so if it has anything in func (f *FsAcd) purgeCheck(check bool) error { if f.root == "" { return fmt.Errorf("Can't purge root directory") } dc := f.dirCache err := dc.FindRoot(false) if err != nil { return err } rootID := dc.RootID() if check { // check directory is empty empty := true _, err := f.listAll(rootID, "", false, false, func(node *acd.Node) bool { switch *node.Kind { case folderKind: empty = false return true case fileKind: empty = false return true default: fs.Debug("Found ASSET %s", *node.Id) } return false }) if err != nil { return err } if !empty { return fmt.Errorf("Directory not empty") } } node := acd.NodeFromId(rootID, f.c.Nodes) var resp *http.Response resp, err = node.Trash() err = f.errChk(resp, err) if err != nil { return err } f.dirCache.ResetRoot() if err != nil { return err } return nil } // Rmdir deletes the root folder // // Returns an error if it isn't empty func (f *FsAcd) Rmdir() error { return f.purgeCheck(true) } // Return the precision func (f *FsAcd) Precision() time.Duration { return fs.ModTimeNotSupported } // Copy src to this remote using server side copy operations. // // This is stored with the remote path given // // It returns the destination Object and a possible error // // Will only be called if src.Fs().Name() == f.Name() // // If it isn't possible then return fs.ErrorCantCopy //func (f *FsAcd) Copy(src fs.Object, remote string) (fs.Object, error) { // srcObj, ok := src.(*FsObjectAcd) // if !ok { // fs.Debug(src, "Can't copy - not same remote type") // return nil, fs.ErrorCantCopy // } // srcFs := srcObj.acd // _, err := f.c.ObjectCopy(srcFs.container, srcFs.root+srcObj.remote, f.container, f.root+remote, nil) // if err != nil { // return nil, err // } // return f.NewFsObject(remote), nil //} // Purge deletes all the files and the container // // Optional interface: Only implement this if you have a way of // deleting all the files quicker than just running Remove() on the // result of List() func (f *FsAcd) Purge() error { return f.purgeCheck(false) } // ------------------------------------------------------------ // Return the parent Fs func (o *FsObjectAcd) Fs() fs.Fs { return o.acd } // Return a string version func (o *FsObjectAcd) String() string { if o == nil { return "" } return o.remote } // Return the remote path func (o *FsObjectAcd) Remote() string { return o.remote } // Md5sum returns the Md5sum of an object returning a lowercase hex string func (o *FsObjectAcd) Md5sum() (string, error) { if o.info.ContentProperties.Md5 != nil { return *o.info.ContentProperties.Md5, nil } return "", nil } // Size returns the size of an object in bytes func (o *FsObjectAcd) Size() int64 { return int64(*o.info.ContentProperties.Size) } // readMetaData gets the metadata if it hasn't already been fetched // // it also sets the info func (o *FsObjectAcd) readMetaData() (err error) { if o.info != nil { return nil } leaf, directoryID, err := o.acd.dirCache.FindPath(o.remote, false) if err != nil { return err } folder := acd.FolderFromId(directoryID, o.acd.c.Nodes) info, _, err := folder.GetFile(leaf) // err = o.acd.errChk(resp, err) if err != nil { fs.Debug(o, "Failed to read info: %s", err) return err } o.info = info.Node return nil } // ModTime returns the modification time of the object // // // It attempts to read the objects mtime and if that isn't present the // LastModified returned in the http headers func (o *FsObjectAcd) ModTime() time.Time { err := o.readMetaData() if err != nil { fs.Log(o, "Failed to read metadata: %s", err) return time.Now() } modTime, err := time.Parse(timeFormat, *o.info.ModifiedDate) if err != nil { fs.Log(o, "Failed to read mtime from object: %s", err) return time.Now() } return modTime } // Sets the modification time of the local fs object func (o *FsObjectAcd) SetModTime(modTime time.Time) { // FIXME not implemented return } // Is this object storable func (o *FsObjectAcd) Storable() bool { return true } // Open an object for read func (o *FsObjectAcd) Open() (in io.ReadCloser, err error) { file := acd.File{Node: o.info} var resp *http.Response in, resp, err = file.Open() err = o.acd.errChk(resp, err) return in, err } // Update the object with the contents of the io.Reader, modTime and size // // The new object may have been created if an error is returned func (o *FsObjectAcd) Update(in io.Reader, modTime time.Time, size int64) error { file := acd.File{Node: o.info} var info *acd.File var resp *http.Response var err error if size != 0 { info, resp, err = file.OverwriteSized(in, size) } else { info, resp, err = file.Overwrite(in) } err = o.acd.errChk(resp, err) if err != nil { return err } o.info = info.Node return nil } // Remove an object func (o *FsObjectAcd) Remove() error { resp, err := o.info.Trash() err = o.acd.errChk(resp, err) return err } // Check the interfaces are satisfied var ( _ fs.Fs = (*FsAcd)(nil) _ fs.Purger = (*FsAcd)(nil) // _ fs.Copier = (*FsAcd)(nil) // _ fs.Mover = (*FsAcd)(nil) // _ fs.DirMover = (*FsAcd)(nil) _ fs.Object = (*FsObjectAcd)(nil) )