2015-09-22 17:47:16 +00:00
|
|
|
// Package dropbox provides an interface to Dropbox object storage
|
2014-07-08 20:59:30 +00:00
|
|
|
package dropbox
|
|
|
|
|
2017-05-21 20:35:33 +00:00
|
|
|
// FIXME dropbox for business would be quite easy to add
|
|
|
|
|
2014-07-08 20:59:30 +00:00
|
|
|
/*
|
2017-06-11 21:43:31 +00:00
|
|
|
The Case folding of PathDisplay problem
|
2017-05-21 20:35:33 +00:00
|
|
|
|
2017-06-11 21:43:31 +00:00
|
|
|
From the docs:
|
2014-07-08 20:59:30 +00:00
|
|
|
|
2017-05-21 20:35:33 +00:00
|
|
|
path_display String. The cased path to be used for display purposes
|
|
|
|
only. In rare instances the casing will not correctly match the user's
|
|
|
|
filesystem, but this behavior will match the path provided in the Core
|
|
|
|
API v1, and at least the last path component will have the correct
|
|
|
|
casing. Changes to only the casing of paths won't be returned by
|
|
|
|
list_folder/continue. This field will be null if the file or folder is
|
|
|
|
not mounted. This field is optional.
|
2017-06-05 15:14:24 +00:00
|
|
|
|
2017-08-10 11:06:50 +00:00
|
|
|
We solve this by not implementing the ListR interface. The dropbox
|
|
|
|
remote will recurse directory by directory only using the last element
|
|
|
|
of path_display and all will be well.
|
2014-07-08 20:59:30 +00:00
|
|
|
*/
|
|
|
|
|
|
|
|
import (
|
2019-06-17 08:34:30 +00:00
|
|
|
"context"
|
2014-07-08 20:59:30 +00:00
|
|
|
"fmt"
|
|
|
|
"io"
|
|
|
|
"log"
|
2014-07-14 10:24:04 +00:00
|
|
|
"path"
|
2015-08-20 17:36:06 +00:00
|
|
|
"regexp"
|
2014-07-08 20:59:30 +00:00
|
|
|
"strings"
|
|
|
|
"time"
|
2020-11-27 11:49:37 +00:00
|
|
|
"unicode/utf8"
|
2014-07-08 20:59:30 +00:00
|
|
|
|
2017-08-08 09:13:15 +00:00
|
|
|
"github.com/dropbox/dropbox-sdk-go-unofficial/dropbox"
|
2018-12-05 15:11:41 +00:00
|
|
|
"github.com/dropbox/dropbox-sdk-go-unofficial/dropbox/auth"
|
2018-05-28 09:50:31 +00:00
|
|
|
"github.com/dropbox/dropbox-sdk-go-unofficial/dropbox/common"
|
2017-08-08 09:13:15 +00:00
|
|
|
"github.com/dropbox/dropbox-sdk-go-unofficial/dropbox/files"
|
2018-03-29 07:10:19 +00:00
|
|
|
"github.com/dropbox/dropbox-sdk-go-unofficial/dropbox/sharing"
|
2018-09-18 00:35:12 +00:00
|
|
|
"github.com/dropbox/dropbox-sdk-go-unofficial/dropbox/team"
|
2018-04-16 21:19:25 +00:00
|
|
|
"github.com/dropbox/dropbox-sdk-go-unofficial/dropbox/users"
|
2016-06-12 14:06:02 +00:00
|
|
|
"github.com/pkg/errors"
|
2019-09-23 13:32:36 +00:00
|
|
|
"github.com/rclone/rclone/backend/dropbox/dbhash"
|
2019-07-28 17:47:38 +00:00
|
|
|
"github.com/rclone/rclone/fs"
|
|
|
|
"github.com/rclone/rclone/fs/config"
|
|
|
|
"github.com/rclone/rclone/fs/config/configmap"
|
|
|
|
"github.com/rclone/rclone/fs/config/configstruct"
|
|
|
|
"github.com/rclone/rclone/fs/config/obscure"
|
|
|
|
"github.com/rclone/rclone/fs/fserrors"
|
|
|
|
"github.com/rclone/rclone/fs/hash"
|
2020-01-14 17:33:35 +00:00
|
|
|
"github.com/rclone/rclone/lib/encoder"
|
2019-07-28 17:47:38 +00:00
|
|
|
"github.com/rclone/rclone/lib/oauthutil"
|
|
|
|
"github.com/rclone/rclone/lib/pacer"
|
|
|
|
"github.com/rclone/rclone/lib/readers"
|
2017-05-28 16:55:18 +00:00
|
|
|
"golang.org/x/oauth2"
|
2014-07-08 20:59:30 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
// Constants
|
|
|
|
const (
|
2017-05-21 20:35:33 +00:00
|
|
|
rcloneClientID = "5jcck7diasz0rqy"
|
|
|
|
rcloneEncryptedClientSecret = "fRS5vVLr2v6FbyXYnIgjwBuUAt0osq_QZTXAEcmZ7g"
|
2017-05-28 16:55:18 +00:00
|
|
|
minSleep = 10 * time.Millisecond
|
|
|
|
maxSleep = 2 * time.Second
|
|
|
|
decayConstant = 2 // bigger for slower decay, exponential
|
2015-08-25 18:01:37 +00:00
|
|
|
// Upload chunk size - setting too small makes uploads slow.
|
2017-11-21 17:30:50 +00:00
|
|
|
// Chunks are buffered into memory for retries.
|
|
|
|
//
|
|
|
|
// Speed vs chunk size uploading a 1 GB file on 2017-11-22
|
|
|
|
//
|
|
|
|
// Chunk Size MB, Speed Mbyte/s, % of max
|
|
|
|
// 1 1.364 11%
|
|
|
|
// 2 2.443 19%
|
|
|
|
// 4 4.288 33%
|
|
|
|
// 8 6.79 52%
|
|
|
|
// 16 8.916 69%
|
|
|
|
// 24 10.195 79%
|
|
|
|
// 32 10.427 81%
|
|
|
|
// 40 10.96 85%
|
|
|
|
// 48 11.828 91%
|
|
|
|
// 56 11.763 91%
|
|
|
|
// 64 12.047 93%
|
|
|
|
// 96 12.302 95%
|
|
|
|
// 128 12.945 100%
|
|
|
|
//
|
|
|
|
// Choose 48MB which is 91% of Maximum speed. rclone by
|
|
|
|
// default does 4 transfers so this should use 4*48MB = 192MB
|
|
|
|
// by default.
|
2018-09-07 11:02:27 +00:00
|
|
|
defaultChunkSize = 48 * fs.MebiByte
|
|
|
|
maxChunkSize = 150 * fs.MebiByte
|
2020-11-27 11:49:37 +00:00
|
|
|
// Max length of filename parts: https://help.dropbox.com/installs-integrations/sync-uploads/files-not-syncing
|
|
|
|
maxFileNameLength = 255
|
2018-05-14 17:06:57 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
var (
|
|
|
|
// Description of how to auth for this app
|
|
|
|
dropboxConfig = &oauth2.Config{
|
|
|
|
Scopes: []string{},
|
|
|
|
// Endpoint: oauth2.Endpoint{
|
|
|
|
// AuthURL: "https://www.dropbox.com/1/oauth2/authorize",
|
|
|
|
// TokenURL: "https://api.dropboxapi.com/1/oauth2/token",
|
|
|
|
// },
|
|
|
|
Endpoint: dropbox.OAuthEndpoint(""),
|
|
|
|
ClientID: rcloneClientID,
|
|
|
|
ClientSecret: obscure.MustReveal(rcloneEncryptedClientSecret),
|
|
|
|
RedirectURL: oauthutil.RedirectLocalhostURL,
|
|
|
|
}
|
|
|
|
// A regexp matching path names for files Dropbox ignores
|
|
|
|
// See https://www.dropbox.com/en/help/145 - Ignored files
|
|
|
|
ignoredFiles = regexp.MustCompile(`(?i)(^|/)(desktop\.ini|thumbs\.db|\.ds_store|icon\r|\.dropbox|\.dropbox.attr)$`)
|
2019-09-23 13:32:36 +00:00
|
|
|
|
|
|
|
// DbHashType is the hash.Type for Dropbox
|
|
|
|
DbHashType hash.Type
|
2020-11-26 11:58:43 +00:00
|
|
|
|
|
|
|
// Errors
|
|
|
|
errNotSupportedInSharedMode = fserrors.NoRetryError(errors.New("not supported in shared files mode"))
|
2015-08-25 18:01:37 +00:00
|
|
|
)
|
2015-08-20 17:36:06 +00:00
|
|
|
|
2014-07-08 20:59:30 +00:00
|
|
|
// Register with Fs
|
|
|
|
func init() {
|
2019-10-30 10:31:32 +00:00
|
|
|
DbHashType = hash.RegisterHash("DropboxHash", 64, dbhash.New)
|
2016-02-18 11:35:25 +00:00
|
|
|
fs.Register(&fs.RegInfo{
|
2016-02-15 18:11:53 +00:00
|
|
|
Name: "dropbox",
|
|
|
|
Description: "Dropbox",
|
|
|
|
NewFs: NewFs,
|
2020-11-05 18:02:26 +00:00
|
|
|
Config: func(ctx context.Context, name string, m configmap.Mapper) {
|
2020-05-25 14:06:08 +00:00
|
|
|
opt := oauthutil.Options{
|
|
|
|
NoOffline: true,
|
2020-11-23 10:38:52 +00:00
|
|
|
OAuth2Opts: []oauth2.AuthCodeOption{
|
|
|
|
oauth2.SetAuthURLParam("token_access_type", "offline"),
|
|
|
|
},
|
2020-05-25 14:06:08 +00:00
|
|
|
}
|
2020-11-05 18:02:26 +00:00
|
|
|
err := oauthutil.Config(ctx, "dropbox", name, m, dropboxConfig, &opt)
|
2017-05-21 20:35:33 +00:00
|
|
|
if err != nil {
|
|
|
|
log.Fatalf("Failed to configure token: %v", err)
|
|
|
|
}
|
|
|
|
},
|
2020-08-01 23:32:21 +00:00
|
|
|
Options: append(oauthutil.SharedOptions, []fs.Option{{
|
2018-10-01 17:36:15 +00:00
|
|
|
Name: "chunk_size",
|
|
|
|
Help: fmt.Sprintf(`Upload chunk size. (< %v).
|
|
|
|
|
|
|
|
Any files larger than this will be uploaded in chunks of this size.
|
|
|
|
|
|
|
|
Note that chunks are buffered in memory (one at a time) so rclone can
|
|
|
|
deal with retries. Setting this larger will increase the speed
|
|
|
|
slightly (at most 10%% for 128MB in tests) at the cost of using more
|
2019-01-11 17:17:46 +00:00
|
|
|
memory. It can be set smaller if you are tight on memory.`, maxChunkSize),
|
|
|
|
Default: defaultChunkSize,
|
2018-05-14 17:06:57 +00:00
|
|
|
Advanced: true,
|
2018-09-18 00:35:12 +00:00
|
|
|
}, {
|
|
|
|
Name: "impersonate",
|
|
|
|
Help: "Impersonate this user when using a business account.",
|
|
|
|
Default: "",
|
|
|
|
Advanced: true,
|
2020-08-23 14:10:30 +00:00
|
|
|
}, {
|
|
|
|
Name: "shared_files",
|
|
|
|
Help: `Instructs rclone to work on individual shared files.
|
|
|
|
|
|
|
|
In this mode rclone's features are extremely limited - only list (ls, lsl, etc.)
|
|
|
|
operations and read operations (e.g. downloading) are supported in this mode.
|
|
|
|
All other operations will be disabled.`,
|
|
|
|
Default: false,
|
|
|
|
Advanced: true,
|
|
|
|
}, {
|
|
|
|
Name: "shared_folders",
|
|
|
|
Help: `Instructs rclone to work on shared folders.
|
|
|
|
|
|
|
|
When this flag is used with no path only the List operation is supported and
|
|
|
|
all available shared folders will be listed. If you specify a path the first part
|
|
|
|
will be interpreted as the name of shared folder. Rclone will then try to mount this
|
|
|
|
shared to the root namespace. On success shared folder rclone proceeds normally.
|
|
|
|
The shared folder is now pretty much a normal folder and all normal operations
|
|
|
|
are supported.
|
|
|
|
|
|
|
|
Note that we don't unmount the shared folder afterwards so the
|
|
|
|
--dropbox-shared-folders can be omitted after the first use of a particular
|
|
|
|
shared folder.`,
|
|
|
|
Default: false,
|
|
|
|
Advanced: true,
|
2020-01-14 17:33:35 +00:00
|
|
|
}, {
|
|
|
|
Name: config.ConfigEncoding,
|
|
|
|
Help: config.ConfigEncodingHelp,
|
|
|
|
Advanced: true,
|
2020-01-14 21:51:49 +00:00
|
|
|
// https://www.dropbox.com/help/syncing-uploads/files-not-syncing lists / and \
|
|
|
|
// as invalid characters.
|
|
|
|
// Testing revealed names with trailing spaces and the DEL character don't work.
|
|
|
|
// Also encode invalid UTF-8 bytes as json doesn't handle them properly.
|
|
|
|
Default: (encoder.Base |
|
|
|
|
encoder.EncodeBackSlash |
|
|
|
|
encoder.EncodeDel |
|
|
|
|
encoder.EncodeRightSpace |
|
|
|
|
encoder.EncodeInvalidUtf8),
|
2020-08-01 23:32:21 +00:00
|
|
|
}}...),
|
2014-07-08 20:59:30 +00:00
|
|
|
})
|
2018-05-14 17:06:57 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Options defines the configuration for this backend
|
|
|
|
type Options struct {
|
2020-08-23 14:10:30 +00:00
|
|
|
ChunkSize fs.SizeSuffix `config:"chunk_size"`
|
|
|
|
Impersonate string `config:"impersonate"`
|
|
|
|
SharedFiles bool `config:"shared_files"`
|
|
|
|
SharedFolders bool `config:"shared_folders"`
|
|
|
|
Enc encoder.MultiEncoder `config:"encoding"`
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
2015-11-07 11:14:46 +00:00
|
|
|
// Fs represents a remote dropbox server
|
|
|
|
type Fs struct {
|
2018-03-29 07:10:19 +00:00
|
|
|
name string // name of this remote
|
|
|
|
root string // the path we are working on
|
2018-05-14 17:06:57 +00:00
|
|
|
opt Options // parsed options
|
2018-03-29 07:10:19 +00:00
|
|
|
features *fs.Features // optional features
|
|
|
|
srv files.Client // the connection to the dropbox server
|
2018-05-28 09:50:31 +00:00
|
|
|
sharing sharing.Client // as above, but for generating sharing links
|
2018-04-16 21:19:25 +00:00
|
|
|
users users.Client // as above, but for accessing user information
|
2018-09-18 00:35:12 +00:00
|
|
|
team team.Client // for the Teams API
|
2018-03-29 07:10:19 +00:00
|
|
|
slashRoot string // root with "/" prefix, lowercase
|
|
|
|
slashRootSlash string // root with "/" prefix and postfix, lowercase
|
2019-02-09 20:52:15 +00:00
|
|
|
pacer *fs.Pacer // To pace the API calls
|
2018-05-28 09:50:31 +00:00
|
|
|
ns string // The namespace we are using or "" for none
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
2015-11-07 11:14:46 +00:00
|
|
|
// Object describes a dropbox object
|
2017-05-26 14:09:31 +00:00
|
|
|
//
|
|
|
|
// Dropbox Objects always have full metadata
|
2015-11-07 11:14:46 +00:00
|
|
|
type Object struct {
|
2020-08-23 14:10:30 +00:00
|
|
|
fs *Fs // what this object is part of
|
|
|
|
id string
|
|
|
|
url string
|
2017-05-26 14:09:31 +00:00
|
|
|
remote string // The remote path
|
|
|
|
bytes int64 // size of the object
|
|
|
|
modTime time.Time // time it was last modified
|
|
|
|
hash string // content_hash of the object
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// ------------------------------------------------------------
|
|
|
|
|
2015-09-22 17:47:16 +00:00
|
|
|
// Name of the remote (as passed into NewFs)
|
2015-11-07 11:14:46 +00:00
|
|
|
func (f *Fs) Name() string {
|
2015-08-22 15:53:11 +00:00
|
|
|
return f.name
|
|
|
|
}
|
|
|
|
|
2015-09-22 17:47:16 +00:00
|
|
|
// Root of the remote (as passed into NewFs)
|
2015-11-07 11:14:46 +00:00
|
|
|
func (f *Fs) Root() string {
|
2015-09-01 19:45:27 +00:00
|
|
|
return f.root
|
|
|
|
}
|
|
|
|
|
2015-11-07 11:14:46 +00:00
|
|
|
// String converts this Fs to a string
|
|
|
|
func (f *Fs) String() string {
|
2014-07-08 20:59:30 +00:00
|
|
|
return fmt.Sprintf("Dropbox root '%s'", f.root)
|
|
|
|
}
|
|
|
|
|
2017-01-13 17:21:47 +00:00
|
|
|
// Features returns the optional features of this Fs
|
|
|
|
func (f *Fs) Features() *fs.Features {
|
|
|
|
return f.features
|
|
|
|
}
|
|
|
|
|
2017-05-28 16:55:18 +00:00
|
|
|
// shouldRetry returns a boolean as to whether this err deserves to be
|
|
|
|
// retried. It returns the err as a convenience
|
|
|
|
func shouldRetry(err error) (bool, error) {
|
|
|
|
if err == nil {
|
|
|
|
return false, err
|
|
|
|
}
|
|
|
|
baseErrString := errors.Cause(err).Error()
|
2020-11-26 12:10:46 +00:00
|
|
|
// First check for specific errors
|
2020-03-26 08:45:35 +00:00
|
|
|
if strings.Contains(baseErrString, "insufficient_space") {
|
|
|
|
return false, fserrors.FatalError(err)
|
2020-11-26 12:10:46 +00:00
|
|
|
} else if strings.Contains(baseErrString, "malformed_path") {
|
|
|
|
return false, fserrors.NoRetryError(err)
|
2020-03-26 08:45:35 +00:00
|
|
|
}
|
|
|
|
// Then handle any official Retry-After header from Dropbox's SDK
|
2018-12-05 15:11:41 +00:00
|
|
|
switch e := err.(type) {
|
|
|
|
case auth.RateLimitAPIError:
|
|
|
|
if e.RateLimitError.RetryAfter > 0 {
|
|
|
|
fs.Debugf(baseErrString, "Too many requests or write operations. Trying again in %d seconds.", e.RateLimitError.RetryAfter)
|
2019-02-09 20:52:15 +00:00
|
|
|
err = pacer.RetryAfterError(err, time.Duration(e.RateLimitError.RetryAfter)*time.Second)
|
2018-12-05 15:11:41 +00:00
|
|
|
}
|
|
|
|
return true, err
|
|
|
|
}
|
2019-02-07 17:41:17 +00:00
|
|
|
// Keep old behavior for backward compatibility
|
2019-02-06 10:23:59 +00:00
|
|
|
if strings.Contains(baseErrString, "too_many_write_operations") || strings.Contains(baseErrString, "too_many_requests") || baseErrString == "" {
|
2017-05-28 16:55:18 +00:00
|
|
|
return true, err
|
|
|
|
}
|
2018-01-12 16:30:54 +00:00
|
|
|
return fserrors.ShouldRetry(err), err
|
2017-05-28 16:55:18 +00:00
|
|
|
}
|
|
|
|
|
2018-09-07 11:02:27 +00:00
|
|
|
func checkUploadChunkSize(cs fs.SizeSuffix) error {
|
|
|
|
const minChunkSize = fs.Byte
|
|
|
|
if cs < minChunkSize {
|
|
|
|
return errors.Errorf("%s is less than %s", cs, minChunkSize)
|
|
|
|
}
|
|
|
|
if cs > maxChunkSize {
|
|
|
|
return errors.Errorf("%s is greater than %s", cs, maxChunkSize)
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (f *Fs) setUploadChunkSize(cs fs.SizeSuffix) (old fs.SizeSuffix, err error) {
|
|
|
|
err = checkUploadChunkSize(cs)
|
|
|
|
if err == nil {
|
|
|
|
old, f.opt.ChunkSize = f.opt.ChunkSize, cs
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2019-02-07 17:41:17 +00:00
|
|
|
// NewFs constructs an Fs from the path, container:path
|
2020-11-05 15:18:51 +00:00
|
|
|
func NewFs(ctx context.Context, name, root string, m configmap.Mapper) (fs.Fs, error) {
|
2018-05-14 17:06:57 +00:00
|
|
|
// Parse config into Options struct
|
|
|
|
opt := new(Options)
|
|
|
|
err := configstruct.Set(m, opt)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
2018-09-07 11:02:27 +00:00
|
|
|
err = checkUploadChunkSize(opt.ChunkSize)
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "dropbox: chunk size")
|
2015-08-25 18:01:37 +00:00
|
|
|
}
|
2017-05-21 20:35:33 +00:00
|
|
|
|
|
|
|
// Convert the old token if it exists. The old token was just
|
|
|
|
// just a string, the new one is a JSON blob
|
2018-05-14 17:06:57 +00:00
|
|
|
oldToken, ok := m.Get(config.ConfigToken)
|
|
|
|
oldToken = strings.TrimSpace(oldToken)
|
|
|
|
if ok && oldToken != "" && oldToken[0] != '{' {
|
2017-05-21 20:35:33 +00:00
|
|
|
fs.Infof(name, "Converting token to new format")
|
|
|
|
newToken := fmt.Sprintf(`{"access_token":"%s","token_type":"bearer","expiry":"0001-01-01T00:00:00Z"}`, oldToken)
|
2018-01-12 16:30:54 +00:00
|
|
|
err := config.SetValueAndSave(name, config.ConfigToken, newToken)
|
2017-05-21 20:35:33 +00:00
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "NewFS convert token")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-11-05 18:02:26 +00:00
|
|
|
oAuthClient, _, err := oauthutil.NewClient(ctx, name, m, dropboxConfig)
|
2015-09-22 06:31:12 +00:00
|
|
|
if err != nil {
|
2018-05-28 09:50:31 +00:00
|
|
|
return nil, errors.Wrap(err, "failed to configure dropbox")
|
2017-05-21 20:35:33 +00:00
|
|
|
}
|
|
|
|
|
2015-11-07 11:14:46 +00:00
|
|
|
f := &Fs{
|
2018-05-28 09:50:31 +00:00
|
|
|
name: name,
|
2018-05-14 17:06:57 +00:00
|
|
|
opt: *opt,
|
2020-11-05 11:33:32 +00:00
|
|
|
pacer: fs.NewPacer(ctx, pacer.NewDefault(pacer.MinSleep(minSleep), pacer.MaxSleep(maxSleep), pacer.DecayConstant(decayConstant))),
|
2018-05-28 09:50:31 +00:00
|
|
|
}
|
|
|
|
config := dropbox.Config{
|
|
|
|
LogLevel: dropbox.LogOff, // logging in the SDK: LogOff, LogDebug, LogInfo
|
|
|
|
Client: oAuthClient, // maybe???
|
|
|
|
HeaderGenerator: f.headerGenerator,
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
2018-09-18 00:35:12 +00:00
|
|
|
|
|
|
|
// NOTE: needs to be created pre-impersonation so we can look up the impersonated user
|
|
|
|
f.team = team.New(config)
|
|
|
|
|
|
|
|
if opt.Impersonate != "" {
|
|
|
|
|
|
|
|
user := team.UserSelectorArg{
|
|
|
|
Email: opt.Impersonate,
|
|
|
|
}
|
|
|
|
user.Tag = "email"
|
|
|
|
|
|
|
|
members := []*team.UserSelectorArg{&user}
|
|
|
|
args := team.NewMembersGetInfoArgs(members)
|
|
|
|
|
|
|
|
memberIds, err := f.team.MembersGetInfo(args)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrapf(err, "invalid dropbox team member: %q", opt.Impersonate)
|
|
|
|
}
|
|
|
|
|
|
|
|
config.AsMemberID = memberIds[0].MemberInfo.Profile.MemberProfile.TeamMemberId
|
|
|
|
}
|
|
|
|
|
2018-05-28 09:50:31 +00:00
|
|
|
f.srv = files.New(config)
|
|
|
|
f.sharing = sharing.New(config)
|
|
|
|
f.users = users.New(config)
|
2017-08-09 14:27:43 +00:00
|
|
|
f.features = (&fs.Features{
|
|
|
|
CaseInsensitive: true,
|
2020-11-29 15:11:03 +00:00
|
|
|
ReadMimeType: false,
|
2017-08-09 14:27:43 +00:00
|
|
|
CanHaveEmptyDirectories: true,
|
2020-08-23 14:10:30 +00:00
|
|
|
})
|
|
|
|
|
|
|
|
// do not fill features yet
|
|
|
|
if f.opt.SharedFiles {
|
|
|
|
f.setRoot(root)
|
|
|
|
if f.root == "" {
|
|
|
|
return f, nil
|
|
|
|
}
|
|
|
|
_, err := f.findSharedFile(f.root)
|
|
|
|
f.root = ""
|
|
|
|
if err == nil {
|
|
|
|
return f, fs.ErrorIsFile
|
|
|
|
}
|
|
|
|
return f, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
if f.opt.SharedFolders {
|
|
|
|
f.setRoot(root)
|
|
|
|
if f.root == "" {
|
|
|
|
return f, nil // our root it empty so we probably want to list shared folders
|
|
|
|
}
|
|
|
|
|
|
|
|
dir := path.Dir(f.root)
|
|
|
|
if dir == "." {
|
|
|
|
dir = f.root
|
|
|
|
}
|
|
|
|
|
|
|
|
// root is not empty so we have find the right shared folder if it exists
|
|
|
|
id, err := f.findSharedFolder(dir)
|
|
|
|
if err != nil {
|
|
|
|
// if we didn't find the specified shared folder we have to bail out here
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
// we found the specified shared folder so let's mount it
|
|
|
|
// this will add it to the users normal root namespace and allows us
|
|
|
|
// to actually perform operations on it using the normal api endpoints.
|
|
|
|
err = f.mountSharedFolder(id)
|
|
|
|
if err != nil {
|
|
|
|
switch e := err.(type) {
|
|
|
|
case sharing.MountFolderAPIError:
|
|
|
|
if e.EndpointError == nil || (e.EndpointError != nil && e.EndpointError.Tag != sharing.MountFolderErrorAlreadyMounted) {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
// if the moint failed we have to abort here
|
|
|
|
}
|
|
|
|
// if the mount succeeded it's now a normal folder in the users root namespace
|
|
|
|
// we disable shared folder mode and proceed normally
|
|
|
|
f.opt.SharedFolders = false
|
|
|
|
}
|
|
|
|
|
2020-11-05 16:00:40 +00:00
|
|
|
f.features.Fill(ctx, f)
|
2014-07-08 20:59:30 +00:00
|
|
|
|
2018-05-28 09:50:31 +00:00
|
|
|
// If root starts with / then use the actual root
|
|
|
|
if strings.HasPrefix(root, "/") {
|
|
|
|
var acc *users.FullAccount
|
|
|
|
err = f.pacer.Call(func() (bool, error) {
|
|
|
|
acc, err = f.users.GetCurrentAccount()
|
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "get current account failed")
|
|
|
|
}
|
|
|
|
switch x := acc.RootInfo.(type) {
|
|
|
|
case *common.TeamRootInfo:
|
|
|
|
f.ns = x.RootNamespaceId
|
|
|
|
case *common.UserRootInfo:
|
|
|
|
f.ns = x.RootNamespaceId
|
|
|
|
default:
|
|
|
|
return nil, errors.Errorf("unknown RootInfo type %v %T", acc.RootInfo, acc.RootInfo)
|
|
|
|
}
|
|
|
|
fs.Debugf(f, "Using root namespace %q", f.ns)
|
|
|
|
}
|
2020-08-23 14:10:30 +00:00
|
|
|
f.setRoot(root)
|
2018-05-28 09:50:31 +00:00
|
|
|
|
2014-07-14 10:24:04 +00:00
|
|
|
// See if the root is actually an object
|
2017-05-21 20:35:33 +00:00
|
|
|
_, err = f.getFileMetadata(f.slashRoot)
|
|
|
|
if err == nil {
|
2014-07-14 10:24:04 +00:00
|
|
|
newRoot := path.Dir(f.root)
|
|
|
|
if newRoot == "." {
|
|
|
|
newRoot = ""
|
2014-07-12 11:38:30 +00:00
|
|
|
}
|
2014-07-14 10:24:04 +00:00
|
|
|
f.setRoot(newRoot)
|
2016-06-21 17:01:53 +00:00
|
|
|
// return an error with an fs which points to the parent
|
|
|
|
return f, fs.ErrorIsFile
|
2014-07-14 10:24:04 +00:00
|
|
|
}
|
2014-07-08 20:59:30 +00:00
|
|
|
return f, nil
|
|
|
|
}
|
|
|
|
|
2018-05-28 09:50:31 +00:00
|
|
|
// headerGenerator for dropbox sdk
|
|
|
|
func (f *Fs) headerGenerator(hostType string, style string, namespace string, route string) map[string]string {
|
|
|
|
if f.ns == "" {
|
|
|
|
return map[string]string{}
|
|
|
|
}
|
|
|
|
return map[string]string{
|
|
|
|
"Dropbox-API-Path-Root": `{".tag": "namespace_id", "namespace_id": "` + f.ns + `"}`,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-07-14 10:24:04 +00:00
|
|
|
// Sets root in f
|
2015-11-07 11:14:46 +00:00
|
|
|
func (f *Fs) setRoot(root string) {
|
2014-07-14 10:24:04 +00:00
|
|
|
f.root = strings.Trim(root, "/")
|
2018-04-05 14:28:40 +00:00
|
|
|
f.slashRoot = "/" + f.root
|
2014-07-14 10:24:04 +00:00
|
|
|
f.slashRootSlash = f.slashRoot
|
2018-04-05 14:28:40 +00:00
|
|
|
if f.root != "" {
|
2014-07-14 10:24:04 +00:00
|
|
|
f.slashRootSlash += "/"
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-05-21 20:35:33 +00:00
|
|
|
// getMetadata gets the metadata for a file or directory
|
|
|
|
func (f *Fs) getMetadata(objPath string) (entry files.IsMetadata, notFound bool, err error) {
|
2017-05-28 16:55:18 +00:00
|
|
|
err = f.pacer.Call(func() (bool, error) {
|
2018-11-02 12:14:02 +00:00
|
|
|
entry, err = f.srv.GetMetadata(&files.GetMetadataArg{
|
2020-01-14 17:33:35 +00:00
|
|
|
Path: f.opt.Enc.FromStandardPath(objPath),
|
2018-11-02 12:14:02 +00:00
|
|
|
})
|
2017-05-28 16:55:18 +00:00
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
2017-05-21 20:35:33 +00:00
|
|
|
if err != nil {
|
|
|
|
switch e := err.(type) {
|
|
|
|
case files.GetMetadataAPIError:
|
2019-10-10 15:34:09 +00:00
|
|
|
if e.EndpointError != nil && e.EndpointError.Path != nil && e.EndpointError.Path.Tag == files.LookupErrorNotFound {
|
2017-05-21 20:35:33 +00:00
|
|
|
notFound = true
|
|
|
|
err = nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// getFileMetadata gets the metadata for a file
|
|
|
|
func (f *Fs) getFileMetadata(filePath string) (fileInfo *files.FileMetadata, err error) {
|
|
|
|
entry, notFound, err := f.getMetadata(filePath)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
if notFound {
|
|
|
|
return nil, fs.ErrorObjectNotFound
|
|
|
|
}
|
|
|
|
fileInfo, ok := entry.(*files.FileMetadata)
|
|
|
|
if !ok {
|
|
|
|
return nil, fs.ErrorNotAFile
|
|
|
|
}
|
|
|
|
return fileInfo, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// getDirMetadata gets the metadata for a directory
|
|
|
|
func (f *Fs) getDirMetadata(dirPath string) (dirInfo *files.FolderMetadata, err error) {
|
|
|
|
entry, notFound, err := f.getMetadata(dirPath)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
if notFound {
|
|
|
|
return nil, fs.ErrorDirNotFound
|
|
|
|
}
|
|
|
|
dirInfo, ok := entry.(*files.FolderMetadata)
|
|
|
|
if !ok {
|
|
|
|
return nil, fs.ErrorIsFile
|
|
|
|
}
|
|
|
|
return dirInfo, nil
|
|
|
|
}
|
|
|
|
|
2016-06-25 20:58:34 +00:00
|
|
|
// Return an Object from a path
|
2014-07-29 16:50:07 +00:00
|
|
|
//
|
2016-06-25 20:23:20 +00:00
|
|
|
// If it can't be found it returns the error fs.ErrorObjectNotFound.
|
2017-05-21 20:35:33 +00:00
|
|
|
func (f *Fs) newObjectWithInfo(remote string, info *files.FileMetadata) (fs.Object, error) {
|
2017-02-25 15:23:27 +00:00
|
|
|
o := &Object{
|
2015-11-07 11:14:46 +00:00
|
|
|
fs: f,
|
|
|
|
remote: remote,
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
2017-02-25 15:23:27 +00:00
|
|
|
var err error
|
2014-07-12 10:46:45 +00:00
|
|
|
if info != nil {
|
2017-02-25 11:09:57 +00:00
|
|
|
err = o.setMetadataFromEntry(info)
|
2014-07-08 20:59:30 +00:00
|
|
|
} else {
|
2017-02-25 11:09:57 +00:00
|
|
|
err = o.readEntryAndSetMetadata()
|
|
|
|
}
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
2016-06-25 20:23:20 +00:00
|
|
|
return o, nil
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
2016-06-25 20:23:20 +00:00
|
|
|
// NewObject finds the Object at remote. If it can't be found
|
|
|
|
// it returns the error fs.ErrorObjectNotFound.
|
2019-06-17 08:34:30 +00:00
|
|
|
func (f *Fs) NewObject(ctx context.Context, remote string) (fs.Object, error) {
|
2020-08-23 14:10:30 +00:00
|
|
|
if f.opt.SharedFiles {
|
|
|
|
return f.findSharedFile(remote)
|
|
|
|
}
|
2016-06-25 20:58:34 +00:00
|
|
|
return f.newObjectWithInfo(remote, nil)
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
2020-08-23 14:10:30 +00:00
|
|
|
// listSharedFoldersApi lists all available shared folders mounted and not mounted
|
|
|
|
// we'll need the id later so we have to return them in original format
|
2020-08-26 23:12:10 +00:00
|
|
|
func (f *Fs) listSharedFolders() (entries fs.DirEntries, err error) {
|
2020-08-23 14:10:30 +00:00
|
|
|
started := false
|
|
|
|
var res *sharing.ListFoldersResult
|
|
|
|
for {
|
|
|
|
if !started {
|
|
|
|
arg := sharing.ListFoldersArgs{
|
|
|
|
Limit: 100,
|
|
|
|
}
|
|
|
|
err := f.pacer.Call(func() (bool, error) {
|
|
|
|
res, err = f.sharing.ListFolders(&arg)
|
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
started = true
|
|
|
|
} else {
|
|
|
|
arg := sharing.ListFoldersContinueArg{
|
|
|
|
Cursor: res.Cursor,
|
|
|
|
}
|
|
|
|
err := f.pacer.Call(func() (bool, error) {
|
|
|
|
res, err = f.sharing.ListFoldersContinue(&arg)
|
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "list continue")
|
|
|
|
}
|
|
|
|
}
|
2020-08-26 23:12:10 +00:00
|
|
|
for _, entry := range res.Entries {
|
|
|
|
leaf := f.opt.Enc.ToStandardName(entry.Name)
|
|
|
|
d := fs.NewDir(leaf, time.Now()).SetID(entry.SharedFolderId)
|
|
|
|
entries = append(entries, d)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
}
|
2020-08-23 14:10:30 +00:00
|
|
|
if res.Cursor == "" {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return entries, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// findSharedFolder find the id for a given shared folder name
|
|
|
|
// somewhat annoyingly there is no endpoint to query a shared folder by it's name
|
|
|
|
// so our only option is to iterate over all shared folders
|
|
|
|
func (f *Fs) findSharedFolder(name string) (id string, err error) {
|
2020-08-26 23:12:10 +00:00
|
|
|
entries, err := f.listSharedFolders()
|
2020-08-23 14:10:30 +00:00
|
|
|
if err != nil {
|
|
|
|
return "", err
|
|
|
|
}
|
|
|
|
for _, entry := range entries {
|
2020-08-26 23:12:10 +00:00
|
|
|
if entry.(*fs.Dir).Remote() == name {
|
|
|
|
return entry.(*fs.Dir).ID(), nil
|
2020-08-23 14:10:30 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return "", fs.ErrorDirNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
// mountSharedFolders mount a shared folder to the root namespace
|
|
|
|
func (f *Fs) mountSharedFolder(id string) error {
|
|
|
|
arg := sharing.MountFolderArg{
|
|
|
|
SharedFolderId: id,
|
|
|
|
}
|
|
|
|
err := f.pacer.Call(func() (bool, error) {
|
|
|
|
_, err := f.sharing.MountFolder(&arg)
|
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// listSharedFolders lists shared the user as access to (note this means individual
|
|
|
|
// files not files contained in shared folders)
|
|
|
|
func (f *Fs) listReceivedFiles() (entries fs.DirEntries, err error) {
|
|
|
|
started := false
|
|
|
|
var res *sharing.ListFilesResult
|
|
|
|
for {
|
|
|
|
if !started {
|
|
|
|
arg := sharing.ListFilesArg{
|
|
|
|
Limit: 100,
|
|
|
|
}
|
|
|
|
err := f.pacer.Call(func() (bool, error) {
|
|
|
|
res, err = f.sharing.ListReceivedFiles(&arg)
|
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
started = true
|
|
|
|
} else {
|
|
|
|
arg := sharing.ListFilesContinueArg{
|
|
|
|
Cursor: res.Cursor,
|
|
|
|
}
|
|
|
|
err := f.pacer.Call(func() (bool, error) {
|
|
|
|
res, err = f.sharing.ListReceivedFilesContinue(&arg)
|
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "list continue")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for _, entry := range res.Entries {
|
|
|
|
fmt.Printf("%+v\n", entry)
|
|
|
|
entryPath := entry.Name
|
|
|
|
o := &Object{
|
|
|
|
fs: f,
|
|
|
|
url: entry.PreviewUrl,
|
|
|
|
remote: entryPath,
|
|
|
|
modTime: entry.TimeInvited,
|
|
|
|
}
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
entries = append(entries, o)
|
|
|
|
}
|
|
|
|
if res.Cursor == "" {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return entries, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (f *Fs) findSharedFile(name string) (o *Object, err error) {
|
|
|
|
files, err := f.listReceivedFiles()
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
for _, entry := range files {
|
|
|
|
if entry.(*Object).remote == name {
|
|
|
|
return entry.(*Object), nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil, fs.ErrorObjectNotFound
|
|
|
|
}
|
|
|
|
|
2017-06-11 21:43:31 +00:00
|
|
|
// List the objects and directories in dir into entries. The
|
|
|
|
// entries can be returned in any order but should be for a
|
|
|
|
// complete directory.
|
|
|
|
//
|
|
|
|
// dir should be "" to list the root, and should not have
|
|
|
|
// trailing slashes.
|
|
|
|
//
|
|
|
|
// This should return ErrDirNotFound if the directory isn't
|
|
|
|
// found.
|
2019-06-17 08:34:30 +00:00
|
|
|
func (f *Fs) List(ctx context.Context, dir string) (entries fs.DirEntries, err error) {
|
2020-08-23 14:10:30 +00:00
|
|
|
if f.opt.SharedFiles {
|
|
|
|
return f.listReceivedFiles()
|
|
|
|
}
|
|
|
|
if f.opt.SharedFolders {
|
|
|
|
return f.listSharedFolders()
|
|
|
|
}
|
|
|
|
|
2016-04-23 20:46:52 +00:00
|
|
|
root := f.slashRoot
|
|
|
|
if dir != "" {
|
|
|
|
root += "/" + dir
|
|
|
|
}
|
2017-05-21 20:35:33 +00:00
|
|
|
|
|
|
|
started := false
|
|
|
|
var res *files.ListFolderResult
|
2014-07-12 10:46:45 +00:00
|
|
|
for {
|
2017-05-21 20:35:33 +00:00
|
|
|
if !started {
|
|
|
|
arg := files.ListFolderArg{
|
2020-01-14 17:33:35 +00:00
|
|
|
Path: f.opt.Enc.FromStandardPath(root),
|
2017-06-11 21:43:31 +00:00
|
|
|
Recursive: false,
|
2017-05-21 20:35:33 +00:00
|
|
|
}
|
|
|
|
if root == "/" {
|
|
|
|
arg.Path = "" // Specify root folder as empty string
|
|
|
|
}
|
2017-05-28 16:55:18 +00:00
|
|
|
err = f.pacer.Call(func() (bool, error) {
|
|
|
|
res, err = f.srv.ListFolder(&arg)
|
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
2017-05-21 20:35:33 +00:00
|
|
|
if err != nil {
|
|
|
|
switch e := err.(type) {
|
|
|
|
case files.ListFolderAPIError:
|
2019-10-10 15:34:09 +00:00
|
|
|
if e.EndpointError != nil && e.EndpointError.Path != nil && e.EndpointError.Path.Tag == files.LookupErrorNotFound {
|
2017-05-21 20:35:33 +00:00
|
|
|
err = fs.ErrorDirNotFound
|
|
|
|
}
|
2016-05-07 13:50:35 +00:00
|
|
|
}
|
2017-06-11 21:43:31 +00:00
|
|
|
return nil, err
|
2017-05-21 20:35:33 +00:00
|
|
|
}
|
2017-07-07 20:20:07 +00:00
|
|
|
started = true
|
2017-05-21 20:35:33 +00:00
|
|
|
} else {
|
|
|
|
arg := files.ListFolderContinueArg{
|
|
|
|
Cursor: res.Cursor,
|
|
|
|
}
|
2017-05-28 16:55:18 +00:00
|
|
|
err = f.pacer.Call(func() (bool, error) {
|
|
|
|
res, err = f.srv.ListFolderContinue(&arg)
|
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
2017-05-21 20:35:33 +00:00
|
|
|
if err != nil {
|
2017-06-11 21:43:31 +00:00
|
|
|
return nil, errors.Wrap(err, "list continue")
|
2017-05-21 20:35:33 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
for _, entry := range res.Entries {
|
|
|
|
var fileInfo *files.FileMetadata
|
|
|
|
var folderInfo *files.FolderMetadata
|
|
|
|
var metadata *files.Metadata
|
|
|
|
switch info := entry.(type) {
|
|
|
|
case *files.FolderMetadata:
|
|
|
|
folderInfo = info
|
|
|
|
metadata = &info.Metadata
|
|
|
|
case *files.FileMetadata:
|
|
|
|
fileInfo = info
|
|
|
|
metadata = &info.Metadata
|
|
|
|
default:
|
|
|
|
fs.Errorf(f, "Unknown type %T", entry)
|
|
|
|
continue
|
|
|
|
}
|
2015-05-23 18:56:48 +00:00
|
|
|
|
2017-08-10 11:06:50 +00:00
|
|
|
// Only the last element is reliably cased in PathDisplay
|
|
|
|
entryPath := metadata.PathDisplay
|
2020-01-14 17:33:35 +00:00
|
|
|
leaf := f.opt.Enc.ToStandardName(path.Base(entryPath))
|
2017-08-10 11:06:50 +00:00
|
|
|
remote := path.Join(dir, leaf)
|
2017-05-21 20:35:33 +00:00
|
|
|
if folderInfo != nil {
|
2020-08-26 23:12:10 +00:00
|
|
|
d := fs.NewDir(remote, time.Now()).SetID(folderInfo.Id)
|
2017-08-10 11:06:50 +00:00
|
|
|
entries = append(entries, d)
|
2017-05-21 20:35:33 +00:00
|
|
|
} else if fileInfo != nil {
|
2017-08-10 11:06:50 +00:00
|
|
|
o, err := f.newObjectWithInfo(remote, fileInfo)
|
2017-05-21 20:35:33 +00:00
|
|
|
if err != nil {
|
2017-06-11 21:43:31 +00:00
|
|
|
return nil, err
|
2017-05-21 20:35:33 +00:00
|
|
|
}
|
2017-06-11 21:43:31 +00:00
|
|
|
entries = append(entries, o)
|
2014-07-12 10:46:45 +00:00
|
|
|
}
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
2017-05-21 20:35:33 +00:00
|
|
|
if !res.HasMore {
|
2016-05-07 13:50:35 +00:00
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
2017-06-11 21:43:31 +00:00
|
|
|
return entries, nil
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Put the object
|
|
|
|
//
|
|
|
|
// Copy the reader in to the new object which is returned
|
|
|
|
//
|
|
|
|
// The new object may have been created if an error is returned
|
2019-06-17 08:34:30 +00:00
|
|
|
func (f *Fs) Put(ctx context.Context, in io.Reader, src fs.ObjectInfo, options ...fs.OpenOption) (fs.Object, error) {
|
2020-08-23 14:10:30 +00:00
|
|
|
if f.opt.SharedFiles || f.opt.SharedFolders {
|
2020-11-26 11:58:43 +00:00
|
|
|
return nil, errNotSupportedInSharedMode
|
2020-08-23 14:10:30 +00:00
|
|
|
}
|
2015-11-07 11:14:46 +00:00
|
|
|
// Temporary Object under construction
|
|
|
|
o := &Object{
|
|
|
|
fs: f,
|
2016-02-18 11:35:25 +00:00
|
|
|
remote: src.Remote(),
|
2015-11-07 11:14:46 +00:00
|
|
|
}
|
2019-06-17 08:34:30 +00:00
|
|
|
return o, o.Update(ctx, in, src, options...)
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
2017-08-03 19:42:35 +00:00
|
|
|
// PutStream uploads to the remote path with the modTime given of indeterminate size
|
2019-06-17 08:34:30 +00:00
|
|
|
func (f *Fs) PutStream(ctx context.Context, in io.Reader, src fs.ObjectInfo, options ...fs.OpenOption) (fs.Object, error) {
|
|
|
|
return f.Put(ctx, in, src, options...)
|
2017-08-03 19:42:35 +00:00
|
|
|
}
|
|
|
|
|
2014-07-08 20:59:30 +00:00
|
|
|
// Mkdir creates the container if it doesn't exist
|
2019-06-17 08:34:30 +00:00
|
|
|
func (f *Fs) Mkdir(ctx context.Context, dir string) error {
|
2020-08-23 14:10:30 +00:00
|
|
|
if f.opt.SharedFiles || f.opt.SharedFolders {
|
2020-11-26 11:58:43 +00:00
|
|
|
return errNotSupportedInSharedMode
|
2020-08-23 14:10:30 +00:00
|
|
|
}
|
2016-11-25 21:52:43 +00:00
|
|
|
root := path.Join(f.slashRoot, dir)
|
2017-05-21 20:35:33 +00:00
|
|
|
|
|
|
|
// can't create or run metadata on root
|
|
|
|
if root == "/" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// check directory doesn't exist
|
|
|
|
_, err := f.getDirMetadata(root)
|
2014-07-13 09:51:47 +00:00
|
|
|
if err == nil {
|
2017-05-21 20:35:33 +00:00
|
|
|
return nil // directory exists already
|
|
|
|
} else if err != fs.ErrorDirNotFound {
|
|
|
|
return err // some other error
|
|
|
|
}
|
|
|
|
|
|
|
|
// create it
|
|
|
|
arg2 := files.CreateFolderArg{
|
2020-01-14 17:33:35 +00:00
|
|
|
Path: f.opt.Enc.FromStandardPath(root),
|
2014-07-13 09:51:47 +00:00
|
|
|
}
|
2020-11-27 11:49:37 +00:00
|
|
|
// Don't attempt to create filenames that are too long
|
|
|
|
if cErr := checkPathLength(arg2.Path); cErr != nil {
|
|
|
|
return cErr
|
|
|
|
}
|
2017-05-28 16:55:18 +00:00
|
|
|
err = f.pacer.Call(func() (bool, error) {
|
2017-08-09 11:58:48 +00:00
|
|
|
_, err = f.srv.CreateFolderV2(&arg2)
|
2017-05-28 16:55:18 +00:00
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
2014-07-08 20:59:30 +00:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2020-06-04 21:25:14 +00:00
|
|
|
// purgeCheck removes the root directory, if check is set then it
|
|
|
|
// refuses to do so if it has anything in
|
|
|
|
func (f *Fs) purgeCheck(ctx context.Context, dir string, check bool) (err error) {
|
2016-11-25 21:52:43 +00:00
|
|
|
root := path.Join(f.slashRoot, dir)
|
2017-05-21 20:35:33 +00:00
|
|
|
|
|
|
|
// can't remove root
|
|
|
|
if root == "/" {
|
|
|
|
return errors.New("can't remove root directory")
|
|
|
|
}
|
|
|
|
|
2020-06-04 21:25:14 +00:00
|
|
|
if check {
|
|
|
|
// check directory exists
|
|
|
|
_, err = f.getDirMetadata(root)
|
|
|
|
if err != nil {
|
|
|
|
return errors.Wrap(err, "Rmdir")
|
|
|
|
}
|
2017-05-21 20:35:33 +00:00
|
|
|
|
2020-06-04 21:25:14 +00:00
|
|
|
root = f.opt.Enc.FromStandardPath(root)
|
|
|
|
// check directory empty
|
|
|
|
arg := files.ListFolderArg{
|
|
|
|
Path: root,
|
|
|
|
Recursive: false,
|
|
|
|
}
|
|
|
|
if root == "/" {
|
|
|
|
arg.Path = "" // Specify root folder as empty string
|
|
|
|
}
|
|
|
|
var res *files.ListFolderResult
|
|
|
|
err = f.pacer.Call(func() (bool, error) {
|
|
|
|
res, err = f.srv.ListFolder(&arg)
|
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return errors.Wrap(err, "Rmdir")
|
|
|
|
}
|
|
|
|
if len(res.Entries) != 0 {
|
|
|
|
return errors.New("directory not empty")
|
|
|
|
}
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
2017-05-21 20:35:33 +00:00
|
|
|
|
|
|
|
// remove it
|
2017-05-28 16:55:18 +00:00
|
|
|
err = f.pacer.Call(func() (bool, error) {
|
2017-08-09 11:58:48 +00:00
|
|
|
_, err = f.srv.DeleteV2(&files.DeleteArg{Path: root})
|
2017-05-28 16:55:18 +00:00
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
2016-11-25 21:52:43 +00:00
|
|
|
return err
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
2020-06-04 21:25:14 +00:00
|
|
|
// Rmdir deletes the container
|
|
|
|
//
|
|
|
|
// Returns an error if it isn't empty
|
|
|
|
func (f *Fs) Rmdir(ctx context.Context, dir string) error {
|
2020-08-23 14:10:30 +00:00
|
|
|
if f.opt.SharedFiles || f.opt.SharedFolders {
|
2020-11-26 11:58:43 +00:00
|
|
|
return errNotSupportedInSharedMode
|
2020-08-23 14:10:30 +00:00
|
|
|
}
|
2020-06-04 21:25:14 +00:00
|
|
|
return f.purgeCheck(ctx, dir, true)
|
|
|
|
}
|
|
|
|
|
2015-09-22 17:47:16 +00:00
|
|
|
// Precision returns the precision
|
2015-11-07 11:14:46 +00:00
|
|
|
func (f *Fs) Precision() time.Duration {
|
2017-05-21 20:35:33 +00:00
|
|
|
return time.Second
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
2020-10-13 21:43:40 +00:00
|
|
|
// Copy src to this remote using server-side copy operations.
|
2015-02-14 18:48:08 +00:00
|
|
|
//
|
|
|
|
// This is stored with the remote path given
|
|
|
|
//
|
|
|
|
// It returns the destination Object and a possible error
|
|
|
|
//
|
|
|
|
// Will only be called if src.Fs().Name() == f.Name()
|
|
|
|
//
|
|
|
|
// If it isn't possible then return fs.ErrorCantCopy
|
2019-06-17 08:34:30 +00:00
|
|
|
func (f *Fs) Copy(ctx context.Context, src fs.Object, remote string) (fs.Object, error) {
|
2015-11-07 11:14:46 +00:00
|
|
|
srcObj, ok := src.(*Object)
|
2015-02-14 18:48:08 +00:00
|
|
|
if !ok {
|
2017-02-09 11:01:20 +00:00
|
|
|
fs.Debugf(src, "Can't copy - not same remote type")
|
2015-02-14 18:48:08 +00:00
|
|
|
return nil, fs.ErrorCantCopy
|
|
|
|
}
|
|
|
|
|
2015-11-07 11:14:46 +00:00
|
|
|
// Temporary Object under construction
|
|
|
|
dstObj := &Object{
|
|
|
|
fs: f,
|
|
|
|
remote: remote,
|
|
|
|
}
|
2015-02-14 18:48:08 +00:00
|
|
|
|
2017-05-21 20:35:33 +00:00
|
|
|
// Copy
|
2018-11-02 12:14:02 +00:00
|
|
|
arg := files.RelocationArg{
|
|
|
|
RelocationPath: files.RelocationPath{
|
2020-01-14 17:33:35 +00:00
|
|
|
FromPath: f.opt.Enc.FromStandardPath(srcObj.remotePath()),
|
|
|
|
ToPath: f.opt.Enc.FromStandardPath(dstObj.remotePath()),
|
2018-11-02 12:14:02 +00:00
|
|
|
},
|
|
|
|
}
|
2017-05-28 16:55:18 +00:00
|
|
|
var err error
|
2017-09-24 19:01:30 +00:00
|
|
|
var result *files.RelocationResult
|
2017-05-28 16:55:18 +00:00
|
|
|
err = f.pacer.Call(func() (bool, error) {
|
2017-09-24 19:01:30 +00:00
|
|
|
result, err = f.srv.CopyV2(&arg)
|
2017-05-28 16:55:18 +00:00
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
2015-02-14 18:48:08 +00:00
|
|
|
if err != nil {
|
2016-06-12 14:06:02 +00:00
|
|
|
return nil, errors.Wrap(err, "copy failed")
|
2015-02-14 18:48:08 +00:00
|
|
|
}
|
2017-05-21 20:35:33 +00:00
|
|
|
|
|
|
|
// Set the metadata
|
2017-09-24 19:01:30 +00:00
|
|
|
fileInfo, ok := result.Metadata.(*files.FileMetadata)
|
2017-05-21 20:35:33 +00:00
|
|
|
if !ok {
|
|
|
|
return nil, fs.ErrorNotAFile
|
|
|
|
}
|
|
|
|
err = dstObj.setMetadataFromEntry(fileInfo)
|
2017-02-25 11:09:57 +00:00
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "copy failed")
|
|
|
|
}
|
2017-05-21 20:35:33 +00:00
|
|
|
|
2015-02-14 18:48:08 +00:00
|
|
|
return dstObj, nil
|
|
|
|
}
|
|
|
|
|
2014-07-08 20:59:30 +00:00
|
|
|
// Purge deletes all the files and the container
|
|
|
|
//
|
2014-07-13 09:53:53 +00:00
|
|
|
// Optional interface: Only implement this if you have a way of
|
|
|
|
// deleting all the files quicker than just running Remove() on the
|
|
|
|
// result of List()
|
2020-06-04 21:25:14 +00:00
|
|
|
func (f *Fs) Purge(ctx context.Context, dir string) (err error) {
|
|
|
|
return f.purgeCheck(ctx, dir, false)
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
2020-10-13 21:43:40 +00:00
|
|
|
// Move src to this remote using server-side move operations.
|
2015-08-31 20:05:51 +00:00
|
|
|
//
|
|
|
|
// This is stored with the remote path given
|
|
|
|
//
|
|
|
|
// It returns the destination Object and a possible error
|
|
|
|
//
|
|
|
|
// Will only be called if src.Fs().Name() == f.Name()
|
|
|
|
//
|
|
|
|
// If it isn't possible then return fs.ErrorCantMove
|
2019-06-17 08:34:30 +00:00
|
|
|
func (f *Fs) Move(ctx context.Context, src fs.Object, remote string) (fs.Object, error) {
|
2015-11-07 11:14:46 +00:00
|
|
|
srcObj, ok := src.(*Object)
|
2015-08-31 20:05:51 +00:00
|
|
|
if !ok {
|
2017-02-09 11:01:20 +00:00
|
|
|
fs.Debugf(src, "Can't move - not same remote type")
|
2015-08-31 20:05:51 +00:00
|
|
|
return nil, fs.ErrorCantMove
|
|
|
|
}
|
|
|
|
|
2015-11-07 11:14:46 +00:00
|
|
|
// Temporary Object under construction
|
|
|
|
dstObj := &Object{
|
|
|
|
fs: f,
|
|
|
|
remote: remote,
|
|
|
|
}
|
2015-08-31 20:05:51 +00:00
|
|
|
|
2017-05-21 20:35:33 +00:00
|
|
|
// Do the move
|
2018-11-02 12:14:02 +00:00
|
|
|
arg := files.RelocationArg{
|
|
|
|
RelocationPath: files.RelocationPath{
|
2020-01-14 17:33:35 +00:00
|
|
|
FromPath: f.opt.Enc.FromStandardPath(srcObj.remotePath()),
|
|
|
|
ToPath: f.opt.Enc.FromStandardPath(dstObj.remotePath()),
|
2018-11-02 12:14:02 +00:00
|
|
|
},
|
|
|
|
}
|
2017-05-28 16:55:18 +00:00
|
|
|
var err error
|
2017-09-24 19:01:30 +00:00
|
|
|
var result *files.RelocationResult
|
2017-05-28 16:55:18 +00:00
|
|
|
err = f.pacer.Call(func() (bool, error) {
|
2017-09-24 19:01:30 +00:00
|
|
|
result, err = f.srv.MoveV2(&arg)
|
2017-05-28 16:55:18 +00:00
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
2015-08-31 20:05:51 +00:00
|
|
|
if err != nil {
|
2016-06-12 14:06:02 +00:00
|
|
|
return nil, errors.Wrap(err, "move failed")
|
2015-08-31 20:05:51 +00:00
|
|
|
}
|
2017-05-21 20:35:33 +00:00
|
|
|
|
|
|
|
// Set the metadata
|
2017-09-24 19:01:30 +00:00
|
|
|
fileInfo, ok := result.Metadata.(*files.FileMetadata)
|
2017-05-21 20:35:33 +00:00
|
|
|
if !ok {
|
|
|
|
return nil, fs.ErrorNotAFile
|
|
|
|
}
|
|
|
|
err = dstObj.setMetadataFromEntry(fileInfo)
|
2017-02-25 11:09:57 +00:00
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "move failed")
|
|
|
|
}
|
2015-08-31 20:05:51 +00:00
|
|
|
return dstObj, nil
|
|
|
|
}
|
|
|
|
|
2018-03-29 07:10:19 +00:00
|
|
|
// PublicLink adds a "readable by anyone with link" permission on the given file or folder.
|
2020-05-31 21:18:01 +00:00
|
|
|
func (f *Fs) PublicLink(ctx context.Context, remote string, expire fs.Duration, unlink bool) (link string, err error) {
|
2020-01-14 17:33:35 +00:00
|
|
|
absPath := f.opt.Enc.FromStandardPath(path.Join(f.slashRoot, remote))
|
2018-03-29 07:10:19 +00:00
|
|
|
fs.Debugf(f, "attempting to share '%s' (absolute path: %s)", remote, absPath)
|
|
|
|
createArg := sharing.CreateSharedLinkWithSettingsArg{
|
|
|
|
Path: absPath,
|
2020-06-18 17:40:33 +00:00
|
|
|
// FIXME this gives settings_error/not_authorized/.. errors
|
|
|
|
// and the expires setting isn't in the documentation so remove
|
|
|
|
// for now.
|
|
|
|
// Settings: &sharing.SharedLinkSettings{
|
|
|
|
// Expires: time.Now().Add(time.Duration(expire)).UTC().Round(time.Second),
|
|
|
|
// },
|
2018-03-29 07:10:19 +00:00
|
|
|
}
|
|
|
|
var linkRes sharing.IsSharedLinkMetadata
|
|
|
|
err = f.pacer.Call(func() (bool, error) {
|
2018-05-28 09:50:31 +00:00
|
|
|
linkRes, err = f.sharing.CreateSharedLinkWithSettings(&createArg)
|
2018-03-29 07:10:19 +00:00
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
|
|
|
|
2018-11-02 12:14:02 +00:00
|
|
|
if err != nil && strings.Contains(err.Error(),
|
|
|
|
sharing.CreateSharedLinkWithSettingsErrorSharedLinkAlreadyExists) {
|
2018-03-29 07:10:19 +00:00
|
|
|
fs.Debugf(absPath, "has a public link already, attempting to retrieve it")
|
|
|
|
listArg := sharing.ListSharedLinksArg{
|
|
|
|
Path: absPath,
|
|
|
|
DirectOnly: true,
|
|
|
|
}
|
|
|
|
var listRes *sharing.ListSharedLinksResult
|
|
|
|
err = f.pacer.Call(func() (bool, error) {
|
2018-05-28 09:50:31 +00:00
|
|
|
listRes, err = f.sharing.ListSharedLinks(&listArg)
|
2018-03-29 07:10:19 +00:00
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
if len(listRes.Links) == 0 {
|
|
|
|
err = errors.New("Dropbox says the sharing link already exists, but list came back empty")
|
|
|
|
return
|
|
|
|
}
|
|
|
|
linkRes = listRes.Links[0]
|
|
|
|
}
|
|
|
|
if err == nil {
|
|
|
|
switch res := linkRes.(type) {
|
|
|
|
case *sharing.FileLinkMetadata:
|
|
|
|
link = res.Url
|
|
|
|
case *sharing.FolderLinkMetadata:
|
|
|
|
link = res.Url
|
|
|
|
default:
|
|
|
|
err = fmt.Errorf("Don't know how to extract link, response has unknown format: %T", res)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2017-02-05 21:20:56 +00:00
|
|
|
// DirMove moves src, srcRemote to this remote at dstRemote
|
2020-10-13 21:43:40 +00:00
|
|
|
// using server-side move operations.
|
2015-08-31 20:05:51 +00:00
|
|
|
//
|
|
|
|
// Will only be called if src.Fs().Name() == f.Name()
|
|
|
|
//
|
|
|
|
// If it isn't possible then return fs.ErrorCantDirMove
|
|
|
|
//
|
|
|
|
// If destination exists then return fs.ErrorDirExists
|
2019-06-17 08:34:30 +00:00
|
|
|
func (f *Fs) DirMove(ctx context.Context, src fs.Fs, srcRemote, dstRemote string) error {
|
2015-11-07 11:14:46 +00:00
|
|
|
srcFs, ok := src.(*Fs)
|
2015-08-31 20:05:51 +00:00
|
|
|
if !ok {
|
2017-02-09 11:01:20 +00:00
|
|
|
fs.Debugf(srcFs, "Can't move directory - not same remote type")
|
2015-08-31 20:05:51 +00:00
|
|
|
return fs.ErrorCantDirMove
|
|
|
|
}
|
2017-02-05 21:20:56 +00:00
|
|
|
srcPath := path.Join(srcFs.slashRoot, srcRemote)
|
|
|
|
dstPath := path.Join(f.slashRoot, dstRemote)
|
2015-08-31 20:05:51 +00:00
|
|
|
|
|
|
|
// Check if destination exists
|
2017-10-02 10:21:16 +00:00
|
|
|
_, err := f.getDirMetadata(dstPath)
|
2017-05-21 20:35:33 +00:00
|
|
|
if err == nil {
|
2015-08-31 20:05:51 +00:00
|
|
|
return fs.ErrorDirExists
|
2017-05-21 20:35:33 +00:00
|
|
|
} else if err != fs.ErrorDirNotFound {
|
|
|
|
return err
|
2015-08-31 20:05:51 +00:00
|
|
|
}
|
|
|
|
|
2017-02-05 21:20:56 +00:00
|
|
|
// Make sure the parent directory exists
|
|
|
|
// ...apparently not necessary
|
|
|
|
|
2015-08-31 20:05:51 +00:00
|
|
|
// Do the move
|
2018-11-02 12:14:02 +00:00
|
|
|
arg := files.RelocationArg{
|
|
|
|
RelocationPath: files.RelocationPath{
|
2020-01-14 17:33:35 +00:00
|
|
|
FromPath: f.opt.Enc.FromStandardPath(srcPath),
|
|
|
|
ToPath: f.opt.Enc.FromStandardPath(dstPath),
|
2018-11-02 12:14:02 +00:00
|
|
|
},
|
|
|
|
}
|
2017-05-28 16:55:18 +00:00
|
|
|
err = f.pacer.Call(func() (bool, error) {
|
2017-09-24 19:01:30 +00:00
|
|
|
_, err = f.srv.MoveV2(&arg)
|
2017-05-28 16:55:18 +00:00
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
2015-08-31 20:05:51 +00:00
|
|
|
if err != nil {
|
2016-06-12 14:06:02 +00:00
|
|
|
return errors.Wrap(err, "MoveDir failed")
|
2015-08-31 20:05:51 +00:00
|
|
|
}
|
2017-05-21 20:35:33 +00:00
|
|
|
|
2015-08-31 20:05:51 +00:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2018-04-16 21:19:25 +00:00
|
|
|
// About gets quota information
|
2019-06-17 08:34:30 +00:00
|
|
|
func (f *Fs) About(ctx context.Context) (usage *fs.Usage, err error) {
|
2018-04-16 21:19:25 +00:00
|
|
|
var q *users.SpaceUsage
|
|
|
|
err = f.pacer.Call(func() (bool, error) {
|
|
|
|
q, err = f.users.GetSpaceUsage()
|
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "about failed")
|
|
|
|
}
|
|
|
|
var total uint64
|
|
|
|
if q.Allocation != nil {
|
|
|
|
if q.Allocation.Individual != nil {
|
|
|
|
total += q.Allocation.Individual.Allocated
|
|
|
|
}
|
|
|
|
if q.Allocation.Team != nil {
|
|
|
|
total += q.Allocation.Team.Allocated
|
|
|
|
}
|
|
|
|
}
|
|
|
|
usage = &fs.Usage{
|
|
|
|
Total: fs.NewUsageValue(int64(total)), // quota of bytes that can be used
|
|
|
|
Used: fs.NewUsageValue(int64(q.Used)), // bytes in use
|
|
|
|
Free: fs.NewUsageValue(int64(total - q.Used)), // bytes which can be uploaded before reaching the quota
|
|
|
|
}
|
|
|
|
return usage, nil
|
|
|
|
}
|
|
|
|
|
2016-01-11 12:39:33 +00:00
|
|
|
// Hashes returns the supported hash sets.
|
2018-01-12 16:30:54 +00:00
|
|
|
func (f *Fs) Hashes() hash.Set {
|
2019-09-23 13:32:36 +00:00
|
|
|
return hash.Set(DbHashType)
|
2016-01-11 12:39:33 +00:00
|
|
|
}
|
|
|
|
|
2014-07-08 20:59:30 +00:00
|
|
|
// ------------------------------------------------------------
|
|
|
|
|
2015-09-22 17:47:16 +00:00
|
|
|
// Fs returns the parent Fs
|
2016-02-18 11:35:25 +00:00
|
|
|
func (o *Object) Fs() fs.Info {
|
2015-11-07 11:14:46 +00:00
|
|
|
return o.fs
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Return a string version
|
2015-11-07 11:14:46 +00:00
|
|
|
func (o *Object) String() string {
|
2014-07-08 20:59:30 +00:00
|
|
|
if o == nil {
|
|
|
|
return "<nil>"
|
|
|
|
}
|
|
|
|
return o.remote
|
|
|
|
}
|
|
|
|
|
2015-09-22 17:47:16 +00:00
|
|
|
// Remote returns the remote path
|
2015-11-07 11:14:46 +00:00
|
|
|
func (o *Object) Remote() string {
|
2014-07-08 20:59:30 +00:00
|
|
|
return o.remote
|
|
|
|
}
|
|
|
|
|
2020-08-26 23:12:10 +00:00
|
|
|
// ID returns the object id
|
|
|
|
func (o *Object) ID() string {
|
|
|
|
return o.id
|
|
|
|
}
|
|
|
|
|
2017-05-26 14:09:31 +00:00
|
|
|
// Hash returns the dropbox special hash
|
2019-06-17 08:34:30 +00:00
|
|
|
func (o *Object) Hash(ctx context.Context, t hash.Type) (string, error) {
|
2020-08-23 14:10:30 +00:00
|
|
|
if o.fs.opt.SharedFiles || o.fs.opt.SharedFolders {
|
2020-11-26 11:58:43 +00:00
|
|
|
return "", errNotSupportedInSharedMode
|
2020-08-23 14:10:30 +00:00
|
|
|
}
|
2019-09-23 13:32:36 +00:00
|
|
|
if t != DbHashType {
|
2018-01-18 20:27:52 +00:00
|
|
|
return "", hash.ErrUnsupported
|
2017-05-26 14:09:31 +00:00
|
|
|
}
|
|
|
|
err := o.readMetaData()
|
|
|
|
if err != nil {
|
|
|
|
return "", errors.Wrap(err, "failed to read hash from metadata")
|
|
|
|
}
|
|
|
|
return o.hash, nil
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Size returns the size of an object in bytes
|
2015-11-07 11:14:46 +00:00
|
|
|
func (o *Object) Size() int64 {
|
2014-07-08 20:59:30 +00:00
|
|
|
return o.bytes
|
|
|
|
}
|
|
|
|
|
2017-05-21 20:35:33 +00:00
|
|
|
// setMetadataFromEntry sets the fs data from a files.FileMetadata
|
2014-07-09 23:17:40 +00:00
|
|
|
//
|
Spelling fixes
Fix spelling of: above, already, anonymous, associated,
authentication, bandwidth, because, between, blocks, calculate,
candidates, cautious, changelog, cleaner, clipboard, command,
completely, concurrently, considered, constructs, corrupt, current,
daemon, dependencies, deprecated, directory, dispatcher, download,
eligible, ellipsis, encrypter, endpoint, entrieslist, essentially,
existing writers, existing, expires, filesystem, flushing, frequently,
hierarchy, however, implementation, implements, inaccurate,
individually, insensitive, longer, maximum, metadata, modified,
multipart, namedirfirst, nextcloud, obscured, opened, optional,
owncloud, pacific, passphrase, password, permanently, persimmon,
positive, potato, protocol, quota, receiving, recommends, referring,
requires, revisited, satisfied, satisfies, satisfy, semver,
serialized, session, storage, strategies, stringlist, successful,
supported, surprise, temporarily, temporary, transactions, unneeded,
update, uploads, wrapped
Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>
2020-10-09 00:17:24 +00:00
|
|
|
// This isn't a complete set of metadata and has an inaccurate date
|
2017-05-21 20:35:33 +00:00
|
|
|
func (o *Object) setMetadataFromEntry(info *files.FileMetadata) error {
|
2020-08-23 14:10:30 +00:00
|
|
|
o.id = info.Id
|
2017-05-21 20:35:33 +00:00
|
|
|
o.bytes = int64(info.Size)
|
|
|
|
o.modTime = info.ClientModified
|
2017-05-26 14:09:31 +00:00
|
|
|
o.hash = info.ContentHash
|
2017-02-25 11:09:57 +00:00
|
|
|
return nil
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
2017-05-21 20:35:33 +00:00
|
|
|
// Reads the entry for a file from dropbox
|
|
|
|
func (o *Object) readEntry() (*files.FileMetadata, error) {
|
|
|
|
return o.fs.getFileMetadata(o.remotePath())
|
2014-07-09 23:17:40 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Read entry if not set and set metadata from it
|
2015-11-07 11:14:46 +00:00
|
|
|
func (o *Object) readEntryAndSetMetadata() error {
|
2014-07-09 23:17:40 +00:00
|
|
|
// Last resort set time from client
|
|
|
|
if !o.modTime.IsZero() {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
entry, err := o.readEntry()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2017-02-25 11:09:57 +00:00
|
|
|
return o.setMetadataFromEntry(entry)
|
2014-07-09 23:17:40 +00:00
|
|
|
}
|
|
|
|
|
2014-07-08 20:59:30 +00:00
|
|
|
// Returns the remote path for the object
|
2015-11-07 11:14:46 +00:00
|
|
|
func (o *Object) remotePath() string {
|
|
|
|
return o.fs.slashRootSlash + o.remote
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// readMetaData gets the info if it hasn't already been fetched
|
2015-11-07 11:14:46 +00:00
|
|
|
func (o *Object) readMetaData() (err error) {
|
2017-05-26 14:09:31 +00:00
|
|
|
if !o.modTime.IsZero() {
|
2014-07-08 20:59:30 +00:00
|
|
|
return nil
|
|
|
|
}
|
2014-07-09 23:17:40 +00:00
|
|
|
// Last resort
|
2014-07-25 17:19:49 +00:00
|
|
|
return o.readEntryAndSetMetadata()
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// ModTime returns the modification time of the object
|
|
|
|
//
|
|
|
|
// It attempts to read the objects mtime and if that isn't present the
|
|
|
|
// LastModified returned in the http headers
|
2019-06-17 08:34:30 +00:00
|
|
|
func (o *Object) ModTime(ctx context.Context) time.Time {
|
2014-07-08 20:59:30 +00:00
|
|
|
err := o.readMetaData()
|
|
|
|
if err != nil {
|
2017-02-09 17:08:51 +00:00
|
|
|
fs.Debugf(o, "Failed to read metadata: %v", err)
|
2014-07-08 20:59:30 +00:00
|
|
|
return time.Now()
|
|
|
|
}
|
|
|
|
return o.modTime
|
|
|
|
}
|
|
|
|
|
2015-09-22 17:47:16 +00:00
|
|
|
// SetModTime sets the modification time of the local fs object
|
2014-07-09 23:17:40 +00:00
|
|
|
//
|
|
|
|
// Commits the datastore
|
2019-06-17 08:34:30 +00:00
|
|
|
func (o *Object) SetModTime(ctx context.Context, modTime time.Time) error {
|
2017-05-21 20:35:33 +00:00
|
|
|
// Dropbox doesn't have a way of doing this so returning this
|
2017-06-13 12:58:39 +00:00
|
|
|
// error will cause the file to be deleted first then
|
|
|
|
// re-uploaded to set the time.
|
2018-04-06 14:34:56 +00:00
|
|
|
return fs.ErrorCantSetModTimeWithoutDelete
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
2015-09-22 17:47:16 +00:00
|
|
|
// Storable returns whether this object is storable
|
2015-11-07 11:14:46 +00:00
|
|
|
func (o *Object) Storable() bool {
|
2014-07-08 20:59:30 +00:00
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// Open an object for read
|
2019-06-17 08:34:30 +00:00
|
|
|
func (o *Object) Open(ctx context.Context, options ...fs.OpenOption) (in io.ReadCloser, err error) {
|
2020-08-23 14:10:30 +00:00
|
|
|
if o.fs.opt.SharedFiles {
|
|
|
|
if len(options) != 0 {
|
|
|
|
return nil, errors.New("OpenOptions not supported for shared files")
|
|
|
|
}
|
|
|
|
arg := sharing.GetSharedLinkMetadataArg{
|
|
|
|
Url: o.url,
|
|
|
|
}
|
|
|
|
err = o.fs.pacer.Call(func() (bool, error) {
|
|
|
|
_, in, err = o.fs.sharing.GetSharedLinkFile(&arg)
|
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2019-08-06 14:18:08 +00:00
|
|
|
fs.FixRangeOption(options, o.bytes)
|
2017-05-21 20:35:33 +00:00
|
|
|
headers := fs.OpenOptionHeaders(options)
|
2018-11-02 12:14:02 +00:00
|
|
|
arg := files.DownloadArg{
|
2020-08-23 14:10:30 +00:00
|
|
|
Path: o.id,
|
2018-11-02 12:14:02 +00:00
|
|
|
ExtraHeaders: headers,
|
|
|
|
}
|
2017-05-28 16:55:18 +00:00
|
|
|
err = o.fs.pacer.Call(func() (bool, error) {
|
|
|
|
_, in, err = o.fs.srv.Download(&arg)
|
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
2017-05-21 20:35:33 +00:00
|
|
|
|
|
|
|
switch e := err.(type) {
|
|
|
|
case files.DownloadAPIError:
|
|
|
|
// Don't attempt to retry copyright violation errors
|
2019-10-10 15:34:09 +00:00
|
|
|
if e.EndpointError != nil && e.EndpointError.Path != nil && e.EndpointError.Path.Tag == files.LookupErrorRestrictedContent {
|
2018-01-12 16:30:54 +00:00
|
|
|
return nil, fserrors.NoRetryError(err)
|
2016-09-10 10:29:57 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-05-21 20:35:33 +00:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// uploadChunked uploads the object in parts
|
|
|
|
//
|
2017-08-03 19:42:35 +00:00
|
|
|
// Will work optimally if size is >= uploadChunkSize. If the size is either
|
|
|
|
// unknown (i.e. -1) or smaller than uploadChunkSize, the method incurs an
|
|
|
|
// avoidable request to the Dropbox API that does not carry payload.
|
|
|
|
func (o *Object) uploadChunked(in0 io.Reader, commitInfo *files.CommitInfo, size int64) (entry *files.FileMetadata, err error) {
|
2018-05-14 17:06:57 +00:00
|
|
|
chunkSize := int64(o.fs.opt.ChunkSize)
|
2017-08-03 19:42:35 +00:00
|
|
|
chunks := 0
|
|
|
|
if size != -1 {
|
|
|
|
chunks = int(size/chunkSize) + 1
|
|
|
|
}
|
2018-01-12 16:30:54 +00:00
|
|
|
in := readers.NewCountingReader(in0)
|
2017-11-23 10:59:28 +00:00
|
|
|
buf := make([]byte, int(chunkSize))
|
2017-08-03 19:42:35 +00:00
|
|
|
|
|
|
|
fmtChunk := func(cur int, last bool) {
|
|
|
|
if chunks == 0 && last {
|
|
|
|
fs.Debugf(o, "Streaming chunk %d/%d", cur, cur)
|
|
|
|
} else if chunks == 0 {
|
|
|
|
fs.Debugf(o, "Streaming chunk %d/unknown", cur)
|
|
|
|
} else {
|
|
|
|
fs.Debugf(o, "Uploading chunk %d/%d", cur, chunks)
|
|
|
|
}
|
|
|
|
}
|
2017-05-21 20:35:33 +00:00
|
|
|
|
2017-08-03 19:42:35 +00:00
|
|
|
// write the first chunk
|
|
|
|
fmtChunk(1, false)
|
2017-05-28 16:55:18 +00:00
|
|
|
var res *files.UploadSessionStartResult
|
2018-01-12 16:30:54 +00:00
|
|
|
chunk := readers.NewRepeatableLimitReaderBuffer(in, buf, chunkSize)
|
2017-11-08 09:18:16 +00:00
|
|
|
err = o.fs.pacer.Call(func() (bool, error) {
|
|
|
|
// seek to the start in case this is a retry
|
2018-04-06 18:53:06 +00:00
|
|
|
if _, err = chunk.Seek(0, io.SeekStart); err != nil {
|
2017-11-08 09:18:16 +00:00
|
|
|
return false, nil
|
|
|
|
}
|
|
|
|
res, err = o.fs.srv.UploadSessionStart(&files.UploadSessionStartArg{}, chunk)
|
2017-05-28 16:55:18 +00:00
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
2017-05-21 20:35:33 +00:00
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
cursor := files.UploadSessionCursor{
|
|
|
|
SessionId: res.SessionId,
|
2017-08-03 19:42:35 +00:00
|
|
|
Offset: 0,
|
2017-05-21 20:35:33 +00:00
|
|
|
}
|
|
|
|
appendArg := files.UploadSessionAppendArg{
|
|
|
|
Cursor: &cursor,
|
|
|
|
Close: false,
|
|
|
|
}
|
|
|
|
|
|
|
|
// write more whole chunks (if any)
|
2017-08-03 19:42:35 +00:00
|
|
|
currentChunk := 2
|
|
|
|
for {
|
|
|
|
if chunks > 0 && currentChunk >= chunks {
|
|
|
|
// if the size is known, only upload full chunks. Remaining bytes are uploaded with
|
|
|
|
// the UploadSessionFinish request.
|
|
|
|
break
|
2017-08-19 12:07:23 +00:00
|
|
|
} else if chunks == 0 && in.BytesRead()-cursor.Offset < uint64(chunkSize) {
|
2017-08-03 19:42:35 +00:00
|
|
|
// if the size is unknown, upload as long as we can read full chunks from the reader.
|
|
|
|
// The UploadSessionFinish request will not contain any payload.
|
|
|
|
break
|
|
|
|
}
|
2017-08-19 12:07:23 +00:00
|
|
|
cursor.Offset = in.BytesRead()
|
2017-08-03 19:42:35 +00:00
|
|
|
fmtChunk(currentChunk, false)
|
2018-01-12 16:30:54 +00:00
|
|
|
chunk = readers.NewRepeatableLimitReaderBuffer(in, buf, chunkSize)
|
2017-11-08 09:18:16 +00:00
|
|
|
err = o.fs.pacer.Call(func() (bool, error) {
|
|
|
|
// seek to the start in case this is a retry
|
2018-04-06 18:53:06 +00:00
|
|
|
if _, err = chunk.Seek(0, io.SeekStart); err != nil {
|
2017-11-08 09:18:16 +00:00
|
|
|
return false, nil
|
|
|
|
}
|
|
|
|
err = o.fs.srv.UploadSessionAppendV2(&appendArg, chunk)
|
2017-11-21 10:03:44 +00:00
|
|
|
// after the first chunk is uploaded, we retry everything
|
|
|
|
return err != nil, err
|
2017-05-28 16:55:18 +00:00
|
|
|
})
|
2017-05-21 20:35:33 +00:00
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
2016-07-04 12:45:10 +00:00
|
|
|
}
|
2017-08-03 19:42:35 +00:00
|
|
|
currentChunk++
|
2016-07-04 12:45:10 +00:00
|
|
|
}
|
2017-05-21 20:35:33 +00:00
|
|
|
|
|
|
|
// write the remains
|
2017-08-19 12:07:23 +00:00
|
|
|
cursor.Offset = in.BytesRead()
|
2017-05-21 20:35:33 +00:00
|
|
|
args := &files.UploadSessionFinishArg{
|
|
|
|
Cursor: &cursor,
|
|
|
|
Commit: commitInfo,
|
|
|
|
}
|
2017-08-03 19:42:35 +00:00
|
|
|
fmtChunk(currentChunk, true)
|
2018-01-12 16:30:54 +00:00
|
|
|
chunk = readers.NewRepeatableReaderBuffer(in, buf)
|
2017-11-08 09:18:16 +00:00
|
|
|
err = o.fs.pacer.Call(func() (bool, error) {
|
|
|
|
// seek to the start in case this is a retry
|
2018-04-06 18:53:06 +00:00
|
|
|
if _, err = chunk.Seek(0, io.SeekStart); err != nil {
|
2017-11-08 09:18:16 +00:00
|
|
|
return false, nil
|
|
|
|
}
|
|
|
|
entry, err = o.fs.srv.UploadSessionFinish(args, chunk)
|
2020-01-08 12:57:21 +00:00
|
|
|
// If error is insufficient space then don't retry
|
|
|
|
if e, ok := err.(files.UploadSessionFinishAPIError); ok {
|
|
|
|
if e.EndpointError != nil && e.EndpointError.Path != nil && e.EndpointError.Path.Tag == files.WriteErrorInsufficientSpace {
|
|
|
|
err = fserrors.NoRetryError(err)
|
|
|
|
return false, err
|
|
|
|
}
|
|
|
|
}
|
2017-11-21 10:03:44 +00:00
|
|
|
// after the first chunk is uploaded, we retry everything
|
|
|
|
return err != nil, err
|
2017-05-28 16:55:18 +00:00
|
|
|
})
|
2017-05-21 20:35:33 +00:00
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
return entry, nil
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
2020-11-27 11:49:37 +00:00
|
|
|
// checks all the parts of name to see they are below
|
|
|
|
// maxFileNameLength runes.
|
|
|
|
//
|
|
|
|
// This checks the length as runes which isn't quite right as dropbox
|
|
|
|
// seems to encode some symbols (eg ☺) as two "characters". This seems
|
|
|
|
// like utf-16 except that ☺ doesn't need two characters in utf-16.
|
|
|
|
//
|
|
|
|
// Using runes instead of what dropbox is using will work for most
|
|
|
|
// cases, and when it goes wrong we will upload something we should
|
|
|
|
// have detected as too long which is the least damaging way to fail.
|
|
|
|
func checkPathLength(name string) (err error) {
|
|
|
|
for next := ""; len(name) > 0; name = next {
|
|
|
|
if slash := strings.IndexRune(name, '/'); slash >= 0 {
|
|
|
|
name, next = name[:slash], name[slash+1:]
|
|
|
|
} else {
|
|
|
|
next = ""
|
|
|
|
}
|
|
|
|
length := utf8.RuneCountInString(name)
|
|
|
|
if length > maxFileNameLength {
|
|
|
|
return fserrors.NoRetryError(fs.ErrorFileNameTooLong)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2014-07-08 20:59:30 +00:00
|
|
|
// Update the already existing object
|
|
|
|
//
|
|
|
|
// Copy the reader into the object updating modTime and size
|
|
|
|
//
|
|
|
|
// The new object may have been created if an error is returned
|
2019-06-17 08:34:30 +00:00
|
|
|
func (o *Object) Update(ctx context.Context, in io.Reader, src fs.ObjectInfo, options ...fs.OpenOption) error {
|
2020-08-23 14:10:30 +00:00
|
|
|
if o.fs.opt.SharedFiles || o.fs.opt.SharedFolders {
|
2020-11-26 11:58:43 +00:00
|
|
|
return errNotSupportedInSharedMode
|
2020-08-23 14:10:30 +00:00
|
|
|
}
|
2015-08-20 17:36:06 +00:00
|
|
|
remote := o.remotePath()
|
|
|
|
if ignoredFiles.MatchString(remote) {
|
2019-10-08 18:56:12 +00:00
|
|
|
return fserrors.NoRetryError(errors.Errorf("file name %q is disallowed - not uploading", path.Base(remote)))
|
2015-08-20 17:36:06 +00:00
|
|
|
}
|
2020-01-14 17:33:35 +00:00
|
|
|
commitInfo := files.NewCommitInfo(o.fs.opt.Enc.FromStandardPath(o.remotePath()))
|
2017-05-21 20:35:33 +00:00
|
|
|
commitInfo.Mode.Tag = "overwrite"
|
|
|
|
// The Dropbox API only accepts timestamps in UTC with second precision.
|
2019-06-17 08:34:30 +00:00
|
|
|
commitInfo.ClientModified = src.ModTime(ctx).UTC().Round(time.Second)
|
2020-11-27 11:49:37 +00:00
|
|
|
// Don't attempt to create filenames that are too long
|
|
|
|
if cErr := checkPathLength(commitInfo.Path); cErr != nil {
|
|
|
|
return cErr
|
|
|
|
}
|
2017-05-21 20:35:33 +00:00
|
|
|
|
|
|
|
size := src.Size()
|
|
|
|
var err error
|
|
|
|
var entry *files.FileMetadata
|
2018-05-14 17:06:57 +00:00
|
|
|
if size > int64(o.fs.opt.ChunkSize) || size == -1 {
|
2017-05-21 20:35:33 +00:00
|
|
|
entry, err = o.uploadChunked(in, commitInfo, size)
|
|
|
|
} else {
|
2017-05-28 16:55:18 +00:00
|
|
|
err = o.fs.pacer.CallNoRetry(func() (bool, error) {
|
|
|
|
entry, err = o.fs.srv.Upload(commitInfo, in)
|
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
2017-05-21 20:35:33 +00:00
|
|
|
}
|
2014-07-08 20:59:30 +00:00
|
|
|
if err != nil {
|
2016-06-12 14:06:02 +00:00
|
|
|
return errors.Wrap(err, "upload failed")
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
2017-02-25 11:09:57 +00:00
|
|
|
return o.setMetadataFromEntry(entry)
|
2014-07-08 20:59:30 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Remove an object
|
2019-06-17 08:34:30 +00:00
|
|
|
func (o *Object) Remove(ctx context.Context) (err error) {
|
2020-08-23 14:10:30 +00:00
|
|
|
if o.fs.opt.SharedFiles || o.fs.opt.SharedFolders {
|
2020-11-26 11:58:43 +00:00
|
|
|
return errNotSupportedInSharedMode
|
2020-08-23 14:10:30 +00:00
|
|
|
}
|
2017-11-08 09:18:16 +00:00
|
|
|
err = o.fs.pacer.Call(func() (bool, error) {
|
2018-11-02 12:14:02 +00:00
|
|
|
_, err = o.fs.srv.DeleteV2(&files.DeleteArg{
|
2020-01-14 17:33:35 +00:00
|
|
|
Path: o.fs.opt.Enc.FromStandardPath(o.remotePath()),
|
2018-11-02 12:14:02 +00:00
|
|
|
})
|
2017-05-28 16:55:18 +00:00
|
|
|
return shouldRetry(err)
|
|
|
|
})
|
2014-07-08 20:59:30 +00:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check the interfaces are satisfied
|
2015-08-31 20:05:51 +00:00
|
|
|
var (
|
2018-03-29 07:10:19 +00:00
|
|
|
_ fs.Fs = (*Fs)(nil)
|
|
|
|
_ fs.Copier = (*Fs)(nil)
|
|
|
|
_ fs.Purger = (*Fs)(nil)
|
|
|
|
_ fs.PutStreamer = (*Fs)(nil)
|
|
|
|
_ fs.Mover = (*Fs)(nil)
|
|
|
|
_ fs.PublicLinker = (*Fs)(nil)
|
|
|
|
_ fs.DirMover = (*Fs)(nil)
|
2018-04-16 21:19:25 +00:00
|
|
|
_ fs.Abouter = (*Fs)(nil)
|
2018-03-29 07:10:19 +00:00
|
|
|
_ fs.Object = (*Object)(nil)
|
2020-08-26 23:12:10 +00:00
|
|
|
_ fs.IDer = (*Object)(nil)
|
2015-08-31 20:05:51 +00:00
|
|
|
)
|