2018-04-13 10:51:28 +00:00
|
|
|
package mega
|
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
|
|
|
"crypto/aes"
|
|
|
|
"crypto/cipher"
|
|
|
|
"crypto/rand"
|
|
|
|
"encoding/json"
|
|
|
|
"errors"
|
|
|
|
"fmt"
|
|
|
|
"io"
|
|
|
|
"io/ioutil"
|
|
|
|
"log"
|
|
|
|
"math/big"
|
|
|
|
mrand "math/rand"
|
|
|
|
"net/http"
|
|
|
|
"os"
|
|
|
|
"path/filepath"
|
|
|
|
"strings"
|
|
|
|
"sync"
|
|
|
|
"time"
|
|
|
|
)
|
|
|
|
|
|
|
|
// Default settings
|
|
|
|
const (
|
|
|
|
API_URL = "https://g.api.mega.co.nz"
|
|
|
|
BASE_DOWNLOAD_URL = "https://mega.co.nz"
|
|
|
|
RETRIES = 10
|
|
|
|
DOWNLOAD_WORKERS = 3
|
|
|
|
MAX_DOWNLOAD_WORKERS = 30
|
|
|
|
UPLOAD_WORKERS = 1
|
|
|
|
MAX_UPLOAD_WORKERS = 30
|
|
|
|
TIMEOUT = time.Second * 10
|
|
|
|
minSleepTime = 10 * time.Millisecond // for retries
|
|
|
|
maxSleepTime = 5 * time.Second // for retries
|
|
|
|
)
|
|
|
|
|
|
|
|
type config struct {
|
|
|
|
baseurl string
|
|
|
|
retries int
|
|
|
|
dl_workers int
|
|
|
|
ul_workers int
|
|
|
|
timeout time.Duration
|
|
|
|
}
|
|
|
|
|
|
|
|
func newConfig() config {
|
|
|
|
return config{
|
|
|
|
baseurl: API_URL,
|
|
|
|
retries: RETRIES,
|
|
|
|
dl_workers: DOWNLOAD_WORKERS,
|
|
|
|
ul_workers: UPLOAD_WORKERS,
|
|
|
|
timeout: TIMEOUT,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set mega service base url
|
|
|
|
func (c *config) SetAPIUrl(u string) {
|
|
|
|
if strings.HasSuffix(u, "/") {
|
|
|
|
u = strings.TrimRight(u, "/")
|
|
|
|
}
|
|
|
|
c.baseurl = u
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set number of retries for api calls
|
|
|
|
func (c *config) SetRetries(r int) {
|
|
|
|
c.retries = r
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set concurrent download workers
|
|
|
|
func (c *config) SetDownloadWorkers(w int) error {
|
|
|
|
if w <= MAX_DOWNLOAD_WORKERS {
|
|
|
|
c.dl_workers = w
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
return EWORKER_LIMIT_EXCEEDED
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set connection timeout
|
|
|
|
func (c *config) SetTimeOut(t time.Duration) {
|
|
|
|
c.timeout = t
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set concurrent upload workers
|
|
|
|
func (c *config) SetUploadWorkers(w int) error {
|
|
|
|
if w <= MAX_UPLOAD_WORKERS {
|
|
|
|
c.ul_workers = w
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
return EWORKER_LIMIT_EXCEEDED
|
|
|
|
}
|
|
|
|
|
|
|
|
type Mega struct {
|
|
|
|
config
|
|
|
|
// Sequence number
|
|
|
|
sn int64
|
|
|
|
// Server state sn
|
|
|
|
ssn string
|
|
|
|
// Session ID
|
|
|
|
sid []byte
|
|
|
|
// Master key
|
|
|
|
k []byte
|
|
|
|
// User handle
|
|
|
|
uh []byte
|
|
|
|
// Filesystem object
|
|
|
|
FS *MegaFS
|
|
|
|
// HTTP Client
|
|
|
|
client *http.Client
|
|
|
|
// Loggers
|
|
|
|
logf func(format string, v ...interface{})
|
|
|
|
debugf func(format string, v ...interface{})
|
|
|
|
// serialize the API requests
|
|
|
|
apiMu sync.Mutex
|
2018-06-30 10:13:06 +00:00
|
|
|
// mutex to protext waitEvents
|
|
|
|
waitEventsMu sync.Mutex
|
|
|
|
// Outstanding channels to close to indicate events all received
|
|
|
|
waitEvents []chan struct{}
|
2018-04-13 10:51:28 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Filesystem node types
|
|
|
|
const (
|
|
|
|
FILE = 0
|
|
|
|
FOLDER = 1
|
|
|
|
ROOT = 2
|
|
|
|
INBOX = 3
|
|
|
|
TRASH = 4
|
|
|
|
)
|
|
|
|
|
|
|
|
// Filesystem node
|
|
|
|
type Node struct {
|
|
|
|
fs *MegaFS
|
|
|
|
name string
|
|
|
|
hash string
|
|
|
|
parent *Node
|
|
|
|
children []*Node
|
|
|
|
ntype int
|
|
|
|
size int64
|
|
|
|
ts time.Time
|
|
|
|
meta NodeMeta
|
|
|
|
}
|
|
|
|
|
|
|
|
func (n *Node) removeChild(c *Node) bool {
|
|
|
|
index := -1
|
|
|
|
for i, v := range n.children {
|
|
|
|
if v.hash == c.hash {
|
|
|
|
index = i
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if index >= 0 {
|
|
|
|
n.children[index] = n.children[len(n.children)-1]
|
|
|
|
n.children = n.children[:len(n.children)-1]
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
func (n *Node) addChild(c *Node) {
|
|
|
|
if n != nil {
|
|
|
|
n.children = append(n.children, c)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (n *Node) getChildren() []*Node {
|
|
|
|
return n.children
|
|
|
|
}
|
|
|
|
|
|
|
|
func (n *Node) GetType() int {
|
|
|
|
n.fs.mutex.Lock()
|
|
|
|
defer n.fs.mutex.Unlock()
|
|
|
|
return n.ntype
|
|
|
|
}
|
|
|
|
|
|
|
|
func (n *Node) GetSize() int64 {
|
|
|
|
n.fs.mutex.Lock()
|
|
|
|
defer n.fs.mutex.Unlock()
|
|
|
|
return n.size
|
|
|
|
}
|
|
|
|
|
|
|
|
func (n *Node) GetTimeStamp() time.Time {
|
|
|
|
n.fs.mutex.Lock()
|
|
|
|
defer n.fs.mutex.Unlock()
|
|
|
|
return n.ts
|
|
|
|
}
|
|
|
|
|
|
|
|
func (n *Node) GetName() string {
|
|
|
|
n.fs.mutex.Lock()
|
|
|
|
defer n.fs.mutex.Unlock()
|
|
|
|
return n.name
|
|
|
|
}
|
|
|
|
|
|
|
|
func (n *Node) GetHash() string {
|
|
|
|
n.fs.mutex.Lock()
|
|
|
|
defer n.fs.mutex.Unlock()
|
|
|
|
return n.hash
|
|
|
|
}
|
|
|
|
|
|
|
|
type NodeMeta struct {
|
|
|
|
key []byte
|
|
|
|
compkey []byte
|
|
|
|
iv []byte
|
|
|
|
mac []byte
|
|
|
|
}
|
|
|
|
|
|
|
|
// Mega filesystem object
|
|
|
|
type MegaFS struct {
|
|
|
|
root *Node
|
|
|
|
trash *Node
|
|
|
|
inbox *Node
|
|
|
|
sroots []*Node
|
|
|
|
lookup map[string]*Node
|
|
|
|
skmap map[string]string
|
|
|
|
mutex sync.Mutex
|
|
|
|
}
|
|
|
|
|
|
|
|
// Get filesystem root node
|
|
|
|
func (fs *MegaFS) GetRoot() *Node {
|
|
|
|
fs.mutex.Lock()
|
|
|
|
defer fs.mutex.Unlock()
|
|
|
|
return fs.root
|
|
|
|
}
|
|
|
|
|
|
|
|
// Get filesystem trash node
|
|
|
|
func (fs *MegaFS) GetTrash() *Node {
|
|
|
|
fs.mutex.Lock()
|
|
|
|
defer fs.mutex.Unlock()
|
|
|
|
return fs.trash
|
|
|
|
}
|
|
|
|
|
|
|
|
// Get inbox node
|
|
|
|
func (fs *MegaFS) GetInbox() *Node {
|
|
|
|
fs.mutex.Lock()
|
|
|
|
defer fs.mutex.Unlock()
|
|
|
|
return fs.inbox
|
|
|
|
}
|
|
|
|
|
|
|
|
// Get a node pointer from its hash
|
|
|
|
func (fs *MegaFS) HashLookup(h string) *Node {
|
|
|
|
fs.mutex.Lock()
|
|
|
|
defer fs.mutex.Unlock()
|
|
|
|
|
|
|
|
return fs.hashLookup(h)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (fs *MegaFS) hashLookup(h string) *Node {
|
|
|
|
if node, ok := fs.lookup[h]; ok {
|
|
|
|
return node
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Get the list of child nodes for a given node
|
|
|
|
func (fs *MegaFS) GetChildren(n *Node) ([]*Node, error) {
|
|
|
|
fs.mutex.Lock()
|
|
|
|
defer fs.mutex.Unlock()
|
|
|
|
|
|
|
|
var empty []*Node
|
|
|
|
|
|
|
|
if n == nil {
|
|
|
|
return empty, EARGS
|
|
|
|
}
|
|
|
|
|
|
|
|
node := fs.hashLookup(n.hash)
|
|
|
|
if node == nil {
|
|
|
|
return empty, ENOENT
|
|
|
|
}
|
|
|
|
|
|
|
|
return node.getChildren(), nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Retreive all the nodes in the given node tree path by name
|
|
|
|
// This method returns array of nodes upto the matched subpath
|
|
|
|
// (in same order as input names array) even if the target node is not located.
|
|
|
|
func (fs *MegaFS) PathLookup(root *Node, ns []string) ([]*Node, error) {
|
|
|
|
fs.mutex.Lock()
|
|
|
|
defer fs.mutex.Unlock()
|
|
|
|
|
|
|
|
if root == nil {
|
|
|
|
return nil, EARGS
|
|
|
|
}
|
|
|
|
|
|
|
|
var err error
|
|
|
|
var found bool = true
|
|
|
|
|
|
|
|
nodepath := []*Node{}
|
|
|
|
|
|
|
|
children := root.children
|
|
|
|
for _, name := range ns {
|
|
|
|
found = false
|
|
|
|
for _, n := range children {
|
|
|
|
if n.name == name {
|
|
|
|
nodepath = append(nodepath, n)
|
|
|
|
children = n.children
|
|
|
|
found = true
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if found == false {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if found == false {
|
|
|
|
err = ENOENT
|
|
|
|
}
|
|
|
|
|
|
|
|
return nodepath, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Get top level directory nodes shared by other users
|
|
|
|
func (fs *MegaFS) GetSharedRoots() []*Node {
|
|
|
|
fs.mutex.Lock()
|
|
|
|
defer fs.mutex.Unlock()
|
|
|
|
return fs.sroots
|
|
|
|
}
|
|
|
|
|
|
|
|
func newMegaFS() *MegaFS {
|
|
|
|
fs := &MegaFS{
|
|
|
|
lookup: make(map[string]*Node),
|
|
|
|
skmap: make(map[string]string),
|
|
|
|
}
|
|
|
|
return fs
|
|
|
|
}
|
|
|
|
|
|
|
|
func New() *Mega {
|
|
|
|
max := big.NewInt(0x100000000)
|
|
|
|
bigx, _ := rand.Int(rand.Reader, max)
|
|
|
|
cfg := newConfig()
|
|
|
|
mgfs := newMegaFS()
|
|
|
|
m := &Mega{
|
|
|
|
config: cfg,
|
|
|
|
sn: bigx.Int64(),
|
|
|
|
FS: mgfs,
|
|
|
|
client: newHttpClient(cfg.timeout),
|
|
|
|
}
|
|
|
|
m.SetLogger(log.Printf)
|
|
|
|
m.SetDebugger(nil)
|
|
|
|
return m
|
|
|
|
}
|
|
|
|
|
|
|
|
// SetClient sets the HTTP client in use
|
|
|
|
func (m *Mega) SetClient(client *http.Client) *Mega {
|
|
|
|
m.client = client
|
|
|
|
return m
|
|
|
|
}
|
|
|
|
|
|
|
|
// discardLogf discards the log messages
|
|
|
|
func discardLogf(format string, v ...interface{}) {
|
|
|
|
}
|
|
|
|
|
|
|
|
// SetLogger sets the logger for important messages. By default this
|
|
|
|
// is log.Printf. Use nil to discard the messages.
|
|
|
|
func (m *Mega) SetLogger(logf func(format string, v ...interface{})) *Mega {
|
|
|
|
if logf == nil {
|
|
|
|
logf = discardLogf
|
|
|
|
}
|
|
|
|
m.logf = logf
|
|
|
|
return m
|
|
|
|
}
|
|
|
|
|
|
|
|
// SetDebugger sets the logger for debug messages. By default these
|
|
|
|
// messages are not output.
|
|
|
|
func (m *Mega) SetDebugger(debugf func(format string, v ...interface{})) *Mega {
|
|
|
|
if debugf == nil {
|
|
|
|
debugf = discardLogf
|
|
|
|
}
|
|
|
|
m.debugf = debugf
|
|
|
|
return m
|
|
|
|
}
|
|
|
|
|
|
|
|
// backOffSleep sleeps for the time pointed to then adjusts it by
|
|
|
|
// doubling it up to a maximum of maxSleepTime.
|
|
|
|
//
|
|
|
|
// This produces a truncated exponential backoff sleep
|
|
|
|
func backOffSleep(pt *time.Duration) {
|
|
|
|
time.Sleep(*pt)
|
|
|
|
*pt *= 2
|
|
|
|
if *pt > maxSleepTime {
|
|
|
|
*pt = maxSleepTime
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// API request method
|
|
|
|
func (m *Mega) api_request(r []byte) (buf []byte, err error) {
|
|
|
|
var resp *http.Response
|
|
|
|
// serialize the API requests
|
|
|
|
m.apiMu.Lock()
|
|
|
|
defer func() {
|
|
|
|
m.sn++
|
|
|
|
m.apiMu.Unlock()
|
|
|
|
}()
|
|
|
|
|
|
|
|
url := fmt.Sprintf("%s/cs?id=%d", m.baseurl, m.sn)
|
|
|
|
|
|
|
|
if m.sid != nil {
|
|
|
|
url = fmt.Sprintf("%s&sid=%s", url, string(m.sid))
|
|
|
|
}
|
|
|
|
|
|
|
|
sleepTime := minSleepTime // inital backoff time
|
|
|
|
for i := 0; i < m.retries+1; i++ {
|
|
|
|
if i != 0 {
|
|
|
|
m.debugf("Retry API request %d/%d: %v", i, m.retries, err)
|
|
|
|
backOffSleep(&sleepTime)
|
|
|
|
}
|
|
|
|
resp, err = m.client.Post(url, "application/json", bytes.NewBuffer(r))
|
|
|
|
if err != nil {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if resp.StatusCode != 200 {
|
|
|
|
// err must be not-nil on a continue
|
|
|
|
err = errors.New("Http Status: " + resp.Status)
|
|
|
|
_ = resp.Body.Close()
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
buf, err = ioutil.ReadAll(resp.Body)
|
|
|
|
if err != nil {
|
|
|
|
_ = resp.Body.Close()
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
err = resp.Body.Close()
|
|
|
|
if err != nil {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
// at this point the body is read and closed
|
|
|
|
|
|
|
|
if bytes.HasPrefix(buf, []byte("[")) == false && bytes.HasPrefix(buf, []byte("-")) == false {
|
|
|
|
return nil, EBADRESP
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(buf) < 6 {
|
|
|
|
var emsg [1]ErrorMsg
|
|
|
|
err = json.Unmarshal(buf, &emsg)
|
|
|
|
if err != nil {
|
|
|
|
err = json.Unmarshal(buf, &emsg[0])
|
|
|
|
}
|
|
|
|
if err != nil {
|
|
|
|
return buf, EBADRESP
|
|
|
|
}
|
|
|
|
err = parseError(emsg[0])
|
|
|
|
if err == EAGAIN {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
return buf, err
|
|
|
|
}
|
|
|
|
|
|
|
|
if err == nil {
|
|
|
|
return buf, nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Authenticate and start a session
|
|
|
|
func (m *Mega) Login(email string, passwd string) error {
|
|
|
|
var msg [1]LoginMsg
|
|
|
|
var res [1]LoginResp
|
|
|
|
var err error
|
|
|
|
var result []byte
|
|
|
|
|
2018-08-17 19:54:00 +00:00
|
|
|
email = strings.ToLower(email) // mega uses lowercased emails for login purposes
|
|
|
|
|
2018-04-13 10:51:28 +00:00
|
|
|
passkey := password_key(passwd)
|
|
|
|
uhandle := stringhash(email, passkey)
|
|
|
|
m.uh = make([]byte, len(uhandle))
|
|
|
|
copy(m.uh, uhandle)
|
|
|
|
|
|
|
|
msg[0].Cmd = "us"
|
|
|
|
msg[0].User = email
|
|
|
|
msg[0].Handle = string(uhandle)
|
|
|
|
|
|
|
|
req, _ := json.Marshal(msg)
|
|
|
|
result, err = m.api_request(req)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
err = json.Unmarshal(result, &res)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
m.k = base64urldecode([]byte(res[0].Key))
|
|
|
|
cipher, err := aes.NewCipher(passkey)
|
|
|
|
cipher.Decrypt(m.k, m.k)
|
|
|
|
m.sid, err = decryptSessionId([]byte(res[0].Privk), []byte(res[0].Csid), m.k)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-06-30 10:13:06 +00:00
|
|
|
waitEvent := m.WaitEventsStart()
|
|
|
|
|
2018-04-13 10:51:28 +00:00
|
|
|
err = m.getFileSystem()
|
2018-06-30 10:13:06 +00:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2018-04-13 10:51:28 +00:00
|
|
|
|
2018-06-30 10:13:06 +00:00
|
|
|
// Wait until the all the pending events have been received
|
|
|
|
m.WaitEvents(waitEvent, 5*time.Second)
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// WaitEventsStart - call this before you do the action which might
|
|
|
|
// generate events then use the returned channel as a parameter to
|
|
|
|
// WaitEvents to wait for the event(s) to be received.
|
|
|
|
func (m *Mega) WaitEventsStart() <-chan struct{} {
|
|
|
|
ch := make(chan struct{})
|
|
|
|
m.waitEventsMu.Lock()
|
|
|
|
m.waitEvents = append(m.waitEvents, ch)
|
|
|
|
m.waitEventsMu.Unlock()
|
|
|
|
return ch
|
|
|
|
}
|
|
|
|
|
|
|
|
// WaitEvents waits for all outstanding events to be received for a
|
|
|
|
// maximum of duration. eventChan should be a channel as returned
|
|
|
|
// from WaitEventStart.
|
|
|
|
//
|
|
|
|
// If the timeout elapsed then it returns true otherwise false.
|
|
|
|
func (m *Mega) WaitEvents(eventChan <-chan struct{}, duration time.Duration) (timedout bool) {
|
|
|
|
m.debugf("Waiting for events to be finished for %v", duration)
|
|
|
|
timer := time.NewTimer(duration)
|
|
|
|
select {
|
|
|
|
case <-eventChan:
|
|
|
|
m.debugf("Events received")
|
|
|
|
timedout = false
|
|
|
|
case <-timer.C:
|
|
|
|
m.debugf("Timeout waiting for events")
|
|
|
|
timedout = true
|
|
|
|
}
|
|
|
|
timer.Stop()
|
|
|
|
return timedout
|
|
|
|
}
|
|
|
|
|
|
|
|
// waitEventsFire - fire the wait event
|
|
|
|
func (m *Mega) waitEventsFire() {
|
|
|
|
m.waitEventsMu.Lock()
|
|
|
|
if len(m.waitEvents) > 0 {
|
|
|
|
m.debugf("Signalling events received")
|
|
|
|
for _, ch := range m.waitEvents {
|
|
|
|
close(ch)
|
|
|
|
}
|
|
|
|
m.waitEvents = nil
|
|
|
|
}
|
|
|
|
m.waitEventsMu.Unlock()
|
2018-04-13 10:51:28 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Get user information
|
|
|
|
func (m *Mega) GetUser() (UserResp, error) {
|
|
|
|
var msg [1]UserMsg
|
|
|
|
var res [1]UserResp
|
|
|
|
|
|
|
|
msg[0].Cmd = "ug"
|
|
|
|
|
|
|
|
req, _ := json.Marshal(msg)
|
|
|
|
result, err := m.api_request(req)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
return res[0], err
|
|
|
|
}
|
|
|
|
|
|
|
|
err = json.Unmarshal(result, &res)
|
|
|
|
return res[0], err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Get quota information
|
|
|
|
func (m *Mega) GetQuota() (QuotaResp, error) {
|
|
|
|
var msg [1]QuotaMsg
|
|
|
|
var res [1]QuotaResp
|
|
|
|
|
|
|
|
msg[0].Cmd = "uq"
|
|
|
|
msg[0].Xfer = 1
|
|
|
|
msg[0].Strg = 1
|
|
|
|
|
|
|
|
req, _ := json.Marshal(msg)
|
|
|
|
result, err := m.api_request(req)
|
|
|
|
if err != nil {
|
|
|
|
return res[0], err
|
|
|
|
}
|
|
|
|
|
|
|
|
err = json.Unmarshal(result, &res)
|
|
|
|
return res[0], err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Add a node into filesystem
|
|
|
|
func (m *Mega) addFSNode(itm FSNode) (*Node, error) {
|
|
|
|
var compkey, key []uint32
|
|
|
|
var attr FileAttr
|
|
|
|
var node, parent *Node
|
|
|
|
var err error
|
|
|
|
|
|
|
|
master_aes, _ := aes.NewCipher(m.k)
|
|
|
|
|
|
|
|
switch {
|
|
|
|
case itm.T == FOLDER || itm.T == FILE:
|
|
|
|
args := strings.Split(itm.Key, ":")
|
|
|
|
|
|
|
|
switch {
|
|
|
|
// File or folder owned by current user
|
|
|
|
case args[0] == itm.User:
|
|
|
|
buf := base64urldecode([]byte(args[1]))
|
|
|
|
err = blockDecrypt(master_aes, buf, buf)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
compkey = bytes_to_a32(buf)
|
|
|
|
// Shared folder
|
|
|
|
case itm.SUser != "" && itm.SKey != "":
|
|
|
|
sk := base64urldecode([]byte(itm.SKey))
|
|
|
|
err = blockDecrypt(master_aes, sk, sk)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
sk_aes, _ := aes.NewCipher(sk)
|
|
|
|
|
|
|
|
m.FS.skmap[itm.Hash] = itm.SKey
|
|
|
|
buf := base64urldecode([]byte(args[1]))
|
|
|
|
err = blockDecrypt(sk_aes, buf, buf)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
compkey = bytes_to_a32(buf)
|
|
|
|
// Shared file
|
|
|
|
default:
|
|
|
|
k := m.FS.skmap[args[0]]
|
|
|
|
b := base64urldecode([]byte(k))
|
|
|
|
err = blockDecrypt(master_aes, b, b)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
block, _ := aes.NewCipher(b)
|
|
|
|
buf := base64urldecode([]byte(args[1]))
|
|
|
|
err = blockDecrypt(block, buf, buf)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
compkey = bytes_to_a32(buf)
|
|
|
|
}
|
|
|
|
|
|
|
|
switch {
|
|
|
|
case itm.T == FILE:
|
|
|
|
key = []uint32{compkey[0] ^ compkey[4], compkey[1] ^ compkey[5], compkey[2] ^ compkey[6], compkey[3] ^ compkey[7]}
|
|
|
|
default:
|
|
|
|
key = compkey
|
|
|
|
}
|
|
|
|
|
|
|
|
attr, err = decryptAttr(a32_to_bytes(key), []byte(itm.Attr))
|
|
|
|
// FIXME:
|
|
|
|
if err != nil {
|
|
|
|
attr.Name = "BAD ATTRIBUTE"
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
n, ok := m.FS.lookup[itm.Hash]
|
|
|
|
switch {
|
|
|
|
case ok:
|
|
|
|
node = n
|
|
|
|
default:
|
|
|
|
node = &Node{
|
|
|
|
fs: m.FS,
|
|
|
|
ntype: itm.T,
|
|
|
|
size: itm.Sz,
|
|
|
|
ts: time.Unix(itm.Ts, 0),
|
|
|
|
}
|
|
|
|
|
|
|
|
m.FS.lookup[itm.Hash] = node
|
|
|
|
}
|
|
|
|
|
|
|
|
n, ok = m.FS.lookup[itm.Parent]
|
|
|
|
switch {
|
|
|
|
case ok:
|
|
|
|
parent = n
|
|
|
|
parent.removeChild(node)
|
|
|
|
parent.addChild(node)
|
|
|
|
default:
|
|
|
|
parent = nil
|
|
|
|
if itm.Parent != "" {
|
|
|
|
parent = &Node{
|
|
|
|
fs: m.FS,
|
|
|
|
children: []*Node{node},
|
|
|
|
ntype: FOLDER,
|
|
|
|
}
|
|
|
|
m.FS.lookup[itm.Parent] = parent
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
switch {
|
|
|
|
case itm.T == FILE:
|
|
|
|
var meta NodeMeta
|
|
|
|
meta.key = a32_to_bytes(key)
|
|
|
|
meta.iv = a32_to_bytes([]uint32{compkey[4], compkey[5], 0, 0})
|
|
|
|
meta.mac = a32_to_bytes([]uint32{compkey[6], compkey[7]})
|
|
|
|
meta.compkey = a32_to_bytes(compkey)
|
|
|
|
node.meta = meta
|
|
|
|
case itm.T == FOLDER:
|
|
|
|
var meta NodeMeta
|
|
|
|
meta.key = a32_to_bytes(key)
|
|
|
|
meta.compkey = a32_to_bytes(compkey)
|
|
|
|
node.meta = meta
|
|
|
|
case itm.T == ROOT:
|
|
|
|
attr.Name = "Cloud Drive"
|
|
|
|
m.FS.root = node
|
|
|
|
case itm.T == INBOX:
|
|
|
|
attr.Name = "InBox"
|
|
|
|
m.FS.inbox = node
|
|
|
|
case itm.T == TRASH:
|
|
|
|
attr.Name = "Trash"
|
|
|
|
m.FS.trash = node
|
|
|
|
}
|
|
|
|
|
|
|
|
// Shared directories
|
|
|
|
if itm.SUser != "" && itm.SKey != "" {
|
|
|
|
m.FS.sroots = append(m.FS.sroots, node)
|
|
|
|
}
|
|
|
|
|
|
|
|
node.name = attr.Name
|
|
|
|
node.hash = itm.Hash
|
|
|
|
node.parent = parent
|
|
|
|
node.ntype = itm.T
|
|
|
|
|
|
|
|
return node, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Get all nodes from filesystem
|
|
|
|
func (m *Mega) getFileSystem() error {
|
|
|
|
m.FS.mutex.Lock()
|
|
|
|
defer m.FS.mutex.Unlock()
|
|
|
|
|
|
|
|
var msg [1]FilesMsg
|
|
|
|
var res [1]FilesResp
|
|
|
|
|
|
|
|
msg[0].Cmd = "f"
|
|
|
|
msg[0].C = 1
|
|
|
|
|
|
|
|
req, _ := json.Marshal(msg)
|
|
|
|
result, err := m.api_request(req)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
err = json.Unmarshal(result, &res)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, sk := range res[0].Ok {
|
|
|
|
m.FS.skmap[sk.Hash] = sk.Key
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, itm := range res[0].F {
|
|
|
|
_, err = m.addFSNode(itm)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
m.ssn = res[0].Sn
|
|
|
|
|
|
|
|
go m.pollEvents()
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Download contains the internal state of a download
|
|
|
|
type Download struct {
|
|
|
|
m *Mega
|
|
|
|
src *Node
|
|
|
|
resourceUrl string
|
|
|
|
aes_block cipher.Block
|
|
|
|
iv []byte
|
|
|
|
mac_enc cipher.BlockMode
|
|
|
|
mutex sync.Mutex // to protect the following
|
|
|
|
chunks []chunkSize
|
|
|
|
chunk_macs [][]byte
|
|
|
|
}
|
|
|
|
|
|
|
|
// an all nil IV for mac calculations
|
|
|
|
var zero_iv = make([]byte, 16)
|
|
|
|
|
|
|
|
// Create a new Download from the src Node
|
|
|
|
//
|
|
|
|
// Call Chunks to find out how many chunks there are, then for id =
|
|
|
|
// 0..chunks-1 call DownloadChunk. Finally call Finish() to receive
|
|
|
|
// the error status.
|
|
|
|
func (m *Mega) NewDownload(src *Node) (*Download, error) {
|
|
|
|
if src == nil {
|
|
|
|
return nil, EARGS
|
|
|
|
}
|
|
|
|
|
|
|
|
var msg [1]DownloadMsg
|
|
|
|
var res [1]DownloadResp
|
|
|
|
|
|
|
|
m.FS.mutex.Lock()
|
|
|
|
msg[0].Cmd = "g"
|
|
|
|
msg[0].G = 1
|
|
|
|
msg[0].N = src.hash
|
|
|
|
key := src.meta.key
|
|
|
|
m.FS.mutex.Unlock()
|
|
|
|
|
|
|
|
request, err := json.Marshal(msg)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
result, err := m.api_request(request)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
err = json.Unmarshal(result, &res)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
_, err = decryptAttr(key, []byte(res[0].Attr))
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
chunks := getChunkSizes(int64(res[0].Size))
|
|
|
|
|
|
|
|
aes_block, err := aes.NewCipher(key)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
mac_enc := cipher.NewCBCEncrypter(aes_block, zero_iv)
|
2018-05-02 16:09:45 +00:00
|
|
|
m.FS.mutex.Lock()
|
2018-04-13 10:51:28 +00:00
|
|
|
t := bytes_to_a32(src.meta.iv)
|
2018-05-02 16:09:45 +00:00
|
|
|
m.FS.mutex.Unlock()
|
2018-04-13 10:51:28 +00:00
|
|
|
iv := a32_to_bytes([]uint32{t[0], t[1], t[0], t[1]})
|
|
|
|
|
|
|
|
d := &Download{
|
|
|
|
m: m,
|
|
|
|
src: src,
|
|
|
|
resourceUrl: res[0].G,
|
|
|
|
aes_block: aes_block,
|
|
|
|
iv: iv,
|
|
|
|
mac_enc: mac_enc,
|
|
|
|
chunks: chunks,
|
|
|
|
chunk_macs: make([][]byte, len(chunks)),
|
|
|
|
}
|
|
|
|
return d, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Chunks returns The number of chunks in the download.
|
|
|
|
func (d *Download) Chunks() int {
|
|
|
|
return len(d.chunks)
|
|
|
|
}
|
|
|
|
|
|
|
|
// ChunkLocation returns the position in the file and the size of the chunk
|
|
|
|
func (d *Download) ChunkLocation(id int) (position int64, size int, err error) {
|
|
|
|
if id < 0 || id >= len(d.chunks) {
|
|
|
|
return 0, 0, EARGS
|
|
|
|
}
|
|
|
|
d.mutex.Lock()
|
|
|
|
defer d.mutex.Unlock()
|
|
|
|
return d.chunks[id].position, d.chunks[id].size, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// DownloadChunk gets a chunk with the given number and update the
|
|
|
|
// mac, returning the position in the file of the chunk
|
|
|
|
func (d *Download) DownloadChunk(id int) (chunk []byte, err error) {
|
|
|
|
if id < 0 || id >= len(d.chunks) {
|
|
|
|
return nil, EARGS
|
|
|
|
}
|
|
|
|
|
|
|
|
chk_start, chk_size, err := d.ChunkLocation(id)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
var resp *http.Response
|
|
|
|
chunk_url := fmt.Sprintf("%s/%d-%d", d.resourceUrl, chk_start, chk_start+int64(chk_size)-1)
|
|
|
|
sleepTime := minSleepTime // inital backoff time
|
|
|
|
for retry := 0; retry < d.m.retries+1; retry++ {
|
|
|
|
resp, err = d.m.client.Get(chunk_url)
|
|
|
|
if err == nil {
|
|
|
|
if resp.StatusCode == 200 {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
err = errors.New("Http Status: " + resp.Status)
|
|
|
|
_ = resp.Body.Close()
|
|
|
|
}
|
|
|
|
d.m.debugf("%s: Retry download chunk %d/%d: %v", d.src.name, retry, d.m.retries, err)
|
|
|
|
backOffSleep(&sleepTime)
|
|
|
|
}
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
if resp == nil {
|
|
|
|
return nil, errors.New("retries exceeded")
|
|
|
|
}
|
|
|
|
|
|
|
|
chunk, err = ioutil.ReadAll(resp.Body)
|
|
|
|
if err != nil {
|
|
|
|
_ = resp.Body.Close()
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
err = resp.Body.Close()
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// body is read and closed here
|
|
|
|
|
|
|
|
if len(chunk) != chk_size {
|
|
|
|
return nil, errors.New("wrong size for downloaded chunk")
|
|
|
|
}
|
|
|
|
|
|
|
|
// Decrypt the block
|
|
|
|
ctr_iv := bytes_to_a32(d.src.meta.iv)
|
|
|
|
ctr_iv[2] = uint32(uint64(chk_start) / 0x1000000000)
|
|
|
|
ctr_iv[3] = uint32(chk_start / 0x10)
|
|
|
|
ctr_aes := cipher.NewCTR(d.aes_block, a32_to_bytes(ctr_iv))
|
|
|
|
ctr_aes.XORKeyStream(chunk, chunk)
|
|
|
|
|
|
|
|
// Update the chunk_macs
|
|
|
|
enc := cipher.NewCBCEncrypter(d.aes_block, d.iv)
|
|
|
|
i := 0
|
|
|
|
block := make([]byte, 16)
|
|
|
|
paddedChunk := paddnull(chunk, 16)
|
|
|
|
for i = 0; i < len(paddedChunk); i += 16 {
|
|
|
|
enc.CryptBlocks(block, paddedChunk[i:i+16])
|
|
|
|
}
|
|
|
|
|
|
|
|
d.mutex.Lock()
|
|
|
|
if len(d.chunk_macs) > 0 {
|
|
|
|
d.chunk_macs[id] = make([]byte, 16)
|
|
|
|
copy(d.chunk_macs[id], block)
|
|
|
|
}
|
|
|
|
d.mutex.Unlock()
|
|
|
|
|
|
|
|
return chunk, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Finish checks the accumulated MAC for each block.
|
|
|
|
//
|
|
|
|
// If all the chunks weren't downloaded then it will just return nil
|
|
|
|
func (d *Download) Finish() (err error) {
|
|
|
|
// Can't check a 0 sized file
|
|
|
|
if len(d.chunk_macs) == 0 {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
mac_data := make([]byte, 16)
|
|
|
|
for _, v := range d.chunk_macs {
|
|
|
|
// If a chunk_macs hasn't been set then the whole file
|
|
|
|
// wasn't downloaded and we can't check it
|
|
|
|
if v == nil {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
d.mac_enc.CryptBlocks(mac_data, v)
|
|
|
|
}
|
|
|
|
|
|
|
|
tmac := bytes_to_a32(mac_data)
|
|
|
|
if bytes.Equal(a32_to_bytes([]uint32{tmac[0] ^ tmac[1], tmac[2] ^ tmac[3]}), d.src.meta.mac) == false {
|
|
|
|
return EMACMISMATCH
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Download file from filesystem reporting progress if not nil
|
|
|
|
func (m *Mega) DownloadFile(src *Node, dstpath string, progress *chan int) error {
|
|
|
|
defer func() {
|
|
|
|
if progress != nil {
|
|
|
|
close(*progress)
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
d, err := m.NewDownload(src)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
_, err = os.Stat(dstpath)
|
|
|
|
if os.IsExist(err) {
|
|
|
|
err = os.Remove(dstpath)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
outfile, err := os.OpenFile(dstpath, os.O_RDWR|os.O_CREATE, 0600)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
workch := make(chan int)
|
|
|
|
errch := make(chan error, m.dl_workers)
|
|
|
|
wg := sync.WaitGroup{}
|
|
|
|
|
|
|
|
// Fire chunk download workers
|
|
|
|
for w := 0; w < m.dl_workers; w++ {
|
|
|
|
wg.Add(1)
|
|
|
|
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
|
|
|
|
|
|
|
// Wait for work blocked on channel
|
|
|
|
for id := range workch {
|
|
|
|
chunk, err := d.DownloadChunk(id)
|
|
|
|
if err != nil {
|
|
|
|
errch <- err
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
chk_start, _, err := d.ChunkLocation(id)
|
|
|
|
if err != nil {
|
|
|
|
errch <- err
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
_, err = outfile.WriteAt(chunk, chk_start)
|
|
|
|
if err != nil {
|
|
|
|
errch <- err
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
if progress != nil {
|
|
|
|
*progress <- len(chunk)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
|
|
|
|
// Place chunk download jobs to chan
|
|
|
|
err = nil
|
|
|
|
for id := 0; id < d.Chunks() && err == nil; {
|
|
|
|
select {
|
|
|
|
case workch <- id:
|
|
|
|
id++
|
|
|
|
case err = <-errch:
|
|
|
|
}
|
|
|
|
}
|
|
|
|
close(workch)
|
|
|
|
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
closeErr := outfile.Close()
|
|
|
|
if err != nil {
|
|
|
|
_ = os.Remove(dstpath)
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if closeErr != nil {
|
|
|
|
return closeErr
|
|
|
|
}
|
|
|
|
|
|
|
|
return d.Finish()
|
|
|
|
}
|
|
|
|
|
|
|
|
// Upload contains the internal state of a upload
|
|
|
|
type Upload struct {
|
|
|
|
m *Mega
|
|
|
|
parenthash string
|
|
|
|
name string
|
|
|
|
uploadUrl string
|
|
|
|
aes_block cipher.Block
|
|
|
|
iv []byte
|
|
|
|
kiv []byte
|
|
|
|
mac_enc cipher.BlockMode
|
|
|
|
kbytes []byte
|
|
|
|
ukey []uint32
|
|
|
|
mutex sync.Mutex // to protect the following
|
|
|
|
chunks []chunkSize
|
|
|
|
chunk_macs [][]byte
|
|
|
|
completion_handle []byte
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create a new Upload of name into parent of fileSize
|
|
|
|
//
|
|
|
|
// Call Chunks to find out how many chunks there are, then for id =
|
|
|
|
// 0..chunks-1 Call ChunkLocation then UploadChunk. Finally call
|
|
|
|
// Finish() to receive the error status and the *Node.
|
|
|
|
func (m *Mega) NewUpload(parent *Node, name string, fileSize int64) (*Upload, error) {
|
|
|
|
if parent == nil {
|
|
|
|
return nil, EARGS
|
|
|
|
}
|
|
|
|
|
|
|
|
var msg [1]UploadMsg
|
|
|
|
var res [1]UploadResp
|
|
|
|
parenthash := parent.GetHash()
|
|
|
|
|
|
|
|
msg[0].Cmd = "u"
|
|
|
|
msg[0].S = fileSize
|
|
|
|
|
|
|
|
request, err := json.Marshal(msg)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
result, err := m.api_request(request)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
err = json.Unmarshal(result, &res)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
uploadUrl := res[0].P
|
|
|
|
ukey := []uint32{0, 0, 0, 0, 0, 0}
|
|
|
|
for i, _ := range ukey {
|
|
|
|
ukey[i] = uint32(mrand.Int31())
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
kbytes := a32_to_bytes(ukey[:4])
|
|
|
|
kiv := a32_to_bytes([]uint32{ukey[4], ukey[5], 0, 0})
|
|
|
|
aes_block, _ := aes.NewCipher(kbytes)
|
|
|
|
|
|
|
|
mac_enc := cipher.NewCBCEncrypter(aes_block, zero_iv)
|
|
|
|
iv := a32_to_bytes([]uint32{ukey[4], ukey[5], ukey[4], ukey[5]})
|
|
|
|
|
|
|
|
chunks := getChunkSizes(fileSize)
|
|
|
|
|
|
|
|
// File size is zero
|
|
|
|
// Do one empty request to get the completion handle
|
|
|
|
if len(chunks) == 0 {
|
|
|
|
chunks = append(chunks, chunkSize{position: 0, size: 0})
|
|
|
|
}
|
|
|
|
|
|
|
|
u := &Upload{
|
|
|
|
m: m,
|
|
|
|
parenthash: parenthash,
|
|
|
|
name: name,
|
|
|
|
uploadUrl: uploadUrl,
|
|
|
|
aes_block: aes_block,
|
|
|
|
iv: iv,
|
|
|
|
kiv: kiv,
|
|
|
|
mac_enc: mac_enc,
|
|
|
|
kbytes: kbytes,
|
|
|
|
ukey: ukey,
|
|
|
|
chunks: chunks,
|
|
|
|
chunk_macs: make([][]byte, len(chunks)),
|
|
|
|
completion_handle: []byte{},
|
|
|
|
}
|
|
|
|
return u, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Chunks returns The number of chunks in the upload.
|
|
|
|
func (u *Upload) Chunks() int {
|
|
|
|
return len(u.chunks)
|
|
|
|
}
|
|
|
|
|
|
|
|
// ChunkLocation returns the position in the file and the size of the chunk
|
|
|
|
func (u *Upload) ChunkLocation(id int) (position int64, size int, err error) {
|
|
|
|
if id < 0 || id >= len(u.chunks) {
|
|
|
|
return 0, 0, EARGS
|
|
|
|
}
|
|
|
|
return u.chunks[id].position, u.chunks[id].size, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// UploadChunk uploads the chunk of id
|
|
|
|
func (u *Upload) UploadChunk(id int, chunk []byte) (err error) {
|
|
|
|
chk_start, chk_size, err := u.ChunkLocation(id)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if len(chunk) != chk_size {
|
|
|
|
return errors.New("upload chunk is wrong size")
|
|
|
|
}
|
|
|
|
ctr_iv := bytes_to_a32(u.kiv)
|
|
|
|
ctr_iv[2] = uint32(uint64(chk_start) / 0x1000000000)
|
|
|
|
ctr_iv[3] = uint32(chk_start / 0x10)
|
|
|
|
ctr_aes := cipher.NewCTR(u.aes_block, a32_to_bytes(ctr_iv))
|
|
|
|
|
|
|
|
enc := cipher.NewCBCEncrypter(u.aes_block, u.iv)
|
|
|
|
|
|
|
|
i := 0
|
|
|
|
block := make([]byte, 16)
|
|
|
|
paddedchunk := paddnull(chunk, 16)
|
|
|
|
for i = 0; i < len(paddedchunk); i += 16 {
|
|
|
|
copy(block[0:16], paddedchunk[i:i+16])
|
|
|
|
enc.CryptBlocks(block, block)
|
|
|
|
}
|
|
|
|
|
|
|
|
var rsp *http.Response
|
|
|
|
var req *http.Request
|
|
|
|
ctr_aes.XORKeyStream(chunk, chunk)
|
|
|
|
chk_url := fmt.Sprintf("%s/%d", u.uploadUrl, chk_start)
|
|
|
|
|
|
|
|
chunk_resp := []byte{}
|
|
|
|
sleepTime := minSleepTime // inital backoff time
|
|
|
|
for retry := 0; retry < u.m.retries+1; retry++ {
|
|
|
|
reader := bytes.NewBuffer(chunk)
|
|
|
|
req, err = http.NewRequest("POST", chk_url, reader)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
rsp, err = u.m.client.Do(req)
|
|
|
|
if err == nil {
|
|
|
|
if rsp.StatusCode == 200 {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
err = errors.New("Http Status: " + rsp.Status)
|
|
|
|
_ = rsp.Body.Close()
|
|
|
|
}
|
|
|
|
u.m.debugf("%s: Retry upload chunk %d/%d: %v", u.name, retry, u.m.retries, err)
|
|
|
|
backOffSleep(&sleepTime)
|
|
|
|
}
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if rsp == nil {
|
|
|
|
return errors.New("retries exceeded")
|
|
|
|
}
|
|
|
|
|
|
|
|
chunk_resp, err = ioutil.ReadAll(rsp.Body)
|
|
|
|
if err != nil {
|
|
|
|
_ = rsp.Body.Close()
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
err = rsp.Body.Close()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
if bytes.Equal(chunk_resp, nil) == false {
|
|
|
|
u.mutex.Lock()
|
|
|
|
u.completion_handle = chunk_resp
|
|
|
|
u.mutex.Unlock()
|
|
|
|
}
|
|
|
|
|
|
|
|
// Update chunk MACs on success only
|
|
|
|
u.mutex.Lock()
|
|
|
|
if len(u.chunk_macs) > 0 {
|
|
|
|
u.chunk_macs[id] = make([]byte, 16)
|
|
|
|
copy(u.chunk_macs[id], block)
|
|
|
|
}
|
|
|
|
u.mutex.Unlock()
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Finish completes the upload and returns the created node
|
|
|
|
func (u *Upload) Finish() (node *Node, err error) {
|
|
|
|
mac_data := make([]byte, 16)
|
|
|
|
for _, v := range u.chunk_macs {
|
|
|
|
u.mac_enc.CryptBlocks(mac_data, v)
|
|
|
|
}
|
|
|
|
|
|
|
|
t := bytes_to_a32(mac_data)
|
|
|
|
meta_mac := []uint32{t[0] ^ t[1], t[2] ^ t[3]}
|
|
|
|
|
|
|
|
attr := FileAttr{u.name}
|
|
|
|
|
|
|
|
attr_data, err := encryptAttr(u.kbytes, attr)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
key := []uint32{u.ukey[0] ^ u.ukey[4], u.ukey[1] ^ u.ukey[5],
|
|
|
|
u.ukey[2] ^ meta_mac[0], u.ukey[3] ^ meta_mac[1],
|
|
|
|
u.ukey[4], u.ukey[5], meta_mac[0], meta_mac[1]}
|
|
|
|
|
|
|
|
buf := a32_to_bytes(key)
|
|
|
|
master_aes, err := aes.NewCipher(u.m.k)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
enc := cipher.NewCBCEncrypter(master_aes, zero_iv)
|
|
|
|
enc.CryptBlocks(buf[:16], buf[:16])
|
|
|
|
enc = cipher.NewCBCEncrypter(master_aes, zero_iv)
|
|
|
|
enc.CryptBlocks(buf[16:], buf[16:])
|
|
|
|
|
|
|
|
var cmsg [1]UploadCompleteMsg
|
|
|
|
var cres [1]UploadCompleteResp
|
|
|
|
|
|
|
|
cmsg[0].Cmd = "p"
|
|
|
|
cmsg[0].T = u.parenthash
|
|
|
|
cmsg[0].N[0].H = string(u.completion_handle)
|
|
|
|
cmsg[0].N[0].T = FILE
|
|
|
|
cmsg[0].N[0].A = string(attr_data)
|
|
|
|
cmsg[0].N[0].K = string(base64urlencode(buf))
|
|
|
|
|
|
|
|
request, err := json.Marshal(cmsg)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
result, err := u.m.api_request(request)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
err = json.Unmarshal(result, &cres)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
u.m.FS.mutex.Lock()
|
|
|
|
defer u.m.FS.mutex.Unlock()
|
|
|
|
return u.m.addFSNode(cres[0].F[0])
|
|
|
|
}
|
|
|
|
|
|
|
|
// Upload a file to the filesystem
|
|
|
|
func (m *Mega) UploadFile(srcpath string, parent *Node, name string, progress *chan int) (*Node, error) {
|
|
|
|
defer func() {
|
|
|
|
if progress != nil {
|
|
|
|
close(*progress)
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
var infile *os.File
|
|
|
|
var fileSize int64
|
|
|
|
|
|
|
|
info, err := os.Stat(srcpath)
|
|
|
|
if err == nil {
|
|
|
|
fileSize = info.Size()
|
|
|
|
}
|
|
|
|
|
|
|
|
infile, err = os.OpenFile(srcpath, os.O_RDONLY, 0666)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
if name == "" {
|
|
|
|
name = filepath.Base(srcpath)
|
|
|
|
}
|
|
|
|
|
|
|
|
u, err := m.NewUpload(parent, name, fileSize)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
workch := make(chan int)
|
|
|
|
errch := make(chan error, m.ul_workers)
|
|
|
|
wg := sync.WaitGroup{}
|
|
|
|
|
|
|
|
// Fire chunk upload workers
|
|
|
|
for w := 0; w < m.ul_workers; w++ {
|
|
|
|
wg.Add(1)
|
|
|
|
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
|
|
|
|
|
|
|
for id := range workch {
|
|
|
|
chk_start, chk_size, err := u.ChunkLocation(id)
|
|
|
|
if err != nil {
|
|
|
|
errch <- err
|
|
|
|
return
|
|
|
|
}
|
|
|
|
chunk := make([]byte, chk_size)
|
|
|
|
n, err := infile.ReadAt(chunk, chk_start)
|
|
|
|
if err != nil && err != io.EOF {
|
|
|
|
errch <- err
|
|
|
|
return
|
|
|
|
}
|
|
|
|
if n != len(chunk) {
|
|
|
|
errch <- errors.New("chunk too short")
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
err = u.UploadChunk(id, chunk)
|
|
|
|
if err != nil {
|
|
|
|
errch <- err
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
if progress != nil {
|
|
|
|
*progress <- chk_size
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
|
|
|
|
// Place chunk download jobs to chan
|
|
|
|
err = nil
|
|
|
|
for id := 0; id < u.Chunks() && err == nil; {
|
|
|
|
select {
|
|
|
|
case workch <- id:
|
|
|
|
id++
|
|
|
|
case err = <-errch:
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
close(workch)
|
|
|
|
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return u.Finish()
|
|
|
|
}
|
|
|
|
|
|
|
|
// Move a file from one location to another
|
|
|
|
func (m *Mega) Move(src *Node, parent *Node) error {
|
|
|
|
m.FS.mutex.Lock()
|
|
|
|
defer m.FS.mutex.Unlock()
|
|
|
|
|
|
|
|
if src == nil || parent == nil {
|
|
|
|
return EARGS
|
|
|
|
}
|
|
|
|
var msg [1]MoveFileMsg
|
|
|
|
var err error
|
|
|
|
|
|
|
|
msg[0].Cmd = "m"
|
|
|
|
msg[0].N = src.hash
|
|
|
|
msg[0].T = parent.hash
|
|
|
|
msg[0].I, err = randString(10)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
request, _ := json.Marshal(msg)
|
|
|
|
_, err = m.api_request(request)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
if src.parent != nil {
|
|
|
|
src.parent.removeChild(src)
|
|
|
|
}
|
|
|
|
|
|
|
|
parent.addChild(src)
|
|
|
|
src.parent = parent
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Rename a file or folder
|
|
|
|
func (m *Mega) Rename(src *Node, name string) error {
|
|
|
|
m.FS.mutex.Lock()
|
|
|
|
defer m.FS.mutex.Unlock()
|
|
|
|
|
|
|
|
if src == nil {
|
|
|
|
return EARGS
|
|
|
|
}
|
|
|
|
var msg [1]FileAttrMsg
|
|
|
|
|
|
|
|
master_aes, _ := aes.NewCipher(m.k)
|
|
|
|
attr := FileAttr{name}
|
|
|
|
attr_data, _ := encryptAttr(src.meta.key, attr)
|
|
|
|
key := make([]byte, len(src.meta.compkey))
|
|
|
|
err := blockEncrypt(master_aes, key, src.meta.compkey)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
msg[0].Cmd = "a"
|
|
|
|
msg[0].Attr = string(attr_data)
|
|
|
|
msg[0].Key = string(base64urlencode(key))
|
|
|
|
msg[0].N = src.hash
|
|
|
|
msg[0].I, err = randString(10)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
req, _ := json.Marshal(msg)
|
|
|
|
_, err = m.api_request(req)
|
|
|
|
|
|
|
|
src.name = name
|
|
|
|
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create a directory in the filesystem
|
|
|
|
func (m *Mega) CreateDir(name string, parent *Node) (*Node, error) {
|
|
|
|
m.FS.mutex.Lock()
|
|
|
|
defer m.FS.mutex.Unlock()
|
|
|
|
|
|
|
|
if parent == nil {
|
|
|
|
return nil, EARGS
|
|
|
|
}
|
|
|
|
var msg [1]UploadCompleteMsg
|
|
|
|
var res [1]UploadCompleteResp
|
|
|
|
|
|
|
|
compkey := []uint32{0, 0, 0, 0, 0, 0}
|
|
|
|
for i, _ := range compkey {
|
|
|
|
compkey[i] = uint32(mrand.Int31())
|
|
|
|
}
|
|
|
|
|
|
|
|
master_aes, _ := aes.NewCipher(m.k)
|
|
|
|
attr := FileAttr{name}
|
|
|
|
ukey := a32_to_bytes(compkey[:4])
|
|
|
|
attr_data, _ := encryptAttr(ukey, attr)
|
|
|
|
key := make([]byte, len(ukey))
|
|
|
|
err := blockEncrypt(master_aes, key, ukey)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
msg[0].Cmd = "p"
|
|
|
|
msg[0].T = parent.hash
|
|
|
|
msg[0].N[0].H = "xxxxxxxx"
|
|
|
|
msg[0].N[0].T = FOLDER
|
|
|
|
msg[0].N[0].A = string(attr_data)
|
|
|
|
msg[0].N[0].K = string(base64urlencode(key))
|
|
|
|
msg[0].I, err = randString(10)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
req, _ := json.Marshal(msg)
|
|
|
|
result, err := m.api_request(req)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
err = json.Unmarshal(result, &res)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
node, err := m.addFSNode(res[0].F[0])
|
|
|
|
|
|
|
|
return node, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Delete a file or directory from filesystem
|
|
|
|
func (m *Mega) Delete(node *Node, destroy bool) error {
|
|
|
|
if node == nil {
|
|
|
|
return EARGS
|
|
|
|
}
|
|
|
|
if destroy == false {
|
|
|
|
return m.Move(node, m.FS.trash)
|
|
|
|
}
|
|
|
|
|
|
|
|
m.FS.mutex.Lock()
|
|
|
|
defer m.FS.mutex.Unlock()
|
|
|
|
|
|
|
|
var msg [1]FileDeleteMsg
|
|
|
|
var err error
|
|
|
|
msg[0].Cmd = "d"
|
|
|
|
msg[0].N = node.hash
|
|
|
|
msg[0].I, err = randString(10)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
req, _ := json.Marshal(msg)
|
|
|
|
_, err = m.api_request(req)
|
|
|
|
|
|
|
|
parent := m.FS.lookup[node.hash]
|
|
|
|
parent.removeChild(node)
|
|
|
|
delete(m.FS.lookup, node.hash)
|
|
|
|
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// process an add node event
|
|
|
|
func (m *Mega) processAddNode(evRaw []byte) error {
|
|
|
|
m.FS.mutex.Lock()
|
|
|
|
defer m.FS.mutex.Unlock()
|
|
|
|
|
|
|
|
var ev FSEvent
|
|
|
|
err := json.Unmarshal(evRaw, &ev)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, itm := range ev.T.Files {
|
|
|
|
_, err = m.addFSNode(itm)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// process an update node event
|
|
|
|
func (m *Mega) processUpdateNode(evRaw []byte) error {
|
|
|
|
m.FS.mutex.Lock()
|
|
|
|
defer m.FS.mutex.Unlock()
|
|
|
|
|
|
|
|
var ev FSEvent
|
|
|
|
err := json.Unmarshal(evRaw, &ev)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
node := m.FS.hashLookup(ev.N)
|
|
|
|
attr, err := decryptAttr(node.meta.key, []byte(ev.Attr))
|
|
|
|
if err == nil {
|
|
|
|
node.name = attr.Name
|
|
|
|
} else {
|
|
|
|
node.name = "BAD ATTRIBUTE"
|
|
|
|
}
|
|
|
|
|
|
|
|
node.ts = time.Unix(ev.Ts, 0)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// process a delete node event
|
|
|
|
func (m *Mega) processDeleteNode(evRaw []byte) error {
|
|
|
|
m.FS.mutex.Lock()
|
|
|
|
defer m.FS.mutex.Unlock()
|
|
|
|
|
|
|
|
var ev FSEvent
|
|
|
|
err := json.Unmarshal(evRaw, &ev)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
node := m.FS.hashLookup(ev.N)
|
|
|
|
if node != nil && node.parent != nil {
|
|
|
|
node.parent.removeChild(node)
|
|
|
|
delete(m.FS.lookup, node.hash)
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Listen for server event notifications and play actions
|
|
|
|
func (m *Mega) pollEvents() {
|
|
|
|
var err error
|
|
|
|
var resp *http.Response
|
|
|
|
sleepTime := minSleepTime // inital backoff time
|
|
|
|
for {
|
|
|
|
if err != nil {
|
|
|
|
m.debugf("pollEvents: error from server", err)
|
|
|
|
backOffSleep(&sleepTime)
|
|
|
|
} else {
|
|
|
|
// reset sleep time to minimum on success
|
|
|
|
sleepTime = minSleepTime
|
|
|
|
}
|
|
|
|
|
|
|
|
url := fmt.Sprintf("%s/sc?sn=%s&sid=%s", m.baseurl, m.ssn, string(m.sid))
|
|
|
|
resp, err = m.client.Post(url, "application/xml", nil)
|
|
|
|
if err != nil {
|
|
|
|
m.logf("pollEvents: Error fetching status: %s", err)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
if resp.StatusCode != 200 {
|
|
|
|
m.logf("pollEvents: Error from server: %s", resp.Status)
|
|
|
|
_ = resp.Body.Close()
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
buf, err := ioutil.ReadAll(resp.Body)
|
|
|
|
if err != nil {
|
|
|
|
m.logf("pollEvents: Error reading body: %v", err)
|
|
|
|
_ = resp.Body.Close()
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
err = resp.Body.Close()
|
|
|
|
if err != nil {
|
|
|
|
m.logf("pollEvents: Error closing body: %v", err)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
// body is read and closed here
|
|
|
|
|
|
|
|
// First attempt to parse an array
|
|
|
|
var events Events
|
|
|
|
err = json.Unmarshal(buf, &events)
|
|
|
|
if err != nil {
|
|
|
|
// Try parsing as a lone error message
|
|
|
|
var emsg ErrorMsg
|
|
|
|
err = json.Unmarshal(buf, &emsg)
|
|
|
|
if err != nil {
|
|
|
|
m.logf("pollEvents: Bad response received from server: %s", buf)
|
|
|
|
} else {
|
|
|
|
err = parseError(emsg)
|
|
|
|
if err == EAGAIN {
|
|
|
|
} else if err != nil {
|
|
|
|
m.logf("pollEvents: Error received from server: %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
// if wait URL is set, then fetch it and continue - we
|
|
|
|
// don't expect anything else if we have a wait URL.
|
|
|
|
if events.W != "" {
|
2018-06-30 10:13:06 +00:00
|
|
|
m.waitEventsFire()
|
2018-04-13 10:51:28 +00:00
|
|
|
if len(events.E) > 0 {
|
|
|
|
m.logf("pollEvents: Unexpected event with w set: %s", buf)
|
|
|
|
}
|
|
|
|
resp, err = m.client.Get(events.W)
|
|
|
|
if err == nil {
|
|
|
|
_ = resp.Body.Close()
|
|
|
|
}
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
m.ssn = events.Sn
|
|
|
|
|
|
|
|
// For each event in the array, parse it
|
|
|
|
for _, evRaw := range events.E {
|
|
|
|
// First attempt to unmarshal as an error message
|
|
|
|
var emsg ErrorMsg
|
|
|
|
err = json.Unmarshal(evRaw, &emsg)
|
|
|
|
if err == nil {
|
|
|
|
m.logf("pollEvents: Error message received %s", evRaw)
|
|
|
|
err = parseError(emsg)
|
|
|
|
if err != nil {
|
|
|
|
m.logf("pollEvents: Event from server was error: %v", err)
|
|
|
|
}
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now unmarshal as a generic event
|
|
|
|
var gev GenericEvent
|
|
|
|
err = json.Unmarshal(evRaw, &gev)
|
|
|
|
if err != nil {
|
|
|
|
m.logf("pollEvents: Couldn't parse event from server: %v: %s", err, evRaw)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
m.debugf("pollEvents: Parsing event %q: %s", gev.Cmd, evRaw)
|
|
|
|
|
|
|
|
// Work out what to do with the event
|
|
|
|
var process func([]byte) error
|
|
|
|
switch gev.Cmd {
|
|
|
|
case "t": // node addition
|
|
|
|
process = m.processAddNode
|
|
|
|
case "u": // node update
|
|
|
|
process = m.processUpdateNode
|
|
|
|
case "d": // node deletion
|
|
|
|
process = m.processDeleteNode
|
|
|
|
case "s", "s2": // share addition/update/revocation
|
|
|
|
case "c": // contact addition/update
|
|
|
|
case "k": // crypto key request
|
|
|
|
case "fa": // file attribute update
|
|
|
|
case "ua": // user attribute update
|
|
|
|
case "psts": // account updated
|
|
|
|
case "ipc": // incoming pending contact request (to us)
|
|
|
|
case "opc": // outgoing pending contact request (from us)
|
|
|
|
case "upci": // incoming pending contact request update (accept/deny/ignore)
|
|
|
|
case "upco": // outgoing pending contact request update (from them, accept/deny/ignore)
|
|
|
|
case "ph": // public links handles
|
|
|
|
case "se": // set email
|
|
|
|
case "mcc": // chat creation / peer's invitation / peer's removal
|
|
|
|
case "mcna": // granted / revoked access to a node
|
|
|
|
case "uac": // user access control
|
|
|
|
default:
|
|
|
|
m.debugf("pollEvents: Unknown message %q received: %s", gev.Cmd, evRaw)
|
|
|
|
}
|
|
|
|
|
|
|
|
// process the event if we can
|
|
|
|
if process != nil {
|
|
|
|
err := process(evRaw)
|
|
|
|
if err != nil {
|
|
|
|
m.logf("pollEvents: Error processing event %q '%s': %v", gev.Cmd, evRaw, err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (m *Mega) getLink(n *Node) (string, error) {
|
|
|
|
var msg [1]GetLinkMsg
|
|
|
|
var res [1]string
|
|
|
|
|
|
|
|
msg[0].Cmd = "l"
|
|
|
|
msg[0].N = n.GetHash()
|
|
|
|
|
|
|
|
req, _ := json.Marshal(msg)
|
|
|
|
result, err := m.api_request(req)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
return "", err
|
|
|
|
}
|
|
|
|
err = json.Unmarshal(result, &res)
|
|
|
|
if err != nil {
|
|
|
|
return "", err
|
|
|
|
}
|
|
|
|
return res[0], nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Exports public link for node, with or without decryption key included
|
|
|
|
func (m *Mega) Link(n *Node, includeKey bool) (string, error) {
|
|
|
|
id, err := m.getLink(n)
|
|
|
|
if err != nil {
|
|
|
|
return "", err
|
|
|
|
}
|
|
|
|
if includeKey {
|
|
|
|
m.FS.mutex.Lock()
|
|
|
|
key := string(base64urlencode(n.meta.compkey))
|
|
|
|
m.FS.mutex.Unlock()
|
|
|
|
return fmt.Sprintf("%v/#!%v!%v", BASE_DOWNLOAD_URL, id, key), nil
|
|
|
|
} else {
|
|
|
|
return fmt.Sprintf("%v/#!%v", BASE_DOWNLOAD_URL, id), nil
|
|
|
|
}
|
|
|
|
}
|