forked from TrueCloudLab/frostfs-s3-gw
d19ce03072
I'm not sure it works, but it's enough code-wise for now. We're reusing some http-gw components here that are to be moved into sdk-go in future. Signed-off-by: Roman Khimov <roman@nspcc.ru>
393 lines
10 KiB
Go
393 lines
10 KiB
Go
package layer
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
"net/url"
|
|
"time"
|
|
|
|
"github.com/nspcc-dev/neofs-api-go/pkg/client"
|
|
"github.com/nspcc-dev/neofs-api-go/pkg/container"
|
|
"github.com/nspcc-dev/neofs-api-go/pkg/object"
|
|
"github.com/nspcc-dev/neofs-api-go/pkg/owner"
|
|
"github.com/nspcc-dev/neofs-api-go/pkg/token"
|
|
"github.com/nspcc-dev/neofs-http-gw/connections"
|
|
sdk "github.com/nspcc-dev/neofs-http-gw/neofs"
|
|
"github.com/nspcc-dev/neofs-s3-gw/api"
|
|
"github.com/nspcc-dev/neofs-s3-gw/creds/neofs"
|
|
"go.uber.org/zap"
|
|
"google.golang.org/grpc/codes"
|
|
"google.golang.org/grpc/status"
|
|
)
|
|
|
|
type (
|
|
layer struct {
|
|
cli sdk.ClientPlant
|
|
log *zap.Logger
|
|
}
|
|
|
|
// Params stores basic API parameters.
|
|
Params struct {
|
|
Pool connections.Pool
|
|
Logger *zap.Logger
|
|
Timeout time.Duration
|
|
Credential neofs.Credentials
|
|
}
|
|
|
|
// GetObjectParams stores object get request parameters.
|
|
GetObjectParams struct {
|
|
Bucket string
|
|
Object string
|
|
Offset int64
|
|
Length int64
|
|
Writer io.Writer
|
|
}
|
|
|
|
// PutObjectParams stores object put request parameters.
|
|
PutObjectParams struct {
|
|
Bucket string
|
|
Object string
|
|
Size int64
|
|
Reader io.Reader
|
|
Header map[string]string
|
|
}
|
|
|
|
// CopyObjectParams stores object copy request parameters.
|
|
CopyObjectParams struct {
|
|
SrcBucket string
|
|
DstBucket string
|
|
SrcObject string
|
|
DstObject string
|
|
Header map[string]string
|
|
}
|
|
|
|
// NeoFS provides basic NeoFS interface.
|
|
NeoFS interface {
|
|
Get(ctx context.Context, address *object.Address) (*object.Object, error)
|
|
}
|
|
|
|
// Client provides S3 API client interface.
|
|
Client interface {
|
|
NeoFS
|
|
|
|
ListBuckets(ctx context.Context) ([]*BucketInfo, error)
|
|
GetBucketInfo(ctx context.Context, name string) (*BucketInfo, error)
|
|
|
|
GetObject(ctx context.Context, p *GetObjectParams) error
|
|
GetObjectInfo(ctx context.Context, bucketName, objectName string) (*ObjectInfo, error)
|
|
|
|
PutObject(ctx context.Context, p *PutObjectParams) (*ObjectInfo, error)
|
|
|
|
CopyObject(ctx context.Context, p *CopyObjectParams) (*ObjectInfo, error)
|
|
|
|
ListObjects(ctx context.Context, p *ListObjectsParams) (*ListObjectsInfo, error)
|
|
|
|
DeleteObject(ctx context.Context, bucket, object string) error
|
|
DeleteObjects(ctx context.Context, bucket string, objects []string) []error
|
|
}
|
|
)
|
|
|
|
var (
|
|
// ErrObjectExists is returned on attempts to create already existing object.
|
|
ErrObjectExists = errors.New("object exists")
|
|
// ErrObjectNotExists is returned on attempts to work with non-existing object.
|
|
ErrObjectNotExists = errors.New("object not exists")
|
|
)
|
|
|
|
// NewLayer creates instance of layer. It checks credentials
|
|
// and establishes gRPC connection with node.
|
|
func NewLayer(log *zap.Logger, cli sdk.ClientPlant) Client {
|
|
return &layer{
|
|
cli: cli,
|
|
log: log,
|
|
}
|
|
}
|
|
|
|
// Owner returns owner id from BearerToken (context) or from client owner.
|
|
func (n *layer) Owner(ctx context.Context) *owner.ID {
|
|
if tkn, ok := ctx.Value(api.BearerTokenKey).(*token.BearerToken); ok && tkn != nil {
|
|
return tkn.Issuer()
|
|
}
|
|
|
|
return n.cli.OwnerID()
|
|
}
|
|
|
|
// Get NeoFS Object by refs.Address (should be used by auth.Center).
|
|
func (n *layer) Get(ctx context.Context, address *object.Address) (*object.Object, error) {
|
|
conn, tok, err := n.cli.ConnectionArtifacts()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
ops := new(client.GetObjectParams).WithAddress(address)
|
|
return conn.GetObject(ctx, ops, client.WithSession(tok))
|
|
}
|
|
|
|
// GetBucketInfo returns bucket info by name.
|
|
func (n *layer) GetBucketInfo(ctx context.Context, name string) (*BucketInfo, error) {
|
|
name, err := url.QueryUnescape(name)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
list, err := n.containerList(ctx)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
for _, bkt := range list {
|
|
if bkt.Name == name {
|
|
return bkt, nil
|
|
}
|
|
}
|
|
|
|
return nil, status.Error(codes.NotFound, "bucket not found")
|
|
}
|
|
|
|
// ListBuckets returns all user containers. Name of the bucket is a container
|
|
// id. Timestamp is omitted since it is not saved in neofs container.
|
|
func (n *layer) ListBuckets(ctx context.Context) ([]*BucketInfo, error) {
|
|
return n.containerList(ctx)
|
|
}
|
|
|
|
// ListObjects returns objects from the container. It ignores tombstones and
|
|
// storage groups.
|
|
// ctx, bucket, prefix, continuationToken, delimiter, maxKeys
|
|
func (n *layer) ListObjects(ctx context.Context, p *ListObjectsParams) (*ListObjectsInfo, error) {
|
|
// todo: make pagination when search response will be gRPC stream,
|
|
// pagination must be implemented with cache, because search results
|
|
// may be different between search calls
|
|
var (
|
|
err error
|
|
bkt *BucketInfo
|
|
ids []*object.ID
|
|
result ListObjectsInfo
|
|
uniqNames = make(map[string]struct{})
|
|
)
|
|
|
|
if bkt, err = n.GetBucketInfo(ctx, p.Bucket); err != nil {
|
|
return nil, err
|
|
} else if ids, err = n.objectSearch(ctx, &findParams{cid: bkt.CID}); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
ln := len(ids)
|
|
// todo: check what happens if there is more than maxKeys objects
|
|
if ln > p.MaxKeys {
|
|
ln = p.MaxKeys
|
|
result.IsTruncated = true
|
|
}
|
|
|
|
result.Objects = make([]*ObjectInfo, 0, ln)
|
|
|
|
for _, id := range ids {
|
|
addr := object.NewAddress()
|
|
addr.SetObjectID(id)
|
|
addr.SetContainerID(bkt.CID)
|
|
|
|
meta, err := n.objectHead(ctx, addr)
|
|
if err != nil {
|
|
n.log.Warn("could not fetch object meta", zap.Error(err))
|
|
continue
|
|
}
|
|
|
|
// // ignore tombstone objects
|
|
// _, hdr := meta.LastHeader(object.HeaderType(object.TombstoneHdr))
|
|
// if hdr != nil {
|
|
// continue
|
|
// }
|
|
|
|
// ignore storage group objects
|
|
// _, hdr = meta.LastHeader(object.HeaderType(object.StorageGroupHdr))
|
|
// if hdr != nil {
|
|
// continue
|
|
// }
|
|
|
|
// dirs don't exist in neofs, gateway stores full path to the file
|
|
// in object header, e.g. `filename`:`/this/is/path/file.txt`
|
|
|
|
// prefix argument contains full dir path from the root, e.g. `/this/is/`
|
|
|
|
// to emulate dirs we take dirs in path, compare it with prefix
|
|
// and look for entities after prefix. If entity does not have any
|
|
// sub-entities, then it is a file, else directory.
|
|
|
|
if oi := objectInfoFromMeta(bkt, meta, p.Prefix); oi != nil {
|
|
// use only unique dir names
|
|
if _, ok := uniqNames[oi.Name]; !ok {
|
|
uniqNames[oi.Name] = struct{}{}
|
|
|
|
result.Objects = append(result.Objects, oi)
|
|
}
|
|
}
|
|
}
|
|
|
|
return &result, nil
|
|
}
|
|
|
|
// GetObject from storage.
|
|
func (n *layer) GetObject(ctx context.Context, p *GetObjectParams) error {
|
|
var (
|
|
err error
|
|
oid *object.ID
|
|
bkt *BucketInfo
|
|
)
|
|
|
|
if bkt, err = n.GetBucketInfo(ctx, p.Bucket); err != nil {
|
|
return fmt.Errorf("couldn't find bucket: %s : %w", p.Bucket, err)
|
|
} else if oid, err = n.objectFindID(ctx, &findParams{cid: bkt.CID, val: p.Object}); err != nil {
|
|
return fmt.Errorf("search of the object failed: cid: %s, val: %s : %w", bkt.CID, p.Object, err)
|
|
}
|
|
|
|
addr := object.NewAddress()
|
|
addr.SetObjectID(oid)
|
|
addr.SetContainerID(bkt.CID)
|
|
|
|
_, err = n.objectGet(ctx, &getParams{
|
|
Writer: p.Writer,
|
|
|
|
address: addr,
|
|
|
|
offset: p.Offset,
|
|
length: p.Length,
|
|
})
|
|
|
|
if err != nil {
|
|
return fmt.Errorf("couldn't get object, cid: %s : %w", bkt.CID, err)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (n *layer) checkObject(ctx context.Context, cid *container.ID, filename string) error {
|
|
var err error
|
|
|
|
if _, err = n.objectFindID(ctx, &findParams{cid: cid, val: filename}); err == nil {
|
|
return ErrObjectExists
|
|
} else if state, ok := status.FromError(err); !ok || state == nil {
|
|
return err
|
|
} else if state.Code() == codes.NotFound {
|
|
return ErrObjectNotExists
|
|
}
|
|
|
|
return err
|
|
}
|
|
|
|
// GetObjectInfo returns meta information about the object.
|
|
func (n *layer) GetObjectInfo(ctx context.Context, bucketName, filename string) (*ObjectInfo, error) {
|
|
var (
|
|
err error
|
|
oid *object.ID
|
|
bkt *BucketInfo
|
|
meta *object.Object
|
|
)
|
|
|
|
if bkt, err = n.GetBucketInfo(ctx, bucketName); err != nil {
|
|
n.log.Error("could not fetch bucket info", zap.Error(err))
|
|
return nil, err
|
|
} else if oid, err = n.objectFindID(ctx, &findParams{cid: bkt.CID, val: filename}); err != nil {
|
|
n.log.Error("could not find object id", zap.Error(err))
|
|
return nil, err
|
|
}
|
|
|
|
addr := object.NewAddress()
|
|
addr.SetObjectID(oid)
|
|
addr.SetContainerID(bkt.CID)
|
|
|
|
if meta, err = n.objectHead(ctx, addr); err != nil {
|
|
n.log.Error("could not fetch object head", zap.Error(err))
|
|
return nil, err
|
|
}
|
|
|
|
return objectInfoFromMeta(bkt, meta, rootSeparator), nil
|
|
}
|
|
|
|
// PutObject into storage.
|
|
func (n *layer) PutObject(ctx context.Context, p *PutObjectParams) (*ObjectInfo, error) {
|
|
return n.objectPut(ctx, p)
|
|
}
|
|
|
|
// CopyObject from one bucket into another bucket.
|
|
func (n *layer) CopyObject(ctx context.Context, p *CopyObjectParams) (*ObjectInfo, error) {
|
|
info, err := n.GetObjectInfo(ctx, p.SrcBucket, p.SrcObject)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("couldn't get object info: %w", err)
|
|
}
|
|
|
|
pr, pw := io.Pipe()
|
|
|
|
go func() {
|
|
err := n.GetObject(ctx, &GetObjectParams{
|
|
Bucket: p.SrcBucket,
|
|
Object: p.SrcObject,
|
|
Writer: pw,
|
|
})
|
|
|
|
if err = pw.CloseWithError(err); err != nil {
|
|
n.log.Error("could not get object", zap.Error(err))
|
|
}
|
|
}()
|
|
|
|
// set custom headers
|
|
for k, v := range p.Header {
|
|
info.Headers[k] = v
|
|
}
|
|
|
|
return n.PutObject(ctx, &PutObjectParams{
|
|
Bucket: p.DstBucket,
|
|
Object: p.DstObject,
|
|
Size: info.Size,
|
|
Reader: pr,
|
|
Header: info.Headers,
|
|
})
|
|
}
|
|
|
|
// DeleteObject removes all objects with passed nice name.
|
|
func (n *layer) DeleteObject(ctx context.Context, bucket, filename string) error {
|
|
var (
|
|
err error
|
|
ids []*object.ID
|
|
bkt *BucketInfo
|
|
)
|
|
|
|
if bkt, err = n.GetBucketInfo(ctx, bucket); err != nil {
|
|
return &api.DeleteError{
|
|
Err: err,
|
|
Object: filename,
|
|
}
|
|
} else if ids, err = n.objectSearch(ctx, &findParams{cid: bkt.CID, val: filename}); err != nil {
|
|
return &api.DeleteError{
|
|
Err: err,
|
|
Object: filename,
|
|
}
|
|
}
|
|
|
|
for _, id := range ids {
|
|
addr := object.NewAddress()
|
|
addr.SetObjectID(id)
|
|
addr.SetContainerID(bkt.CID)
|
|
|
|
if err = n.objectDelete(ctx, addr); err != nil {
|
|
return &api.DeleteError{
|
|
Err: err,
|
|
Object: filename,
|
|
}
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// DeleteObjects from the storage.
|
|
func (n *layer) DeleteObjects(ctx context.Context, bucket string, objects []string) []error {
|
|
var errs = make([]error, 0, len(objects))
|
|
|
|
for i := range objects {
|
|
if err := n.DeleteObject(ctx, bucket, objects[i]); err != nil {
|
|
errs = append(errs, err)
|
|
}
|
|
}
|
|
|
|
return errs
|
|
}
|