mirror of
https://github.com/ceph/s3-tests.git
synced 2024-11-21 11:51:06 +00:00
Initial commit.
This commit is contained in:
commit
46f2e3b00e
6 changed files with 825 additions and 0 deletions
11
.gitignore
vendored
Normal file
11
.gitignore
vendored
Normal file
|
@ -0,0 +1,11 @@
|
|||
*~
|
||||
.#*
|
||||
## the next line needs to start with a backslash to avoid looking like
|
||||
## a comment
|
||||
\#*#
|
||||
.*.swp
|
||||
|
||||
*.pyc
|
||||
*.pyo
|
||||
|
||||
/virtualenv
|
91
README.rst
Normal file
91
README.rst
Normal file
|
@ -0,0 +1,91 @@
|
|||
========================
|
||||
S3 compatibility tests
|
||||
========================
|
||||
|
||||
This is a set of completely unofficial Amazon AWS S3 compatibility
|
||||
tests, that will hopefully be useful to people implementing software
|
||||
that exposes an S3-like API.
|
||||
|
||||
The tests only cover the REST interface.
|
||||
|
||||
TODO: test direct HTTP downloads, like a web browser would do.
|
||||
|
||||
The tests use the Boto library, so any e.g. HTTP-level differences
|
||||
that Boto papers over, the tests will not be able to discover. Raw
|
||||
HTTP tests may be added later.
|
||||
|
||||
The tests use the Nose test framework. To get started, ensure you have
|
||||
the ``virtualenv`` software installed; e.g. on Debian/Ubuntu::
|
||||
|
||||
sudo apt-get install python-virtualenv
|
||||
|
||||
and then run::
|
||||
|
||||
./bootstrap
|
||||
|
||||
You will need to create a configuration file with the location of the
|
||||
service and two different credentials, something like this::
|
||||
|
||||
[DEFAULT]
|
||||
## this section is just used as default for all the "s3 *"
|
||||
## sections, you can place these variables also directly there
|
||||
|
||||
## replace with e.g. "localhost" to run against local software
|
||||
host = s3.amazonaws.com
|
||||
|
||||
## uncomment the port to use something other than 80
|
||||
# port = 8080
|
||||
|
||||
## say "no" to disable TLS
|
||||
is_secure = yes
|
||||
|
||||
[fixtures]
|
||||
## all the buckets created will start with this prefix;
|
||||
## {random} will be filled with random characters to pad
|
||||
## the prefix to 30 characters long, and avoid collisions
|
||||
bucket prefix = YOURNAMEHERE-{random}-
|
||||
|
||||
[s3 main]
|
||||
## the tests assume two accounts are defined, "main" and "alt".
|
||||
|
||||
## user_id is a 64-character hexstring
|
||||
user_id = 0123456789abcdef0123456789abcdef0123456789abcdef0123456789abcdef
|
||||
|
||||
## display name typically looks more like a unix login, "jdoe" etc
|
||||
display_name = youruseridhere
|
||||
|
||||
## replace these with your access keys
|
||||
access_key = ABCDEFGHIJKLMNOPQRST
|
||||
secret_key = abcdefghijklmnopqrstuvwxyzabcdefghijklmn
|
||||
|
||||
[s3 alt]
|
||||
## another user account, used for ACL-related tests
|
||||
user_id = 56789abcdef0123456789abcdef0123456789abcdef0123456789abcdef01234
|
||||
display_name = john.doe
|
||||
## the "alt" user needs to have email set, too
|
||||
email = john.doe@example.com
|
||||
access_key = NOPQRSTUVWXYZABCDEFG
|
||||
secret_key = nopqrstuvwxyzabcdefghijklmnabcdefghijklm
|
||||
|
||||
Once you have that, you can run the tests with::
|
||||
|
||||
S3TEST_CONF=your.conf ./virtualenv/bin/nosetests
|
||||
|
||||
You can specify what test(s) to run::
|
||||
|
||||
S3TEST_CONF=your.conf ./virtualenv/bin/nosetests test_s3:test_object_acl_grant_public_read
|
||||
|
||||
Some tests have attributes set based on their current reliability and
|
||||
things like AWS not enforcing their spec stricly. You can filter tests
|
||||
based on their attributes::
|
||||
|
||||
S3TEST_CONF=aws.conf ./virtualenv/bin/nosetests -a '!fails_on_aws'
|
||||
|
||||
|
||||
TODO
|
||||
====
|
||||
|
||||
- We should assume read-after-write consistency, and make the tests
|
||||
actually request such a location.
|
||||
|
||||
http://aws.amazon.com/s3/faqs/#What_data_consistency_model_does_Amazon_S3_employ
|
4
bootstrap
Executable file
4
bootstrap
Executable file
|
@ -0,0 +1,4 @@
|
|||
#!/bin/sh
|
||||
set -e
|
||||
virtualenv --no-site-packages --distribute virtualenv
|
||||
./virtualenv/bin/pip install -r requirements.txt
|
3
requirements.txt
Normal file
3
requirements.txt
Normal file
|
@ -0,0 +1,3 @@
|
|||
nose >=1.0.0
|
||||
boto >=2.0b4
|
||||
bunch >=1.0.0
|
702
test_s3.py
Normal file
702
test_s3.py
Normal file
|
@ -0,0 +1,702 @@
|
|||
import ConfigParser
|
||||
import boto.exception
|
||||
import boto.s3.connection
|
||||
import bunch
|
||||
import itertools
|
||||
import os
|
||||
import random
|
||||
import string
|
||||
import time
|
||||
|
||||
from nose.tools import eq_ as eq
|
||||
from nose.plugins.attrib import attr
|
||||
|
||||
from utils import assert_raises
|
||||
|
||||
NONEXISTENT_EMAIL = 'doesnotexist@dreamhost.com.invalid'
|
||||
|
||||
s3 = bunch.Bunch()
|
||||
config = bunch.Bunch()
|
||||
|
||||
# this will be assigned by setup()
|
||||
prefix = None
|
||||
|
||||
|
||||
def choose_bucket_prefix(template, max_len=30):
|
||||
"""
|
||||
Choose a prefix for our test buckets, so they're easy to identify.
|
||||
|
||||
Use template and feed it more and more random filler, until it's
|
||||
as long as possible but still below max_len.
|
||||
"""
|
||||
rand = ''.join(
|
||||
random.choice(string.ascii_lowercase + string.digits)
|
||||
for c in range(255)
|
||||
)
|
||||
|
||||
while rand:
|
||||
s = template.format(random=rand)
|
||||
if len(s) <= max_len:
|
||||
return s
|
||||
rand = rand[:-1]
|
||||
|
||||
raise RuntimeError(
|
||||
'Bucket prefix template is impossible to fulfill: {template!r}'.format(
|
||||
template=template,
|
||||
),
|
||||
)
|
||||
|
||||
|
||||
def nuke_prefixed_buckets():
|
||||
for name, conn in s3.items():
|
||||
print 'Cleaning buckets from connection {name}'.format(name=name)
|
||||
for bucket in conn.get_all_buckets():
|
||||
if bucket.name.startswith(prefix):
|
||||
print 'Cleaning bucket {bucket}'.format(bucket=bucket)
|
||||
try:
|
||||
for key in bucket.list():
|
||||
print 'Cleaning bucket {bucket} key {key}'.format(
|
||||
bucket=bucket,
|
||||
key=key,
|
||||
)
|
||||
key.delete()
|
||||
bucket.delete()
|
||||
except boto.exception.S3ResponseError as e:
|
||||
# TODO workaround for buggy rgw that fails to send
|
||||
# error_code, remove
|
||||
if (e.status == 403
|
||||
and e.error_code is None
|
||||
and e.body == ''):
|
||||
e.error_code = 'AccessDenied'
|
||||
if e.error_code != 'AccessDenied':
|
||||
print 'GOT UNWANTED ERROR', e.error_code
|
||||
raise
|
||||
# seems like we're not the owner of the bucket; ignore
|
||||
pass
|
||||
|
||||
print 'Done with cleanup of test buckets.'
|
||||
|
||||
|
||||
def setup():
|
||||
|
||||
cfg = ConfigParser.RawConfigParser()
|
||||
try:
|
||||
path = os.environ['S3TEST_CONF']
|
||||
except KeyError:
|
||||
raise RuntimeError(
|
||||
'To run tests, point environment '
|
||||
+ 'variable S3TEST_CONF to a config file.',
|
||||
)
|
||||
with file(path) as f:
|
||||
cfg.readfp(f)
|
||||
|
||||
global prefix
|
||||
try:
|
||||
template = cfg.get('fixtures', 'bucket prefix')
|
||||
except (ConfigParser.NoSectionError, ConfigParser.NoOptionError):
|
||||
template = 'test-{random}-'
|
||||
prefix = choose_bucket_prefix(template=template)
|
||||
|
||||
s3.clear()
|
||||
config.clear()
|
||||
for section in cfg.sections():
|
||||
try:
|
||||
(type_, name) = section.split(None, 1)
|
||||
except ValueError:
|
||||
continue
|
||||
if type_ != 's3':
|
||||
continue
|
||||
try:
|
||||
port = cfg.getint(section, 'port')
|
||||
except ConfigParser.NoOptionError:
|
||||
port = None
|
||||
|
||||
config[name] = bunch.Bunch()
|
||||
for var in [
|
||||
'user_id',
|
||||
'display_name',
|
||||
'email',
|
||||
]:
|
||||
try:
|
||||
config[name][var] = cfg.get(section, var)
|
||||
except ConfigParser.NoOptionError:
|
||||
pass
|
||||
conn = boto.s3.connection.S3Connection(
|
||||
aws_access_key_id=cfg.get(section, 'access_key'),
|
||||
aws_secret_access_key=cfg.get(section, 'secret_key'),
|
||||
is_secure=cfg.getboolean(section, 'is_secure'),
|
||||
port=port,
|
||||
host=cfg.get(section, 'host'),
|
||||
# TODO support & test all variations
|
||||
calling_format=boto.s3.connection.OrdinaryCallingFormat(),
|
||||
)
|
||||
s3[name] = conn
|
||||
|
||||
# WARNING! we actively delete all buckets we see with the prefix
|
||||
# we've chosen! Choose your prefix with care, and don't reuse
|
||||
# credentials!
|
||||
|
||||
# We also assume nobody else is going to use buckets with that
|
||||
# prefix. This is racy but given enough randomness, should not
|
||||
# really fail.
|
||||
nuke_prefixed_buckets()
|
||||
|
||||
|
||||
def teardown():
|
||||
# remove our buckets here also, to avoid littering
|
||||
nuke_prefixed_buckets()
|
||||
|
||||
|
||||
bucket_counter = itertools.count(1)
|
||||
|
||||
|
||||
def get_new_bucket(connection=None):
|
||||
"""
|
||||
Get a bucket that exists and is empty.
|
||||
|
||||
Always recreates a bucket from scratch. This is useful to also
|
||||
reset ACLs and such.
|
||||
"""
|
||||
if connection is None:
|
||||
connection = s3.main
|
||||
name = '{prefix}{num}'.format(
|
||||
prefix=prefix,
|
||||
num=next(bucket_counter),
|
||||
)
|
||||
# the only way for this to fail with a pre-existing bucket is if
|
||||
# someone raced us between setup nuke_prefixed_buckets and here;
|
||||
# ignore that as astronomically unlikely
|
||||
bucket = connection.create_bucket(name)
|
||||
return bucket
|
||||
|
||||
|
||||
def check_access_denied(fn, *args, **kwargs):
|
||||
e = assert_raises(boto.exception.S3ResponseError, fn, *args, **kwargs)
|
||||
eq(e.status, 403)
|
||||
eq(e.reason, 'Forbidden')
|
||||
eq(e.error_code, 'AccessDenied')
|
||||
|
||||
|
||||
def test_bucket_list_empty():
|
||||
bucket = get_new_bucket()
|
||||
l = bucket.list()
|
||||
l = list(l)
|
||||
eq(l, [])
|
||||
|
||||
|
||||
# TODO rgw gives NoSuchKey instead of NoSuchBucket
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_create_delete():
|
||||
name = '{prefix}foo'.format(prefix=prefix)
|
||||
print 'Trying bucket {name!r}'.format(name=name)
|
||||
bucket = s3.main.create_bucket(name)
|
||||
# make sure it's actually there
|
||||
s3.main.get_bucket(bucket.name)
|
||||
bucket.delete()
|
||||
# make sure it's gone
|
||||
e = assert_raises(boto.exception.S3ResponseError, bucket.delete)
|
||||
eq(e.status, 404)
|
||||
eq(e.reason, 'Not Found')
|
||||
eq(e.error_code, 'NoSuchBucket')
|
||||
|
||||
|
||||
def test_object_read_notexist():
|
||||
bucket = get_new_bucket()
|
||||
key = bucket.new_key('foobar')
|
||||
e = assert_raises(boto.exception.S3ResponseError, key.get_contents_as_string)
|
||||
eq(e.status, 404)
|
||||
eq(e.reason, 'Not Found')
|
||||
eq(e.error_code, 'NoSuchKey')
|
||||
|
||||
|
||||
def test_object_write_then_read():
|
||||
bucket = get_new_bucket()
|
||||
key = bucket.new_key('foo')
|
||||
key.set_contents_from_string('bar')
|
||||
got = key.get_contents_as_string()
|
||||
eq(got, 'bar')
|
||||
|
||||
|
||||
def check_bad_bucket_name(name):
|
||||
e = assert_raises(boto.exception.S3ResponseError, s3.main.create_bucket, name)
|
||||
eq(e.status, 400)
|
||||
eq(e.reason, 'Bad Request')
|
||||
eq(e.error_code, 'InvalidBucketName')
|
||||
|
||||
|
||||
# AWS does not enforce all documented bucket restrictions.
|
||||
# http://docs.amazonwebservices.com/AmazonS3/2006-03-01/dev/index.html?BucketRestrictions.html
|
||||
@attr('fails_on_aws')
|
||||
# TODO rgw fails to provide error_code
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_create_naming_bad_starts_nonalpha():
|
||||
check_bad_bucket_name('_alphasoup')
|
||||
|
||||
|
||||
# TODO this seems to hang until timeout on rgw?
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_create_naming_bad_short_empty():
|
||||
# bucket creates where name is empty look like PUTs to the parent
|
||||
# resource (with slash), hence their error response is different
|
||||
e = assert_raises(boto.exception.S3ResponseError, s3.main.create_bucket, '')
|
||||
eq(e.status, 405)
|
||||
eq(e.reason, 'Method Not Allowed')
|
||||
eq(e.error_code, 'MethodNotAllowed')
|
||||
|
||||
|
||||
# TODO rgw fails to provide error_code
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_create_naming_bad_short_one():
|
||||
check_bad_bucket_name('a')
|
||||
|
||||
|
||||
# TODO rgw fails to provide error_code
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_create_naming_bad_short_two():
|
||||
check_bad_bucket_name('aa')
|
||||
|
||||
|
||||
# TODO rgw fails to provide error_code
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_create_naming_bad_long():
|
||||
check_bad_bucket_name(256*'a')
|
||||
check_bad_bucket_name(280*'a')
|
||||
check_bad_bucket_name(3000*'a')
|
||||
|
||||
|
||||
def check_good_bucket_name(name):
|
||||
# prefixing to make then unique; tests using this must *not* rely
|
||||
# on being able to set the initial character, or exceed the max
|
||||
# len
|
||||
s3.main.create_bucket('{prefix}{name}'.format(
|
||||
prefix=prefix,
|
||||
name=name,
|
||||
))
|
||||
|
||||
|
||||
def _test_bucket_create_naming_good_long(length):
|
||||
assert len(prefix) < 255
|
||||
num = length - len(prefix)
|
||||
s3.main.create_bucket('{prefix}{name}'.format(
|
||||
prefix=prefix,
|
||||
name=num*'a',
|
||||
))
|
||||
|
||||
|
||||
def test_bucket_create_naming_good_long_250():
|
||||
_test_bucket_create_naming_good_long(250)
|
||||
|
||||
|
||||
# breaks nuke_prefixed_buckets in teardown, claims a bucket from
|
||||
# conn.get_all_buckets() suddenly does not exist
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_create_naming_good_long_251():
|
||||
_test_bucket_create_naming_good_long(251)
|
||||
|
||||
|
||||
# breaks nuke_prefixed_buckets in teardown, claims a bucket from
|
||||
# conn.get_all_buckets() suddenly does not exist
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_create_naming_good_long_252():
|
||||
_test_bucket_create_naming_good_long(252)
|
||||
|
||||
|
||||
# breaks nuke_prefixed_buckets in teardown, claims a bucket from
|
||||
# conn.get_all_buckets() suddenly does not exist
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_create_naming_good_long_253():
|
||||
_test_bucket_create_naming_good_long(253)
|
||||
|
||||
|
||||
# breaks nuke_prefixed_buckets in teardown, claims a bucket from
|
||||
# conn.get_all_buckets() suddenly does not exist
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_create_naming_good_long_254():
|
||||
_test_bucket_create_naming_good_long(254)
|
||||
|
||||
|
||||
# breaks nuke_prefixed_buckets in teardown, claims a bucket from
|
||||
# conn.get_all_buckets() suddenly does not exist
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_create_naming_good_long_255():
|
||||
_test_bucket_create_naming_good_long(255)
|
||||
|
||||
|
||||
# AWS does not enforce all documented bucket restrictions.
|
||||
# http://docs.amazonwebservices.com/AmazonS3/2006-03-01/dev/index.html?BucketRestrictions.html
|
||||
@attr('fails_on_aws')
|
||||
# TODO rgw fails to provide error_code
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_create_naming_bad_ip():
|
||||
check_bad_bucket_name('192.168.5.123')
|
||||
|
||||
|
||||
# TODO rgw fails to provide error_code
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_create_naming_bad_punctuation():
|
||||
# characters other than [a-zA-Z0-9._-]
|
||||
check_bad_bucket_name('alpha!soup')
|
||||
|
||||
|
||||
# test_bucket_create_naming_dns_* are valid but not recommended
|
||||
|
||||
def test_bucket_create_naming_dns_underscore():
|
||||
check_good_bucket_name('foo_bar')
|
||||
|
||||
|
||||
def test_bucket_create_naming_dns_long():
|
||||
assert len(prefix) < 50
|
||||
num = 100 - len(prefix)
|
||||
check_good_bucket_name(num * 'a')
|
||||
|
||||
|
||||
def test_bucket_create_naming_dns_dash_at_end():
|
||||
check_good_bucket_name('foo-')
|
||||
|
||||
|
||||
def test_bucket_create_naming_dns_dot_dot():
|
||||
check_good_bucket_name('foo..bar')
|
||||
|
||||
|
||||
def test_bucket_create_naming_dns_dot_dash():
|
||||
check_good_bucket_name('foo.-bar')
|
||||
|
||||
|
||||
def test_bucket_create_naming_dns_dash_dot():
|
||||
check_good_bucket_name('foo-.bar')
|
||||
|
||||
|
||||
# TODO rgw create_bucket() gives 409 Conflict even when owner matches;
|
||||
# AWS ignores double-PUT
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_create_exists():
|
||||
bucket = get_new_bucket()
|
||||
# REST idempotency means this should be a nop
|
||||
s3.main.create_bucket(bucket.name)
|
||||
|
||||
|
||||
def test_bucket_create_exists_nonowner():
|
||||
# Names are shared across a global namespace. As such, no two
|
||||
# users can create a bucket with that same name.
|
||||
bucket = get_new_bucket()
|
||||
e = assert_raises(boto.exception.S3CreateError, s3.alt.create_bucket, bucket.name)
|
||||
eq(e.status, 409)
|
||||
eq(e.reason, 'Conflict')
|
||||
eq(e.error_code, 'BucketAlreadyExists')
|
||||
|
||||
|
||||
def test_bucket_delete_nonowner():
|
||||
bucket = get_new_bucket()
|
||||
check_access_denied(s3.alt.delete_bucket, bucket.name)
|
||||
|
||||
|
||||
# TODO radosgw returns the access_key instead of user_id
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_acl_default():
|
||||
bucket = get_new_bucket()
|
||||
policy = bucket.get_acl()
|
||||
print repr(policy)
|
||||
eq(policy.owner.type, None)
|
||||
eq(policy.owner.id, config.main.user_id)
|
||||
eq(policy.owner.display_name, config.main.display_name)
|
||||
eq(len(policy.acl.grants), 1)
|
||||
eq(policy.acl.grants[0].permission, 'FULL_CONTROL')
|
||||
eq(policy.acl.grants[0].id, policy.owner.id)
|
||||
eq(policy.acl.grants[0].display_name, policy.owner.display_name)
|
||||
eq(policy.acl.grants[0].uri, None)
|
||||
eq(policy.acl.grants[0].email_address, None)
|
||||
eq(policy.acl.grants[0].type, 'CanonicalUser')
|
||||
|
||||
|
||||
# TODO rgw bucket.set_acl() gives 403 Forbidden
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_acl_canned():
|
||||
bucket = get_new_bucket()
|
||||
# Since it defaults to private, set it public-read first
|
||||
bucket.set_acl('public-read')
|
||||
policy = bucket.get_acl()
|
||||
print repr(policy)
|
||||
eq(len(policy.acl.grants), 2)
|
||||
eq(policy.acl.grants[0].permission, 'FULL_CONTROL')
|
||||
eq(policy.acl.grants[0].id, policy.owner.id)
|
||||
eq(policy.acl.grants[0].display_name, policy.owner.display_name)
|
||||
eq(policy.acl.grants[0].uri, None)
|
||||
eq(policy.acl.grants[0].email_address, None)
|
||||
eq(policy.acl.grants[0].type, 'CanonicalUser')
|
||||
eq(policy.acl.grants[1].permission, 'READ')
|
||||
eq(policy.acl.grants[1].id, None)
|
||||
eq(policy.acl.grants[1].display_name, None)
|
||||
eq(policy.acl.grants[1].uri, 'http://acs.amazonaws.com/groups/global/AllUsers')
|
||||
eq(policy.acl.grants[1].email_address, None)
|
||||
eq(policy.acl.grants[1].type, 'Group')
|
||||
|
||||
# Then back to private.
|
||||
bucket.set_acl('private')
|
||||
policy = bucket.get_acl()
|
||||
print repr(policy)
|
||||
eq(len(policy.acl.grants), 1)
|
||||
eq(policy.acl.grants[0].permission, 'FULL_CONTROL')
|
||||
eq(policy.acl.grants[0].id, policy.owner.id)
|
||||
eq(policy.acl.grants[0].display_name, policy.owner.display_name)
|
||||
eq(policy.acl.grants[0].uri, None)
|
||||
eq(policy.acl.grants[0].email_address, None)
|
||||
eq(policy.acl.grants[0].type, 'CanonicalUser')
|
||||
|
||||
|
||||
# TODO rgw bucket.set_acl() gives 403 Forbidden
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_acl_canned_private_to_private():
|
||||
bucket = get_new_bucket()
|
||||
bucket.set_acl('private')
|
||||
|
||||
|
||||
# TODO rgw bucket.set_acl() gives 403 Forbidden
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_acl_grant_userid():
|
||||
bucket = get_new_bucket()
|
||||
# add alt user
|
||||
policy = bucket.get_acl()
|
||||
policy.acl.add_user_grant('FULL_CONTROL', config.alt.user_id)
|
||||
bucket.set_acl(policy)
|
||||
policy = bucket.get_acl()
|
||||
eq(len(policy.acl.grants), 2)
|
||||
eq(policy.acl.grants[1].permission, 'FULL_CONTROL')
|
||||
eq(policy.acl.grants[1].id, config.alt.user_id)
|
||||
eq(policy.acl.grants[1].display_name, config.alt.display_name)
|
||||
eq(policy.acl.grants[1].uri, None)
|
||||
eq(policy.acl.grants[1].email_address, None)
|
||||
eq(policy.acl.grants[1].type, 'CanonicalUser')
|
||||
|
||||
# alt user can write
|
||||
bucket2 = s3.alt.get_bucket(bucket.name)
|
||||
key = bucket2.new_key('foo')
|
||||
key.set_contents_from_string('bar')
|
||||
|
||||
|
||||
# TODO rgw bucket.set_acl() gives 403 Forbidden
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_acl_grant_email():
|
||||
bucket = get_new_bucket()
|
||||
# add alt user
|
||||
policy = bucket.get_acl()
|
||||
policy.acl.add_email_grant('FULL_CONTROL', config.alt.email)
|
||||
bucket.set_acl(policy)
|
||||
policy = bucket.get_acl()
|
||||
eq(len(policy.acl.grants), 2)
|
||||
eq(policy.acl.grants[1].permission, 'FULL_CONTROL')
|
||||
eq(policy.acl.grants[1].id, config.alt.user_id)
|
||||
eq(policy.acl.grants[1].display_name, config.alt.display_name)
|
||||
eq(policy.acl.grants[1].uri, None)
|
||||
eq(policy.acl.grants[1].email_address, None)
|
||||
eq(policy.acl.grants[1].type, 'CanonicalUser')
|
||||
|
||||
# alt user can write
|
||||
bucket2 = s3.alt.get_bucket(bucket.name)
|
||||
key = bucket2.new_key('foo')
|
||||
key.set_contents_from_string('bar')
|
||||
|
||||
|
||||
# TODO rgw gives 403 error
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_acl_grant_email_notexist():
|
||||
# behavior not documented by amazon
|
||||
bucket = get_new_bucket()
|
||||
policy = bucket.get_acl()
|
||||
policy.acl.add_email_grant('FULL_CONTROL', NONEXISTENT_EMAIL)
|
||||
e = assert_raises(boto.exception.S3ResponseError, bucket.set_acl, policy)
|
||||
eq(e.status, 400)
|
||||
eq(e.reason, 'Bad Request')
|
||||
eq(e.error_code, 'UnresolvableGrantByEmailAddress')
|
||||
|
||||
|
||||
# TODO rgw bucket.set_acl() gives 403 Forbidden
|
||||
@attr('fails_on_rgw')
|
||||
def test_bucket_acl_revoke_all():
|
||||
# revoke all access, including the owner's access
|
||||
bucket = get_new_bucket()
|
||||
policy = bucket.get_acl()
|
||||
policy.acl.grants = []
|
||||
bucket.set_acl(policy)
|
||||
policy = bucket.get_acl()
|
||||
eq(len(policy.acl.grants), 0)
|
||||
|
||||
|
||||
# TODO rgw log_bucket.set_as_logging_target() gives 403 Forbidden
|
||||
@attr('fails_on_rgw')
|
||||
def test_logging_toggle():
|
||||
bucket = get_new_bucket()
|
||||
log_bucket = s3.main.create_bucket(bucket.name + '-log')
|
||||
log_bucket.set_as_logging_target()
|
||||
bucket.enable_logging(target_bucket=log_bucket, target_prefix=bucket.name)
|
||||
bucket.disable_logging()
|
||||
|
||||
|
||||
def _setup_access(bucket_acl, object_acl):
|
||||
"""
|
||||
Simple test fixture: create a bucket with given ACL, with objects:
|
||||
|
||||
- a: given ACL
|
||||
- b: default ACL
|
||||
"""
|
||||
obj = bunch.Bunch()
|
||||
bucket = get_new_bucket()
|
||||
bucket.set_acl(bucket_acl)
|
||||
obj.a = bucket.new_key('foo')
|
||||
obj.a.set_contents_from_string('foocontent')
|
||||
obj.a.set_acl(object_acl)
|
||||
obj.b = bucket.new_key('bar')
|
||||
obj.b.set_contents_from_string('barcontent')
|
||||
|
||||
obj.bucket2 = s3.alt.get_bucket(bucket.name, validate=False)
|
||||
obj.a2 = obj.bucket2.new_key(obj.a.name)
|
||||
obj.b2 = obj.bucket2.new_key(obj.b.name)
|
||||
obj.new = obj.bucket2.new_key('new')
|
||||
|
||||
return obj
|
||||
|
||||
|
||||
def get_bucket_key_names(bucket):
|
||||
return frozenset(k.name for k in bucket.list())
|
||||
|
||||
|
||||
# TODO bucket.set_acl('private') fails on rgw
|
||||
@attr('fails_on_rgw')
|
||||
def test_access_bucket_private_object_private():
|
||||
# all the test_access_* tests follow this template
|
||||
obj = _setup_access(bucket_acl='private', object_acl='private')
|
||||
# acled object read fail
|
||||
check_access_denied(obj.a2.get_contents_as_string)
|
||||
# acled object write fail
|
||||
check_access_denied(obj.a2.set_contents_from_string, 'barcontent')
|
||||
# default object read fail
|
||||
check_access_denied(obj.b2.get_contents_as_string)
|
||||
# default object write fail
|
||||
check_access_denied(obj.b2.set_contents_from_string, 'baroverwrite')
|
||||
# bucket read fail
|
||||
check_access_denied(get_bucket_key_names, obj.bucket2)
|
||||
# bucket write fail
|
||||
check_access_denied(obj.new.set_contents_from_string, 'newcontent')
|
||||
|
||||
|
||||
# TODO bucket.set_acl('private') fails on rgw
|
||||
@attr('fails_on_rgw')
|
||||
def test_access_bucket_private_object_publicread():
|
||||
obj = _setup_access(bucket_acl='private', object_acl='public-read')
|
||||
eq(obj.a2.get_contents_as_string(), 'foocontent')
|
||||
check_access_denied(obj.a2.set_contents_from_string, 'foooverwrite')
|
||||
check_access_denied(obj.b2.get_contents_as_string)
|
||||
check_access_denied(obj.b2.set_contents_from_string, 'baroverwrite')
|
||||
check_access_denied(get_bucket_key_names, obj.bucket2)
|
||||
check_access_denied(obj.new.set_contents_from_string, 'newcontent')
|
||||
|
||||
|
||||
# TODO bucket.set_acl('private') fails on rgw
|
||||
@attr('fails_on_rgw')
|
||||
def test_access_bucket_private_object_publicreadwrite():
|
||||
obj = _setup_access(bucket_acl='private', object_acl='public-read-write')
|
||||
eq(obj.a2.get_contents_as_string(), 'foocontent')
|
||||
### TODO: it seems AWS denies this write, even when we expected it
|
||||
### to complete; as it is unclear what the actual desired behavior
|
||||
### is (the docs are somewhat unclear), we'll just codify current
|
||||
### AWS behavior, at least for now.
|
||||
# obj.a2.set_contents_from_string('foooverwrite')
|
||||
check_access_denied(obj.a2.set_contents_from_string, 'foooverwrite')
|
||||
check_access_denied(obj.b2.get_contents_as_string)
|
||||
check_access_denied(obj.b2.set_contents_from_string, 'baroverwrite')
|
||||
check_access_denied(get_bucket_key_names, obj.bucket2)
|
||||
check_access_denied(obj.new.set_contents_from_string, 'newcontent')
|
||||
|
||||
|
||||
# TODO bucket.set_acl('public-read') fails on rgw
|
||||
@attr('fails_on_rgw')
|
||||
def test_access_bucket_publicread_object_private():
|
||||
obj = _setup_access(bucket_acl='public-read', object_acl='private')
|
||||
check_access_denied(obj.a2.get_contents_as_string)
|
||||
check_access_denied(obj.a2.set_contents_from_string, 'barcontent')
|
||||
### TODO: i don't understand why this gets denied, but codifying what
|
||||
### AWS does
|
||||
# eq(obj.b2.get_contents_as_string(), 'barcontent')
|
||||
check_access_denied(obj.b2.get_contents_as_string)
|
||||
check_access_denied(obj.b2.set_contents_from_string, 'baroverwrite')
|
||||
eq(get_bucket_key_names(obj.bucket2), frozenset(['foo', 'bar']))
|
||||
check_access_denied(obj.new.set_contents_from_string, 'newcontent')
|
||||
|
||||
|
||||
# TODO bucket.set_acl('public-read') fails on rgw
|
||||
@attr('fails_on_rgw')
|
||||
def test_access_bucket_publicread_object_publicread():
|
||||
obj = _setup_access(bucket_acl='public-read', object_acl='public-read')
|
||||
eq(obj.a2.get_contents_as_string(), 'foocontent')
|
||||
check_access_denied(obj.a2.set_contents_from_string, 'foooverwrite')
|
||||
### TODO: i don't understand why this gets denied, but codifying what
|
||||
### AWS does
|
||||
# eq(obj.b2.get_contents_as_string(), 'barcontent')
|
||||
check_access_denied(obj.b2.get_contents_as_string)
|
||||
check_access_denied(obj.b2.set_contents_from_string, 'baroverwrite')
|
||||
eq(get_bucket_key_names(obj.bucket2), frozenset(['foo', 'bar']))
|
||||
check_access_denied(obj.new.set_contents_from_string, 'newcontent')
|
||||
|
||||
|
||||
# TODO bucket.set_acl('public-read') fails on rgw
|
||||
@attr('fails_on_rgw')
|
||||
def test_access_bucket_publicread_object_publicreadwrite():
|
||||
obj = _setup_access(bucket_acl='public-read', object_acl='public-read-write')
|
||||
eq(obj.a2.get_contents_as_string(), 'foocontent')
|
||||
### TODO: it seems AWS denies this write, even when we expected it
|
||||
### to complete; as it is unclear what the actual desired behavior
|
||||
### is (the docs are somewhat unclear), we'll just codify current
|
||||
### AWS behavior, at least for now.
|
||||
# obj.a2.set_contents_from_string('foooverwrite')
|
||||
check_access_denied(obj.a2.set_contents_from_string, 'foooverwrite')
|
||||
### TODO: i don't understand why this gets denied, but codifying what
|
||||
### AWS does
|
||||
# eq(obj.b2.get_contents_as_string(), 'barcontent')
|
||||
check_access_denied(obj.b2.get_contents_as_string)
|
||||
check_access_denied(obj.b2.set_contents_from_string, 'baroverwrite')
|
||||
eq(get_bucket_key_names(obj.bucket2), frozenset(['foo', 'bar']))
|
||||
check_access_denied(obj.new.set_contents_from_string, 'newcontent')
|
||||
|
||||
|
||||
# TODO bucket.set_acl('public-read-write') fails on rgw
|
||||
@attr('fails_on_rgw')
|
||||
def test_access_bucket_publicreadwrite_object_private():
|
||||
obj = _setup_access(bucket_acl='public-read-write', object_acl='private')
|
||||
check_access_denied(obj.a2.get_contents_as_string)
|
||||
obj.a2.set_contents_from_string('barcontent')
|
||||
### TODO: i don't understand why this gets denied, but codifying what
|
||||
### AWS does
|
||||
# eq(obj.b2.get_contents_as_string(), 'barcontent')
|
||||
check_access_denied(obj.b2.get_contents_as_string)
|
||||
obj.b2.set_contents_from_string('baroverwrite')
|
||||
eq(get_bucket_key_names(obj.bucket2), frozenset(['foo', 'bar']))
|
||||
obj.new.set_contents_from_string('newcontent')
|
||||
|
||||
|
||||
# TODO bucket.set_acl('public-read-write') fails on rgw
|
||||
@attr('fails_on_rgw')
|
||||
def test_access_bucket_publicreadwrite_object_publicread():
|
||||
obj = _setup_access(bucket_acl='public-read-write', object_acl='public-read')
|
||||
eq(obj.a2.get_contents_as_string(), 'foocontent')
|
||||
obj.a2.set_contents_from_string('barcontent')
|
||||
### TODO: i don't understand why this gets denied, but codifying what
|
||||
### AWS does
|
||||
# eq(obj.b2.get_contents_as_string(), 'barcontent')
|
||||
check_access_denied(obj.b2.get_contents_as_string)
|
||||
obj.b2.set_contents_from_string('baroverwrite')
|
||||
eq(get_bucket_key_names(obj.bucket2), frozenset(['foo', 'bar']))
|
||||
obj.new.set_contents_from_string('newcontent')
|
||||
|
||||
|
||||
# TODO bucket.set_acl('public-read-write') fails on rgw
|
||||
@attr('fails_on_rgw')
|
||||
def test_access_bucket_publicreadwrite_object_publicreadwrite():
|
||||
obj = _setup_access(bucket_acl='public-read-write', object_acl='public-read-write')
|
||||
eq(obj.a2.get_contents_as_string(), 'foocontent')
|
||||
obj.a2.set_contents_from_string('foooverwrite')
|
||||
### TODO: i don't understand why this gets denied, but codifying what
|
||||
### AWS does
|
||||
# eq(obj.b2.get_contents_as_string(), 'barcontent')
|
||||
check_access_denied(obj.b2.get_contents_as_string)
|
||||
obj.b2.set_contents_from_string('baroverwrite')
|
||||
eq(get_bucket_key_names(obj.bucket2), frozenset(['foo', 'bar']))
|
||||
obj.new.set_contents_from_string('newcontent')
|
14
utils.py
Normal file
14
utils.py
Normal file
|
@ -0,0 +1,14 @@
|
|||
def assert_raises(excClass, callableObj, *args, **kwargs):
|
||||
"""
|
||||
Like unittest.TestCase.assertRaises, but returns the exception.
|
||||
"""
|
||||
try:
|
||||
callableObj(*args, **kwargs)
|
||||
except excClass as e:
|
||||
return e
|
||||
else:
|
||||
if hasattr(excClass, '__name__'):
|
||||
excName = excClass.__name__
|
||||
else:
|
||||
excName = str(excClass)
|
||||
raise AssertionError("%s not raised" % excName)
|
Loading…
Reference in a new issue