forked from TrueCloudLab/s3-tests
Compare commits
8 commits
master
...
hammer-s3d
Author | SHA1 | Date | |
---|---|---|---|
|
3b1b4638f4 | ||
|
7d26245b3a | ||
|
dc26c0d696 | ||
|
34530164b1 | ||
|
5d72a43ffa | ||
|
366fdee269 | ||
|
f041a73153 | ||
|
4342e14114 |
8 changed files with 105 additions and 34 deletions
|
@ -41,5 +41,4 @@ virtualenv --no-site-packages --distribute virtualenv
|
||||||
# easy_install, and we really wanted pip; next line will fail if pip
|
# easy_install, and we really wanted pip; next line will fail if pip
|
||||||
# requirements.txt does not match setup.py requirements -- sucky but
|
# requirements.txt does not match setup.py requirements -- sucky but
|
||||||
# good enough for now
|
# good enough for now
|
||||||
./virtualenv/bin/python setup.py develop \
|
./virtualenv/bin/python setup.py develop
|
||||||
--allow-hosts None
|
|
||||||
|
|
|
@ -3,7 +3,7 @@ nose >=1.0.0
|
||||||
boto >=2.6.0
|
boto >=2.6.0
|
||||||
bunch >=1.0.0
|
bunch >=1.0.0
|
||||||
# 0.14 switches to libev, that means bootstrap needs to change too
|
# 0.14 switches to libev, that means bootstrap needs to change too
|
||||||
gevent ==0.13.6
|
gevent >=1.0
|
||||||
isodate >=0.4.4
|
isodate >=0.4.4
|
||||||
requests ==0.14.0
|
requests ==0.14.0
|
||||||
pytz >=2011k
|
pytz >=2011k
|
||||||
|
|
|
@ -108,6 +108,9 @@ def connect(conf):
|
||||||
raise RuntimeError(
|
raise RuntimeError(
|
||||||
'calling_format unknown: %r' % raw_calling_format
|
'calling_format unknown: %r' % raw_calling_format
|
||||||
)
|
)
|
||||||
|
if conf.has_key('debug'):
|
||||||
|
kwargs['debug']=conf['debug']
|
||||||
|
boto.set_stream_logger('boto')
|
||||||
# TODO test vhost calling format
|
# TODO test vhost calling format
|
||||||
conn = boto.s3.connection.S3Connection(**kwargs)
|
conn = boto.s3.connection.S3Connection(**kwargs)
|
||||||
return conn
|
return conn
|
||||||
|
|
|
@ -64,22 +64,31 @@ def nuke_prefixed_buckets_on_conn(prefix, name, conn):
|
||||||
print 'prefix=',prefix
|
print 'prefix=',prefix
|
||||||
if bucket.name.startswith(prefix):
|
if bucket.name.startswith(prefix):
|
||||||
print 'Cleaning bucket {bucket}'.format(bucket=bucket)
|
print 'Cleaning bucket {bucket}'.format(bucket=bucket)
|
||||||
try:
|
success = False
|
||||||
# bucket.set_canned_acl('private')
|
for i in xrange(2):
|
||||||
for key in bucket.list_versions():
|
try:
|
||||||
print 'Cleaning bucket {bucket} key {key}'.format(
|
for key in bucket.list_versions():
|
||||||
bucket=bucket,
|
print 'Cleaning bucket {bucket} key {key}'.format(
|
||||||
key=key,
|
bucket=bucket,
|
||||||
)
|
key=key,
|
||||||
# key.set_canned_acl('private')
|
)
|
||||||
bucket.delete_key(key.name, version_id = key.version_id)
|
# key.set_canned_acl('private')
|
||||||
bucket.delete()
|
bucket.delete_key(key.name, version_id = key.version_id)
|
||||||
except boto.exception.S3ResponseError as e:
|
bucket.delete()
|
||||||
if e.error_code != 'AccessDenied':
|
success = True
|
||||||
print 'GOT UNWANTED ERROR', e.error_code
|
except boto.exception.S3ResponseError as e:
|
||||||
raise
|
if e.error_code != 'AccessDenied':
|
||||||
# seems like we're not the owner of the bucket; ignore
|
print 'GOT UNWANTED ERROR', e.error_code
|
||||||
pass
|
raise
|
||||||
|
# seems like we don't have permissions set appropriately, we'll
|
||||||
|
# modify permissions and retry
|
||||||
|
pass
|
||||||
|
|
||||||
|
if success:
|
||||||
|
return
|
||||||
|
|
||||||
|
bucket.set_canned_acl('private')
|
||||||
|
|
||||||
|
|
||||||
def nuke_prefixed_buckets(prefix):
|
def nuke_prefixed_buckets(prefix):
|
||||||
# If no regions are specified, use the simple method
|
# If no regions are specified, use the simple method
|
||||||
|
@ -106,7 +115,6 @@ def nuke_prefixed_buckets(prefix):
|
||||||
|
|
||||||
print 'Done with cleanup of test buckets.'
|
print 'Done with cleanup of test buckets.'
|
||||||
|
|
||||||
|
|
||||||
class TargetConfig:
|
class TargetConfig:
|
||||||
def __init__(self, cfg, section):
|
def __init__(self, cfg, section):
|
||||||
self.port = None
|
self.port = None
|
||||||
|
|
|
@ -3261,6 +3261,27 @@ def test_object_acl_canned_bucketownerfullcontrol():
|
||||||
key.delete()
|
key.delete()
|
||||||
bucket.delete()
|
bucket.delete()
|
||||||
|
|
||||||
|
@attr(resource='object.acls')
|
||||||
|
@attr(method='put')
|
||||||
|
@attr(operation='set write-acp')
|
||||||
|
@attr(assertion='does not modify owner')
|
||||||
|
def test_object_acl_full_control_verify_owner():
|
||||||
|
bucket = get_new_bucket(targets.main.default)
|
||||||
|
bucket.set_acl('public-read-write')
|
||||||
|
|
||||||
|
key = bucket.new_key('foo')
|
||||||
|
key.set_contents_from_string('bar')
|
||||||
|
|
||||||
|
key.add_user_grant(permission='FULL_CONTROL', user_id=config.alt.user_id)
|
||||||
|
|
||||||
|
k2 = s3.alt.get_bucket(bucket.name).get_key('foo')
|
||||||
|
|
||||||
|
k2.add_user_grant(permission='READ_ACP', user_id=config.alt.user_id)
|
||||||
|
|
||||||
|
policy = k2.get_acl()
|
||||||
|
eq(policy.owner.id, config.main.user_id)
|
||||||
|
|
||||||
|
|
||||||
@attr(resource='bucket')
|
@attr(resource='bucket')
|
||||||
@attr(method='ACLs')
|
@attr(method='ACLs')
|
||||||
@attr(operation='set acl private')
|
@attr(operation='set acl private')
|
||||||
|
@ -3532,6 +3553,12 @@ def test_bucket_acl_grant_userid_fullcontrol():
|
||||||
# can write acl
|
# can write acl
|
||||||
_check_bucket_acl_grant_can_writeacp(bucket)
|
_check_bucket_acl_grant_can_writeacp(bucket)
|
||||||
|
|
||||||
|
# verify owner did not change
|
||||||
|
bucket2 = s3.main.get_bucket(bucket.name)
|
||||||
|
policy = bucket2.get_acl()
|
||||||
|
eq(policy.owner.id, config.main.user_id)
|
||||||
|
eq(policy.owner.display_name, config.main.display_name)
|
||||||
|
|
||||||
|
|
||||||
@attr(resource='bucket')
|
@attr(resource='bucket')
|
||||||
@attr(method='ACLs')
|
@attr(method='ACLs')
|
||||||
|
@ -5762,7 +5789,7 @@ def test_versioning_multi_object_delete_with_marker_create():
|
||||||
|
|
||||||
keyname = 'key'
|
keyname = 'key'
|
||||||
|
|
||||||
rmkeys = { bucket.new_key(keyname) }
|
rmkeys = [ bucket.new_key(keyname) ]
|
||||||
|
|
||||||
eq(_count_bucket_versioned_objs(bucket), 0)
|
eq(_count_bucket_versioned_objs(bucket), 0)
|
||||||
|
|
||||||
|
@ -5852,17 +5879,18 @@ def test_versioned_object_acl():
|
||||||
check_grants(k.get_acl().acl.grants, default_policy)
|
check_grants(k.get_acl().acl.grants, default_policy)
|
||||||
|
|
||||||
|
|
||||||
def _do_create_object(bucket, objname, i):
|
def _do_create_object(bucket, objname, i, obj_size):
|
||||||
k = bucket.new_key(objname)
|
k = bucket.new_key(objname)
|
||||||
k.set_contents_from_string('data {i}'.format(i=i))
|
s = 'x' * obj_size
|
||||||
|
k.set_contents_from_string(s)
|
||||||
|
|
||||||
def _do_remove_ver(bucket, obj):
|
def _do_remove_ver(bucket, obj):
|
||||||
bucket.delete_key(obj.name, version_id = obj.version_id)
|
bucket.delete_key(obj.name, version_id = obj.version_id)
|
||||||
|
|
||||||
def _do_create_versioned_obj_concurrent(bucket, objname, num):
|
def _do_create_obj_concurrent(bucket, objname, num, obj_size=0):
|
||||||
t = []
|
t = []
|
||||||
for i in range(num):
|
for i in range(num):
|
||||||
thr = threading.Thread(target = _do_create_object, args=(bucket, objname, i))
|
thr = threading.Thread(target = _do_create_object, args=(bucket, objname, i, obj_size))
|
||||||
thr.start()
|
thr.start()
|
||||||
t.append(thr)
|
t.append(thr)
|
||||||
return t
|
return t
|
||||||
|
@ -5894,7 +5922,7 @@ def test_versioned_concurrent_object_create_concurrent_remove():
|
||||||
num_objs = 5
|
num_objs = 5
|
||||||
|
|
||||||
for i in xrange(5):
|
for i in xrange(5):
|
||||||
t = _do_create_versioned_obj_concurrent(bucket, keyname, num_objs)
|
t = _do_create_obj_concurrent(bucket, keyname, num_objs)
|
||||||
_do_wait_completion(t)
|
_do_wait_completion(t)
|
||||||
|
|
||||||
eq(_count_bucket_versioned_objs(bucket), num_objs)
|
eq(_count_bucket_versioned_objs(bucket), num_objs)
|
||||||
|
@ -5923,7 +5951,7 @@ def test_versioned_concurrent_object_create_and_remove():
|
||||||
all_threads = []
|
all_threads = []
|
||||||
|
|
||||||
for i in xrange(3):
|
for i in xrange(3):
|
||||||
t = _do_create_versioned_obj_concurrent(bucket, keyname, num_objs)
|
t = _do_create_obj_concurrent(bucket, keyname, num_objs)
|
||||||
all_threads.append(t)
|
all_threads.append(t)
|
||||||
|
|
||||||
t = _do_clear_versioned_bucket_concurrent(bucket)
|
t = _do_clear_versioned_bucket_concurrent(bucket)
|
||||||
|
@ -5938,3 +5966,36 @@ def test_versioned_concurrent_object_create_and_remove():
|
||||||
|
|
||||||
eq(_count_bucket_versioned_objs(bucket), 0)
|
eq(_count_bucket_versioned_objs(bucket), 0)
|
||||||
eq(len(bucket.get_all_keys()), 0)
|
eq(len(bucket.get_all_keys()), 0)
|
||||||
|
|
||||||
|
@attr(resource='object')
|
||||||
|
@attr(method='put')
|
||||||
|
@attr(operation='concurrent creation of objects, concurrent removal')
|
||||||
|
@attr(assertion='works')
|
||||||
|
def test_non_versioned_concurrent_object_create_concurrent_remove():
|
||||||
|
bucket = get_new_bucket()
|
||||||
|
|
||||||
|
# not configuring versioning here! this test is non-versioned
|
||||||
|
|
||||||
|
keyname = 'myobj'
|
||||||
|
|
||||||
|
num_objs = 5
|
||||||
|
obj_size = 10 # non-zero
|
||||||
|
|
||||||
|
for i in xrange(5):
|
||||||
|
t = _do_create_obj_concurrent(bucket, keyname, num_objs, obj_size)
|
||||||
|
_do_wait_completion(t)
|
||||||
|
|
||||||
|
keys = []
|
||||||
|
for k in bucket.get_all_keys():
|
||||||
|
keys.append(k)
|
||||||
|
eq(len(keys), 1)
|
||||||
|
eq(keys[0].size, obj_size)
|
||||||
|
|
||||||
|
|
||||||
|
t = _do_clear_versioned_bucket_concurrent(bucket)
|
||||||
|
_do_wait_completion(t)
|
||||||
|
|
||||||
|
eq(_count_bucket_versioned_objs(bucket), 0)
|
||||||
|
eq(len(bucket.get_all_keys()), 0)
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -195,7 +195,7 @@ def main():
|
||||||
warmup_pool = gevent.pool.Pool(size=100)
|
warmup_pool = gevent.pool.Pool(size=100)
|
||||||
for file_name in file_names:
|
for file_name in file_names:
|
||||||
fp = next(files)
|
fp = next(files)
|
||||||
warmup_pool.spawn_link_exception(
|
warmup_pool.spawn(
|
||||||
write_file,
|
write_file,
|
||||||
bucket=bucket,
|
bucket=bucket,
|
||||||
file_name=file_name,
|
file_name=file_name,
|
||||||
|
@ -214,7 +214,7 @@ def main():
|
||||||
if not config.readwrite.get('deterministic_file_names'):
|
if not config.readwrite.get('deterministic_file_names'):
|
||||||
for x in xrange(config.readwrite.writers):
|
for x in xrange(config.readwrite.writers):
|
||||||
this_rand = random.Random(rand_writer.randrange(2**32))
|
this_rand = random.Random(rand_writer.randrange(2**32))
|
||||||
group.spawn_link_exception(
|
group.spawn(
|
||||||
writer,
|
writer,
|
||||||
bucket=bucket,
|
bucket=bucket,
|
||||||
worker_id=x,
|
worker_id=x,
|
||||||
|
@ -231,7 +231,7 @@ def main():
|
||||||
rand_reader = random.Random(seeds['reader'])
|
rand_reader = random.Random(seeds['reader'])
|
||||||
for x in xrange(config.readwrite.readers):
|
for x in xrange(config.readwrite.readers):
|
||||||
this_rand = random.Random(rand_reader.randrange(2**32))
|
this_rand = random.Random(rand_reader.randrange(2**32))
|
||||||
group.spawn_link_exception(
|
group.spawn(
|
||||||
reader,
|
reader,
|
||||||
bucket=bucket,
|
bucket=bucket,
|
||||||
worker_id=x,
|
worker_id=x,
|
||||||
|
|
|
@ -161,7 +161,7 @@ def main():
|
||||||
)
|
)
|
||||||
q = gevent.queue.Queue()
|
q = gevent.queue.Queue()
|
||||||
|
|
||||||
logger_g = gevent.spawn_link_exception(yaml.safe_dump_all, q, stream=real_stdout)
|
logger_g = gevent.spawn(yaml.safe_dump_all, q, stream=real_stdout)
|
||||||
|
|
||||||
print "Writing {num} objects with {w} workers...".format(
|
print "Writing {num} objects with {w} workers...".format(
|
||||||
num=config.roundtrip.files.num,
|
num=config.roundtrip.files.num,
|
||||||
|
@ -171,7 +171,7 @@ def main():
|
||||||
start = time.time()
|
start = time.time()
|
||||||
for objname in objnames:
|
for objname in objnames:
|
||||||
fp = next(files)
|
fp = next(files)
|
||||||
pool.spawn_link_exception(
|
pool.spawn(
|
||||||
writer,
|
writer,
|
||||||
bucket=bucket,
|
bucket=bucket,
|
||||||
objname=objname,
|
objname=objname,
|
||||||
|
@ -195,7 +195,7 @@ def main():
|
||||||
pool = gevent.pool.Pool(size=config.roundtrip.readers)
|
pool = gevent.pool.Pool(size=config.roundtrip.readers)
|
||||||
start = time.time()
|
start = time.time()
|
||||||
for objname in objnames:
|
for objname in objnames:
|
||||||
pool.spawn_link_exception(
|
pool.spawn(
|
||||||
reader,
|
reader,
|
||||||
bucket=bucket,
|
bucket=bucket,
|
||||||
objname=objname,
|
objname=objname,
|
||||||
|
|
2
setup.py
2
setup.py
|
@ -16,7 +16,7 @@ setup(
|
||||||
'boto >=2.0b4',
|
'boto >=2.0b4',
|
||||||
'PyYAML',
|
'PyYAML',
|
||||||
'bunch >=1.0.0',
|
'bunch >=1.0.0',
|
||||||
'gevent ==0.13.6',
|
'gevent >=1.0',
|
||||||
'isodate >=0.4.4',
|
'isodate >=0.4.4',
|
||||||
],
|
],
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue