2020-07-01 02:28:31 +00:00
|
|
|
#!/usr/bin/python3
|
|
|
|
|
|
|
|
import subprocess
|
|
|
|
import os
|
|
|
|
import re
|
|
|
|
import binascii
|
|
|
|
import uuid
|
|
|
|
import hashlib
|
|
|
|
from robot.api.deco import keyword
|
|
|
|
from robot.api import logger
|
|
|
|
|
2020-11-30 10:33:05 +00:00
|
|
|
if os.getenv('ROBOT_PROFILE') == 'selectel_smoke':
|
|
|
|
from selectelcdn_smoke_vars import (NEOGO_CLI_PREFIX, NEO_MAINNET_ENDPOINT,
|
|
|
|
NEOFS_NEO_API_ENDPOINT, NEOFS_ENDPOINT)
|
|
|
|
else:
|
|
|
|
from neofs_int_vars import (NEOGO_CLI_PREFIX, NEO_MAINNET_ENDPOINT,
|
|
|
|
NEOFS_NEO_API_ENDPOINT, NEOFS_ENDPOINT)
|
2020-07-01 02:28:31 +00:00
|
|
|
|
|
|
|
ROBOT_AUTO_KEYWORDS = False
|
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
CLI_PREFIX = ""
|
|
|
|
|
|
|
|
@keyword('Form WIF from String')
|
|
|
|
def form_wif_from_string(private_key: str):
|
|
|
|
wif = ""
|
|
|
|
Cmd = f'neofs-cli util keyer -u {private_key}'
|
|
|
|
logger.info("Cmd: %s" % Cmd)
|
|
|
|
complProc = subprocess.run(Cmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=150, shell=True)
|
|
|
|
output = complProc.stdout
|
|
|
|
logger.info("Output: %s" % output)
|
|
|
|
|
|
|
|
m = re.search(r'WIF\s+(\w+)', output)
|
2020-11-30 10:33:05 +00:00
|
|
|
if m.start() != m.end():
|
2020-11-18 15:15:57 +00:00
|
|
|
wif = m.group(1)
|
|
|
|
else:
|
|
|
|
raise Exception("Can not get WIF.")
|
|
|
|
|
|
|
|
return wif
|
|
|
|
|
|
|
|
|
|
|
|
@keyword('Get ScripHash')
|
|
|
|
def get_scripthash(privkey: str):
|
|
|
|
scripthash = ""
|
|
|
|
Cmd = f'neofs-cli util keyer -u {privkey}'
|
|
|
|
logger.info("Cmd: %s" % Cmd)
|
|
|
|
complProc = subprocess.run(Cmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=150, shell=True)
|
|
|
|
output = complProc.stdout
|
|
|
|
logger.info("Output: %s" % output)
|
|
|
|
|
|
|
|
# ScriptHash3.0 00284fc88f8ac31f8e56c03301bfab0757e3f212
|
|
|
|
m = re.search(r'ScriptHash3.0 (\w+)', output)
|
2020-11-30 10:33:05 +00:00
|
|
|
if m.start() != m.end():
|
2020-11-18 15:15:57 +00:00
|
|
|
scripthash = m.group(1)
|
|
|
|
else:
|
|
|
|
raise Exception("Can not get ScriptHash.")
|
|
|
|
|
|
|
|
return scripthash
|
2020-07-01 02:28:31 +00:00
|
|
|
|
2020-07-14 00:05:22 +00:00
|
|
|
|
|
|
|
|
|
|
|
@keyword('Get nodes with object')
|
2020-11-18 15:15:57 +00:00
|
|
|
def get_nodes_with_object(private_key: str, cid, oid):
|
2020-07-14 00:05:22 +00:00
|
|
|
storage_nodes = _get_storage_nodes(private_key)
|
|
|
|
copies = 0
|
|
|
|
|
|
|
|
nodes_list = []
|
|
|
|
|
|
|
|
for node in storage_nodes:
|
2020-09-01 03:23:17 +00:00
|
|
|
search_res = _search_object(node, private_key, cid, oid)
|
|
|
|
if search_res:
|
2020-11-18 15:15:57 +00:00
|
|
|
if re.search(r'(%s)' % (oid), search_res):
|
2020-09-01 03:23:17 +00:00
|
|
|
nodes_list.append(node)
|
2020-07-14 00:05:22 +00:00
|
|
|
|
|
|
|
logger.info("Nodes with object: %s" % nodes_list)
|
|
|
|
|
|
|
|
|
|
|
|
@keyword('Get nodes without object')
|
2020-11-18 15:15:57 +00:00
|
|
|
def get_nodes_without_object(private_key: str, cid, oid):
|
2020-07-14 00:05:22 +00:00
|
|
|
storage_nodes = _get_storage_nodes(private_key)
|
|
|
|
copies = 0
|
|
|
|
|
|
|
|
nodes_list = []
|
|
|
|
|
|
|
|
for node in storage_nodes:
|
2020-09-01 03:23:17 +00:00
|
|
|
search_res = _search_object(node, private_key, cid, oid)
|
|
|
|
if search_res:
|
2020-11-18 15:15:57 +00:00
|
|
|
if not re.search(r'(%s)' % (oid), search_res):
|
2020-09-01 03:23:17 +00:00
|
|
|
nodes_list.append(node)
|
|
|
|
else:
|
2020-07-14 00:05:22 +00:00
|
|
|
nodes_list.append(node)
|
|
|
|
|
|
|
|
logger.info("Nodes with object: %s" % nodes_list)
|
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
|
|
|
|
@keyword('Validate storage policy for object')
|
2020-11-18 15:15:57 +00:00
|
|
|
def validate_storage_policy_for_object(private_key: str, expected_copies: int, cid, oid, *expected_node_list):
|
2020-07-01 02:28:31 +00:00
|
|
|
storage_nodes = _get_storage_nodes(private_key)
|
|
|
|
copies = 0
|
2020-09-01 03:23:17 +00:00
|
|
|
found_nodes = []
|
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
for node in storage_nodes:
|
2020-09-01 03:23:17 +00:00
|
|
|
search_res = _search_object(node, private_key, cid, oid)
|
|
|
|
if search_res:
|
2020-11-18 15:15:57 +00:00
|
|
|
if re.search(r'(%s)' % (oid), search_res):
|
2020-09-01 03:23:17 +00:00
|
|
|
copies += 1
|
|
|
|
found_nodes.append(node)
|
|
|
|
|
|
|
|
if copies != expected_copies:
|
|
|
|
raise Exception("Object copies is not match storage policy. Found: %s, expexted: %s." % (copies, expected_copies))
|
|
|
|
else:
|
|
|
|
logger.info("Found copies: %s, expected: %s" % (copies, expected_copies))
|
|
|
|
|
|
|
|
logger.info("Found nodes: %s" % found_nodes)
|
|
|
|
|
|
|
|
if expected_node_list:
|
|
|
|
if sorted(found_nodes) == sorted(expected_node_list):
|
|
|
|
logger.info("Found node list '{}' is equal for expected list '{}'".format(found_nodes, expected_node_list))
|
|
|
|
else:
|
|
|
|
raise Exception("Found node list '{}' is not equal to expected list '{}'".format(found_nodes, expected_node_list))
|
|
|
|
|
2020-07-14 00:05:22 +00:00
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
|
|
|
|
|
2020-08-13 22:09:00 +00:00
|
|
|
|
|
|
|
@keyword('Get eACL')
|
|
|
|
def get_eacl(private_key: bytes, cid: str):
|
|
|
|
|
|
|
|
Cmd = f'{CLI_PREFIX}neofs-cli --host {NEOFS_ENDPOINT} --key {binascii.hexlify(private_key).decode()} container get-eacl --cid {cid}'
|
|
|
|
logger.info("Cmd: %s" % Cmd)
|
|
|
|
complProc = subprocess.run(Cmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=150, shell=True)
|
|
|
|
output = complProc.stdout
|
|
|
|
logger.info("Output: %s" % output)
|
|
|
|
|
|
|
|
|
2020-08-19 22:31:16 +00:00
|
|
|
|
|
|
|
@keyword('Convert Str to Hex Str with Len')
|
|
|
|
def conver_str_to_hex(string_convert: str):
|
|
|
|
converted = binascii.hexlify(bytes(string_convert, encoding= 'utf-8')).decode("utf-8")
|
|
|
|
prev_len_2 = '{:04x}'.format(int(len(converted)/2))
|
|
|
|
|
|
|
|
return str(prev_len_2)+str(converted)
|
|
|
|
|
|
|
|
|
|
|
|
@keyword('Set custom eACL')
|
|
|
|
def set_custom_eacl(private_key: bytes, cid: str, eacl_prefix: str, eacl_slice: str, eacl_postfix: str):
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-08-19 22:31:16 +00:00
|
|
|
logger.info(str(eacl_prefix))
|
|
|
|
logger.info(str(eacl_slice))
|
|
|
|
logger.info(str(eacl_postfix))
|
|
|
|
|
|
|
|
eacl = str(eacl_prefix) + str(eacl_slice) + str(eacl_postfix)
|
|
|
|
logger.info("Custom eACL: %s" % eacl)
|
|
|
|
|
|
|
|
set_eacl(private_key, cid, eacl)
|
|
|
|
return
|
|
|
|
|
|
|
|
|
|
|
|
|
2020-08-13 22:09:00 +00:00
|
|
|
@keyword('Set eACL')
|
|
|
|
def set_eacl(private_key: bytes, cid: str, eacl: str):
|
|
|
|
|
|
|
|
Cmd = f'{CLI_PREFIX}neofs-cli --host {NEOFS_ENDPOINT} --key {binascii.hexlify(private_key).decode()} container set-eacl --cid {cid} --eacl {eacl}'
|
|
|
|
logger.info("Cmd: %s" % Cmd)
|
|
|
|
complProc = subprocess.run(Cmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=150, shell=True)
|
|
|
|
output = complProc.stdout
|
|
|
|
logger.info("Output: %s" % output)
|
|
|
|
|
2020-07-14 00:05:22 +00:00
|
|
|
|
|
|
|
|
|
|
|
@keyword('Get Range')
|
2020-11-18 15:15:57 +00:00
|
|
|
def get_range(private_key: str, cid: str, oid: str, bearer: str, range_cut: str):
|
2020-07-14 00:05:22 +00:00
|
|
|
|
2020-09-01 03:23:17 +00:00
|
|
|
bearer_token = ""
|
2020-11-30 10:33:05 +00:00
|
|
|
if bearer:
|
2020-09-01 03:23:17 +00:00
|
|
|
bearer_token = f"--bearer {bearer}"
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
Cmd = f'neofs-cli --rpc-endpoint {NEOFS_ENDPOINT} --key {binascii.hexlify(private_key).decode()} object get-range --cid {cid} --oid {oid} {bearer_token} {range_cut} '
|
2020-07-14 00:05:22 +00:00
|
|
|
logger.info("Cmd: %s" % Cmd)
|
|
|
|
complProc = subprocess.run(Cmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=150, shell=True)
|
|
|
|
output = complProc.stdout
|
|
|
|
logger.info("Output: %s" % output)
|
|
|
|
|
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
@keyword('Create container')
|
2020-11-18 15:15:57 +00:00
|
|
|
def create_container(private_key: str, basic_acl:str="", rule:str="REP 2 IN X CBF 1 SELECT 2 FROM * AS X"):
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-07-14 00:05:22 +00:00
|
|
|
if basic_acl != "":
|
2020-11-18 15:15:57 +00:00
|
|
|
basic_acl = "--basic-acl " + basic_acl
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
createContainerCmd = f'neofs-cli --rpc-endpoint {NEOFS_ENDPOINT} --key {private_key} container create --policy "{rule}" {basic_acl} --await'
|
2020-07-14 00:05:22 +00:00
|
|
|
logger.info("Cmd: %s" % createContainerCmd)
|
2020-07-01 02:28:31 +00:00
|
|
|
complProc = subprocess.run(createContainerCmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=150, shell=True)
|
|
|
|
output = complProc.stdout
|
|
|
|
logger.info("Output: %s" % output)
|
|
|
|
cid = _parse_cid(output)
|
|
|
|
logger.info("Created container %s with rule '%s'" % (cid, rule))
|
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
#$ ./bin/neofs-cli -c config.yml container create --policy rule.ql --await
|
|
|
|
#container ID: GePis2sDpYqYPh4F8vfGUqoujtNcqdXhipbLx2pKbUwX
|
|
|
|
|
|
|
|
# REP 1 IN X CBF 1 SELECT 2 IN SAME Location FROM * AS X
|
2020-07-01 02:28:31 +00:00
|
|
|
return cid
|
|
|
|
|
|
|
|
|
|
|
|
@keyword('Container Existing')
|
2020-11-18 15:15:57 +00:00
|
|
|
def container_existing(private_key: str, cid: str):
|
|
|
|
Cmd = f'neofs-cli --rpc-endpoint {NEOFS_ENDPOINT} --key {private_key} container list'
|
|
|
|
logger.info("Cmd: %s" % Cmd)
|
2020-07-01 02:28:31 +00:00
|
|
|
complProc = subprocess.run(Cmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=15, shell=True)
|
|
|
|
logger.info("Output: %s" % complProc.stdout)
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
_find_cid(complProc.stdout, cid)
|
|
|
|
return
|
|
|
|
|
|
|
|
|
|
|
|
@keyword('Generate file of bytes')
|
|
|
|
def generate_file_of_bytes(size):
|
|
|
|
"""
|
|
|
|
generate big binary file with the specified size in bytes
|
|
|
|
:param size: the size in bytes, can be declared as 6e+6 for example
|
|
|
|
:return:string filename
|
|
|
|
"""
|
|
|
|
|
|
|
|
size = int(float(size))
|
|
|
|
|
|
|
|
filename = str(uuid.uuid4())
|
|
|
|
with open('%s'%filename, 'wb') as fout:
|
|
|
|
fout.write(os.urandom(size))
|
|
|
|
|
|
|
|
logger.info("Random binary file with size %s bytes has been generated." % str(size))
|
2020-11-30 10:33:05 +00:00
|
|
|
return os.path.abspath(os.getcwd()) + '/' + filename
|
2020-07-01 02:28:31 +00:00
|
|
|
|
|
|
|
|
|
|
|
@keyword('Search object')
|
2020-11-18 15:15:57 +00:00
|
|
|
def search_object(private_key: str, cid: str, keys: str, bearer: str, filters: str, *expected_objects_list ):
|
2020-07-01 02:28:31 +00:00
|
|
|
|
2020-09-01 03:23:17 +00:00
|
|
|
bearer_token = ""
|
2020-11-30 10:33:05 +00:00
|
|
|
if bearer:
|
2020-09-01 03:23:17 +00:00
|
|
|
bearer_token = f"--bearer {bearer}"
|
2020-07-01 02:28:31 +00:00
|
|
|
|
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
if filters:
|
|
|
|
filters = f"--filters {filters}"
|
|
|
|
|
|
|
|
ObjectCmd = f'neofs-cli --rpc-endpoint {NEOFS_ENDPOINT} --key {private_key} object search {keys} --cid {cid} {bearer_token} {filters}'
|
2020-07-01 02:28:31 +00:00
|
|
|
logger.info("Cmd: %s" % ObjectCmd)
|
|
|
|
try:
|
|
|
|
complProc = subprocess.run(ObjectCmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=15, shell=True)
|
|
|
|
|
|
|
|
logger.info("Output: %s" % complProc.stdout)
|
|
|
|
|
2020-09-01 03:23:17 +00:00
|
|
|
if expected_objects_list:
|
2020-11-18 15:15:57 +00:00
|
|
|
found_objects = re.findall(r'(\w{43,44})', complProc.stdout)
|
2020-07-01 02:28:31 +00:00
|
|
|
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
if sorted(found_objects) == sorted(expected_objects_list):
|
|
|
|
logger.info("Found objects list '{}' is equal for expected list '{}'".format(found_objects, expected_objects_list))
|
|
|
|
else:
|
|
|
|
raise Exception("Found object list '{}' is not equal to expected list '{}'".format(found_objects, expected_objects_list))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
except subprocess.CalledProcessError as e:
|
2020-11-30 10:33:05 +00:00
|
|
|
raise Exception("command '{}' return with error (code {}): {}".format(e.cmd, e.returncode, e.output))
|
2020-07-01 02:28:31 +00:00
|
|
|
|
|
|
|
|
|
|
|
@keyword('Verify Head Tombstone')
|
2020-11-18 15:15:57 +00:00
|
|
|
def verify_head_tombstone(private_key: str, cid: str, oid: str):
|
2020-07-01 02:28:31 +00:00
|
|
|
|
|
|
|
ObjectCmd = f'{CLI_PREFIX}neofs-cli --host {NEOFS_ENDPOINT} --key {binascii.hexlify(private_key).decode()} object head --cid {cid} --oid {oid} --full-headers'
|
|
|
|
logger.info("Cmd: %s" % ObjectCmd)
|
|
|
|
try:
|
|
|
|
complProc = subprocess.run(ObjectCmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=15, shell=True)
|
|
|
|
logger.info("Output: %s" % complProc.stdout)
|
|
|
|
|
|
|
|
if re.search(r'Type=Tombstone\s+Value=MARKED', complProc.stdout):
|
|
|
|
logger.info("Tombstone header 'Type=Tombstone Value=MARKED' was parsed from command output")
|
|
|
|
else:
|
|
|
|
raise Exception("Tombstone header 'Type=Tombstone Value=MARKED' was not found in the command output: \t%s" % (complProc.stdout))
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
except subprocess.CalledProcessError as e:
|
|
|
|
raise Exception("command '{}' return with error (code {}): {}".format(e.cmd, e.returncode, e.output))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def _exec_cli_cmd(private_key: bytes, postfix: str):
|
|
|
|
|
|
|
|
# Get linked objects from first
|
|
|
|
ObjectCmd = f'{CLI_PREFIX}neofs-cli --raw --host {NEOFS_ENDPOINT} --key {binascii.hexlify(private_key).decode()} {postfix}'
|
|
|
|
logger.info("Cmd: %s" % ObjectCmd)
|
|
|
|
try:
|
|
|
|
complProc = subprocess.run(ObjectCmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=15, shell=True)
|
|
|
|
logger.info("Output: %s" % complProc.stdout)
|
|
|
|
|
|
|
|
except subprocess.CalledProcessError as e:
|
|
|
|
raise Exception("command '{}' return with error (code {}): {}".format(e.cmd, e.returncode, e.output))
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
return complProc.stdout
|
|
|
|
|
|
|
|
|
|
|
|
@keyword('Verify linked objects')
|
|
|
|
def verify_linked_objects(private_key: bytes, cid: str, oid: str, payload_size: float):
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
payload_size = int(float(payload_size))
|
|
|
|
|
|
|
|
# Get linked objects from first
|
|
|
|
postfix = f'object head --cid {cid} --oid {oid} --full-headers'
|
|
|
|
output = _exec_cli_cmd(private_key, postfix)
|
|
|
|
child_obj_list = []
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
for m in re.finditer(r'Type=Child ID=([\w-]+)', output):
|
|
|
|
child_obj_list.append(m.group(1))
|
|
|
|
|
|
|
|
if not re.search(r'PayloadLength=0', output):
|
|
|
|
raise Exception("Payload is not equal to zero in the parent object %s." % obj)
|
|
|
|
|
|
|
|
if not child_obj_list:
|
|
|
|
raise Exception("Child objects was not found.")
|
|
|
|
else:
|
|
|
|
logger.info("Child objects: %s" % child_obj_list)
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
# HEAD and validate each child object:
|
|
|
|
payload = 0
|
|
|
|
parent_id = "00000000-0000-0000-0000-000000000000"
|
|
|
|
first_obj = None
|
|
|
|
child_obj_list_headers = {}
|
|
|
|
|
|
|
|
for obj in child_obj_list:
|
|
|
|
postfix = f'object head --cid {cid} --oid {obj} --full-headers'
|
|
|
|
output = _exec_cli_cmd(private_key, postfix)
|
|
|
|
child_obj_list_headers[obj] = output
|
|
|
|
if re.search(r'Type=Previous ID=00000000-0000-0000-0000-000000000000', output):
|
|
|
|
first_obj = obj
|
|
|
|
logger.info("First child object %s has been found" % first_obj)
|
|
|
|
|
|
|
|
if not first_obj:
|
|
|
|
raise Exception("Can not find first object with zero Parent ID.")
|
|
|
|
else:
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
_check_linked_object(first_obj, child_obj_list_headers, payload_size, payload, parent_id)
|
|
|
|
|
|
|
|
return child_obj_list_headers.keys()
|
|
|
|
|
|
|
|
def _check_linked_object(obj:str, child_obj_list_headers:dict, payload_size:int, payload:int, parent_id:str):
|
|
|
|
|
|
|
|
output = child_obj_list_headers[obj]
|
|
|
|
logger.info("Verify headers of the child object %s" % obj)
|
|
|
|
|
|
|
|
if not re.search(r'Type=Previous ID=%s' % parent_id, output):
|
|
|
|
raise Exception("Incorrect previos ID %s in the child object %s." % parent_id, obj)
|
|
|
|
else:
|
|
|
|
logger.info("Previous ID is equal for expected: %s" % parent_id)
|
|
|
|
|
|
|
|
m = re.search(r'PayloadLength=(\d+)', output)
|
2020-11-30 10:33:05 +00:00
|
|
|
if m.start() != m.end():
|
2020-07-01 02:28:31 +00:00
|
|
|
payload += int(m.group(1))
|
|
|
|
else:
|
|
|
|
raise Exception("Can not get payload for the object %s." % obj)
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
if payload > payload_size:
|
|
|
|
raise Exception("Payload exceeds expected total payload %s." % payload_size)
|
|
|
|
|
|
|
|
elif payload == payload_size:
|
|
|
|
if not re.search(r'Type=Next ID=00000000-0000-0000-0000-000000000000', output):
|
|
|
|
raise Exception("Incorrect previos ID in the last child object %s." % obj)
|
|
|
|
else:
|
|
|
|
logger.info("Next ID is correct for the final child object: %s" % obj)
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
else:
|
|
|
|
m = re.search(r'Type=Next ID=([\w-]+)', output)
|
2020-11-30 10:33:05 +00:00
|
|
|
if m:
|
2020-07-01 02:28:31 +00:00
|
|
|
# next object should be in the expected list
|
|
|
|
logger.info(m.group(1))
|
|
|
|
if m.group(1) not in child_obj_list_headers.keys():
|
|
|
|
raise Exception(f'Next object {m.group(1)} is not in the expected list: {child_obj_list_headers.keys()}.')
|
|
|
|
else:
|
|
|
|
logger.info(f'Next object {m.group(1)} is in the expected list: {child_obj_list_headers.keys()}.')
|
|
|
|
|
|
|
|
_check_linked_object(m.group(1), child_obj_list_headers, payload_size, payload, obj)
|
|
|
|
|
|
|
|
else:
|
|
|
|
raise Exception("Can not get Next object ID for the object %s." % obj)
|
|
|
|
|
|
|
|
|
|
|
|
@keyword('Head object')
|
2020-11-18 15:15:57 +00:00
|
|
|
def head_object(private_key: str, cid: str, oid: str, bearer: str, user_headers:str=""):
|
2020-07-01 02:28:31 +00:00
|
|
|
options = ""
|
|
|
|
|
2020-09-01 03:23:17 +00:00
|
|
|
bearer_token = ""
|
|
|
|
if bearer:
|
|
|
|
bearer_token = f"--bearer {bearer}"
|
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
ObjectCmd = f'neofs-cli --rpc-endpoint {NEOFS_ENDPOINT} --key {private_key} object head --cid {cid} --oid {oid} {bearer_token} {options}'
|
2020-07-01 02:28:31 +00:00
|
|
|
logger.info("Cmd: %s" % ObjectCmd)
|
|
|
|
try:
|
|
|
|
complProc = subprocess.run(ObjectCmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=15, shell=True)
|
|
|
|
logger.info("Output: %s" % complProc.stdout)
|
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
for key in user_headers.split(","):
|
|
|
|
# user_header = f'Key={key} Val={user_headers_dict[key]}'
|
|
|
|
if re.search(r'(%s)' % key, complProc.stdout):
|
|
|
|
logger.info("User header %s was parsed from command output" % key)
|
2020-07-01 02:28:31 +00:00
|
|
|
else:
|
2020-11-18 15:15:57 +00:00
|
|
|
raise Exception("User header %s was not found in the command output: \t%s" % (key, complProc.stdout))
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-08-13 22:09:00 +00:00
|
|
|
return complProc.stdout
|
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
except subprocess.CalledProcessError as e:
|
|
|
|
raise Exception("command '{}' return with error (code {}): {}".format(e.cmd, e.returncode, e.output))
|
|
|
|
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-08-13 22:09:00 +00:00
|
|
|
|
|
|
|
|
2020-08-19 22:31:16 +00:00
|
|
|
@keyword('Parse Object System Header')
|
|
|
|
def parse_object_system_header(header: str):
|
2020-08-13 22:09:00 +00:00
|
|
|
result_header = dict()
|
|
|
|
|
|
|
|
#SystemHeader
|
2020-08-19 22:31:16 +00:00
|
|
|
logger.info("Input: %s" % header)
|
|
|
|
# ID
|
|
|
|
m = re.search(r'- ID=([a-zA-Z0-9-]+)', header)
|
|
|
|
if m.start() != m.end(): # e.g., if match found something
|
|
|
|
result_header['ID'] = m.group(1)
|
|
|
|
else:
|
|
|
|
raise Exception("no ID was parsed from object header: \t%s" % output)
|
2020-08-13 22:09:00 +00:00
|
|
|
|
2020-08-19 22:31:16 +00:00
|
|
|
# CID
|
|
|
|
m = re.search(r'- CID=([a-zA-Z0-9]+)', header)
|
|
|
|
if m.start() != m.end(): # e.g., if match found something
|
|
|
|
result_header['CID'] = m.group(1)
|
|
|
|
else:
|
|
|
|
raise Exception("no CID was parsed from object header: \t%s" % output)
|
|
|
|
|
|
|
|
# Owner
|
|
|
|
m = re.search(r'- OwnerID=([a-zA-Z0-9]+)', header)
|
|
|
|
if m.start() != m.end(): # e.g., if match found something
|
|
|
|
result_header['OwnerID'] = m.group(1)
|
|
|
|
else:
|
|
|
|
raise Exception("no OwnerID was parsed from object header: \t%s" % output)
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-08-19 22:31:16 +00:00
|
|
|
# Version
|
|
|
|
m = re.search(r'- Version=(\d+)', header)
|
2020-08-13 22:09:00 +00:00
|
|
|
if m.start() != m.end(): # e.g., if match found something
|
2020-08-19 22:31:16 +00:00
|
|
|
result_header['Version'] = m.group(1)
|
2020-08-13 22:09:00 +00:00
|
|
|
else:
|
2020-08-19 22:31:16 +00:00
|
|
|
raise Exception("no Version was parsed from object header: \t%s" % output)
|
|
|
|
|
|
|
|
|
|
|
|
# PayloadLength
|
|
|
|
m = re.search(r'- PayloadLength=(\d+)', header)
|
|
|
|
if m.start() != m.end(): # e.g., if match found something
|
|
|
|
result_header['PayloadLength'] = m.group(1)
|
|
|
|
else:
|
|
|
|
raise Exception("no PayloadLength was parsed from object header: \t%s" % output)
|
|
|
|
|
|
|
|
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-08-19 22:31:16 +00:00
|
|
|
# CreatedAtUnixTime
|
|
|
|
m = re.search(r'- CreatedAt={UnixTime=(\d+)', header)
|
|
|
|
if m.start() != m.end(): # e.g., if match found something
|
|
|
|
result_header['CreatedAtUnixTime'] = m.group(1)
|
|
|
|
else:
|
|
|
|
raise Exception("no CreatedAtUnixTime was parsed from object header: \t%s" % output)
|
|
|
|
|
|
|
|
# CreatedAtEpoch
|
|
|
|
m = re.search(r'- CreatedAt={UnixTime=\d+ Epoch=(\d+)', header)
|
|
|
|
if m.start() != m.end(): # e.g., if match found something
|
|
|
|
result_header['CreatedAtEpoch'] = m.group(1)
|
|
|
|
else:
|
|
|
|
raise Exception("no CreatedAtEpoch was parsed from object header: \t%s" % output)
|
|
|
|
|
|
|
|
logger.info("Result: %s" % result_header)
|
|
|
|
return result_header
|
2020-08-13 22:09:00 +00:00
|
|
|
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-08-19 22:31:16 +00:00
|
|
|
|
|
|
|
@keyword('Parse Object Extended Header')
|
|
|
|
def parse_object_extended_header(header: str):
|
|
|
|
result_header = dict()
|
|
|
|
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-08-19 22:31:16 +00:00
|
|
|
pattern = re.compile(r'- Type=(\w+)\n.+Value=(.+)\n')
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-08-19 22:31:16 +00:00
|
|
|
|
|
|
|
for (f_type, f_val) in re.findall(pattern, header):
|
|
|
|
logger.info("found: %s - %s" % (f_type, f_val))
|
|
|
|
if f_type not in result_header.keys():
|
|
|
|
result_header[f_type] = []
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-08-19 22:31:16 +00:00
|
|
|
result_header[f_type].append(f_val)
|
|
|
|
|
|
|
|
logger.info("Result: %s" % result_header)
|
|
|
|
return result_header
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
|
|
|
|
@keyword('Delete object')
|
2020-11-18 15:15:57 +00:00
|
|
|
def delete_object(private_key: str, cid: str, oid: str, bearer: str):
|
2020-09-01 03:23:17 +00:00
|
|
|
|
|
|
|
bearer_token = ""
|
|
|
|
if bearer:
|
|
|
|
bearer_token = f"--bearer {bearer}"
|
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
ObjectCmd = f'neofs-cli --rpc-endpoint {NEOFS_ENDPOINT} --key {private_key} object delete --cid {cid} --oid {oid} {bearer_token}'
|
2020-07-01 02:28:31 +00:00
|
|
|
try:
|
|
|
|
complProc = subprocess.run(ObjectCmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=15, shell=True)
|
|
|
|
logger.info("Output: %s" % complProc.stdout)
|
|
|
|
except subprocess.CalledProcessError as e:
|
2020-11-30 10:33:05 +00:00
|
|
|
raise Exception("command '{}' return with error (code {}): {}".format(e.cmd, e.returncode, e.output))
|
2020-07-01 02:28:31 +00:00
|
|
|
|
|
|
|
|
|
|
|
@keyword('Get file hash')
|
|
|
|
def get_file_hash(filename):
|
|
|
|
file_hash = _get_file_hash(filename)
|
|
|
|
return file_hash
|
|
|
|
|
|
|
|
|
|
|
|
@keyword('Verify file hash')
|
|
|
|
def verify_file_hash(filename, expected_hash):
|
|
|
|
file_hash = _get_file_hash(filename)
|
|
|
|
if file_hash == expected_hash:
|
|
|
|
logger.info("Hash is equal to expected: %s" % file_hash)
|
|
|
|
else:
|
|
|
|
raise Exception("File hash '{}' is not equal to {}".format(file_hash, expected_hash))
|
|
|
|
|
|
|
|
|
|
|
|
@keyword('Create storage group')
|
|
|
|
def create_storage_group(private_key: bytes, cid: str, *objects_list):
|
|
|
|
objects = ""
|
|
|
|
|
|
|
|
for oid in objects_list:
|
|
|
|
objects = f'{objects} --oid {oid}'
|
|
|
|
|
|
|
|
ObjectCmd = f'{CLI_PREFIX}neofs-cli --host {NEOFS_ENDPOINT} --key {binascii.hexlify(private_key).decode()} sg put --cid {cid} {objects}'
|
|
|
|
complProc = subprocess.run(ObjectCmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=15, shell=True)
|
|
|
|
logger.info("Output: %s" % complProc.stdout)
|
|
|
|
sgid = _parse_oid(complProc.stdout)
|
|
|
|
return sgid
|
|
|
|
|
|
|
|
|
|
|
|
@keyword('Get storage group')
|
|
|
|
def get_storage_group(private_key: bytes, cid: str, sgid: str):
|
|
|
|
ObjectCmd = f'{CLI_PREFIX}neofs-cli --host {NEOFS_ENDPOINT} --key {binascii.hexlify(private_key).decode()} sg get --cid {cid} --sgid {sgid}'
|
|
|
|
logger.info("Cmd: %s" % ObjectCmd)
|
|
|
|
try:
|
|
|
|
complProc = subprocess.run(ObjectCmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=15, shell=True)
|
|
|
|
logger.info("Output: %s" % complProc.stdout)
|
|
|
|
except subprocess.CalledProcessError as e:
|
|
|
|
raise Exception("command '{}' return with error (code {}): {}".format(e.cmd, e.returncode, e.output))
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
|
|
|
|
@keyword('Cleanup File')
|
|
|
|
# remove temp files
|
|
|
|
def cleanup_file(filename: str):
|
|
|
|
if os.path.isfile(filename):
|
|
|
|
try:
|
|
|
|
os.remove(filename)
|
2020-11-30 10:33:05 +00:00
|
|
|
except OSError as e:
|
2020-07-01 02:28:31 +00:00
|
|
|
raise Exception("Error: '%s' - %s." % (e.filename, e.strerror))
|
2020-11-30 10:33:05 +00:00
|
|
|
else:
|
2020-07-01 02:28:31 +00:00
|
|
|
raise Exception("Error: '%s' file not found" % filename)
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
logger.info("File '%s' has been deleted." % filename)
|
|
|
|
|
|
|
|
|
|
|
|
@keyword('Put object to NeoFS')
|
2020-11-18 15:15:57 +00:00
|
|
|
def put_object(private_key: str, path: str, cid: str, bearer: str, user_headers: str):
|
2020-07-01 02:28:31 +00:00
|
|
|
logger.info("Going to put the object")
|
2020-09-01 03:23:17 +00:00
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
if user_headers:
|
|
|
|
user_headers = f"--attributes {user_headers}"
|
2020-07-01 02:28:31 +00:00
|
|
|
|
2020-09-01 03:23:17 +00:00
|
|
|
if bearer:
|
2020-11-18 15:15:57 +00:00
|
|
|
bearer = f"--bearer {bearer}"
|
2020-07-01 02:28:31 +00:00
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
putObjectCmd = f'neofs-cli --rpc-endpoint {NEOFS_ENDPOINT} --key {private_key} object put --file {path} --cid {cid} {bearer} {user_headers}'
|
2020-07-01 02:28:31 +00:00
|
|
|
logger.info("Cmd: %s" % putObjectCmd)
|
|
|
|
complProc = subprocess.run(putObjectCmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=60, shell=True)
|
|
|
|
logger.info("Output: %s" % complProc.stdout)
|
|
|
|
oid = _parse_oid(complProc.stdout)
|
|
|
|
return oid
|
|
|
|
|
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
@keyword('Get Range Hash')
|
|
|
|
def get_range_hash(private_key: str, cid: str, oid: str, bearer_token: str, range_cut: str):
|
2020-11-30 10:33:05 +00:00
|
|
|
|
|
|
|
if bearer_token:
|
2020-09-01 03:23:17 +00:00
|
|
|
bearer_token = f"--bearer {bearer}"
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
ObjectCmd = f'neofs-cli --rpc-endpoint {NEOFS_ENDPOINT} --key {private_key} object hash --cid {cid} --oid {oid} --range {range_cut} {bearer_token}'
|
2020-09-01 03:23:17 +00:00
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
logger.info("Cmd: %s" % ObjectCmd)
|
|
|
|
try:
|
|
|
|
complProc = subprocess.run(ObjectCmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=60, shell=True)
|
|
|
|
logger.info("Output: %s" % complProc.stdout)
|
|
|
|
except subprocess.CalledProcessError as e:
|
|
|
|
raise Exception("command '{}' return with error (code {}): {}".format(e.cmd, e.returncode, e.output))
|
2020-09-01 03:23:17 +00:00
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
@keyword('Get object from NeoFS')
|
|
|
|
def get_object(private_key: str, cid: str, oid: str, bearer_token: str, read_object: str):
|
2020-09-01 03:23:17 +00:00
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
if bearer_token:
|
|
|
|
bearer_token = f"--bearer {bearer_token}"
|
|
|
|
|
|
|
|
ObjectCmd = f'neofs-cli --rpc-endpoint {NEOFS_ENDPOINT} --key {private_key} object get --cid {cid} --oid {oid} --file {read_object} {bearer_token}'
|
2020-07-01 02:28:31 +00:00
|
|
|
|
|
|
|
logger.info("Cmd: %s" % ObjectCmd)
|
|
|
|
try:
|
|
|
|
complProc = subprocess.run(ObjectCmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=60, shell=True)
|
|
|
|
logger.info("Output: %s" % complProc.stdout)
|
|
|
|
except subprocess.CalledProcessError as e:
|
|
|
|
raise Exception("command '{}' return with error (code {}): {}".format(e.cmd, e.returncode, e.output))
|
|
|
|
|
|
|
|
|
|
|
|
def _get_file_hash(filename):
|
|
|
|
blocksize = 65536
|
|
|
|
hash = hashlib.md5()
|
|
|
|
with open(filename, "rb") as f:
|
|
|
|
for block in iter(lambda: f.read(blocksize), b""):
|
|
|
|
hash.update(block)
|
|
|
|
logger.info("Hash: %s" % hash.hexdigest())
|
|
|
|
|
|
|
|
return hash.hexdigest()
|
|
|
|
|
|
|
|
def _find_cid(output: str, cid: str):
|
|
|
|
"""
|
|
|
|
This function parses CID from given CLI output.
|
|
|
|
Parameters:
|
|
|
|
- output: a string with command run output
|
|
|
|
"""
|
|
|
|
|
|
|
|
if re.search(r'(%s)' % cid, output):
|
|
|
|
logger.info("CID %s was parsed from command output: \t%s" % (cid, output))
|
|
|
|
else:
|
|
|
|
raise Exception("no CID %s was parsed from command output: \t%s" % (cid, output))
|
|
|
|
return cid
|
|
|
|
|
|
|
|
def _parse_oid(output: str):
|
|
|
|
"""
|
|
|
|
This function parses OID from given CLI output.
|
|
|
|
Parameters:
|
|
|
|
- output: a string with command run output
|
|
|
|
"""
|
|
|
|
m = re.search(r'ID: ([a-zA-Z0-9-]+)', output)
|
|
|
|
if m.start() != m.end(): # e.g., if match found something
|
|
|
|
oid = m.group(1)
|
|
|
|
else:
|
|
|
|
raise Exception("no OID was parsed from command output: \t%s" % output)
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
return oid
|
|
|
|
|
|
|
|
def _parse_cid(output: str):
|
|
|
|
"""
|
|
|
|
This function parses CID from given CLI output.
|
|
|
|
Parameters:
|
|
|
|
- output: a string with command run output
|
|
|
|
"""
|
2020-11-18 15:15:57 +00:00
|
|
|
m = re.search(r'container ID: (\w+)', output)
|
|
|
|
if not m.start() != m.end(): # e.g., if match found something
|
2020-07-01 02:28:31 +00:00
|
|
|
raise Exception("no CID was parsed from command output: \t%s" % (output))
|
2020-11-18 15:15:57 +00:00
|
|
|
cid = m.group(1)
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-07-01 02:28:31 +00:00
|
|
|
return cid
|
|
|
|
|
|
|
|
def _get_storage_nodes(private_key: bytes):
|
2020-11-18 15:15:57 +00:00
|
|
|
storage_nodes = ['s01.neofs.devenv:8080', 's02.neofs.devenv:8080','s03.neofs.devenv:8080','s04.neofs.devenv:8080']
|
|
|
|
#NetmapCmd = f'{CLI_PREFIX}neofs-cli --host {NEOFS_ENDPOINT} --key {binascii.hexlify(private_key).decode()} status netmap'
|
|
|
|
#complProc = subprocess.run(NetmapCmd, check=True, universal_newlines=True,
|
|
|
|
# stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=15, shell=True)
|
|
|
|
#output = complProc.stdout
|
|
|
|
#logger.info("Netmap: %s" % output)
|
|
|
|
#for m in re.finditer(r'"address":"/ip4/(\d+\.\d+\.\d+\.\d+)/tcp/(\d+)"', output):
|
|
|
|
# storage_nodes.append(m.group(1)+":"+m.group(2))
|
2020-11-30 10:33:05 +00:00
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
#if not storage_nodes:
|
|
|
|
# raise Exception("Storage nodes was not found.")
|
|
|
|
|
|
|
|
|
|
|
|
# Will be fixed when netmap will be added to cli
|
2020-07-01 02:28:31 +00:00
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
#storage_nodes.append()
|
2020-07-01 02:28:31 +00:00
|
|
|
logger.info("Storage nodes: %s" % storage_nodes)
|
|
|
|
return storage_nodes
|
|
|
|
|
|
|
|
|
2020-11-18 15:15:57 +00:00
|
|
|
def _search_object(node:str, private_key: str, cid:str, oid: str):
|
|
|
|
# --filters objectID={oid}
|
|
|
|
Cmd = f'{CLI_PREFIX}neofs-cli --rpc-endpoint {node} --key {private_key} --ttl 1 object search --root --cid {cid} '
|
2020-07-01 02:28:31 +00:00
|
|
|
|
2020-09-01 03:23:17 +00:00
|
|
|
try:
|
|
|
|
logger.info(Cmd)
|
|
|
|
complProc = subprocess.run(Cmd, check=True, universal_newlines=True,
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE, timeout=15, shell=True)
|
|
|
|
logger.info("Output: %s" % complProc.stdout)
|
2020-11-18 15:15:57 +00:00
|
|
|
|
|
|
|
if re.search(r'%s' % oid, complProc.stdout):
|
|
|
|
return oid
|
|
|
|
else:
|
|
|
|
logger.info("Object is not found.")
|
2020-09-01 03:23:17 +00:00
|
|
|
|
|
|
|
except subprocess.CalledProcessError as e:
|
2020-11-18 15:15:57 +00:00
|
|
|
if re.search(r'local node is outside of object placement', e.output):
|
|
|
|
logger.info("Server is not presented in container.")
|
2020-09-01 03:23:17 +00:00
|
|
|
else:
|
|
|
|
raise Exception("command '{}' return with error (code {}): {}".format(e.cmd, e.returncode, e.output))
|
|
|
|
|
|
|
|
|
2020-11-30 10:33:05 +00:00
|
|
|
|