Parsing k6 results + dataclass for K6 results

Signed-off-by: a.lipay <a.lipay@yadro.com>
This commit is contained in:
a.lipay 2022-10-11 01:09:54 +03:00 committed by Alipayy
parent e63db788c5
commit 7ab737b595

View file

@ -1,3 +1,4 @@
import re
from contextlib import contextmanager from contextlib import contextmanager
from dataclasses import dataclass from dataclasses import dataclass
from time import sleep from time import sleep
@ -16,13 +17,28 @@ class LoadParams:
containers_count: int containers_count: int
out_file: str out_file: str
obj_count: int obj_count: int
writers_percent: int writers: int
readers: int
deleters: int
load_time: int load_time: int
clients_count: int
load_type: str load_type: str
endpoint: str endpoint: str
@dataclass
class LoadResults:
latency_write_min: float = 0.0
latency_write_max: float = 0.0
latency_write_med: float = 0.0
latency_write_avg: float = 0.0
latency_read_min: float = 0.0
latency_read_max: float = 0.0
latency_read_med: float = 0.0
latency_read_avg: float = 0.0
read_ops: float = 0.0
write_ops: float = 0.0
class K6: class K6:
def __init__(self, load_params: LoadParams, host_client: HostClient): def __init__(self, load_params: LoadParams, host_client: HostClient):
@ -54,8 +70,8 @@ class K6:
f"{self.k6_dir}/scenarios/preset/preset_grpc.py " f"{self.k6_dir}/scenarios/preset/preset_grpc.py "
f"--size {self.load_params.obj_size} " f"--size {self.load_params.obj_size} "
f"--containers {self.load_params.containers_count} " f"--containers {self.load_params.containers_count} "
f"--out {self.load_params.load_type}_{self.load_params.out_file} " f"--out {self.k6_dir}/{self.load_params.load_type}_{self.load_params.out_file} "
f"--endpoint {self.load_params.endpoint} " f"--endpoint {self.load_params.endpoint.split(',')[0]} "
f"--preload_obj {self.load_params.obj_count} " f"--preload_obj {self.load_params.obj_count} "
) )
terminal = self.host_client.exec(command) terminal = self.host_client.exec(command)
@ -64,13 +80,14 @@ class K6:
command = ( command = (
f"{self.k6_dir}/scenarios/preset/preset_s3.py --size {self.load_params.obj_size} " f"{self.k6_dir}/scenarios/preset/preset_s3.py --size {self.load_params.obj_size} "
f"--buckets {self.load_params.containers_count} " f"--buckets {self.load_params.containers_count} "
f"--out {self.load_params.load_type}_{self.load_params.out_file} " f"--out {self.k6_dir}/{self.load_params.load_type}_{self.load_params.out_file} "
f"--endpoint {self.load_params.endpoint} " f"--endpoint {self.load_params.endpoint.split(',')[0]} "
f"--preload_obj {self.load_params.obj_count} " f"--preload_obj {self.load_params.obj_count} "
f"--location load-1-1" f"--location load-1-1"
) )
terminal = self.host_client.exec(command) terminal = self.host_client.exec(command)
return terminal.stdout.strip("\n") return terminal.stdout.strip("\n")
else:
raise AssertionError("Wrong K6 load type") raise AssertionError("Wrong K6 load type")
@allure.step("Start K6 on initiator") @allure.step("Start K6 on initiator")
@ -78,11 +95,14 @@ class K6:
self._k6_dir = self.k6_dir self._k6_dir = self.k6_dir
command = ( command = (
f"{self.k6_dir}/k6 run -e " f"{self.k6_dir}/k6 run "
f"PROFILE={self.load_params.writers_percent}:{self.load_params.load_time} " f"-e DURATION={self.load_params.load_time} "
f"-e WRITE_OBJ_SIZE={self.load_params.obj_size} " f"-e WRITE_OBJ_SIZE={self.load_params.obj_size} "
f"-e CLIENTS={self.load_params.clients_count} -e NODES={self.load_params.endpoint} " f"-e WRITERS={self.load_params.writers} -e READERS={self.load_params.readers} "
f"-e PREGEN_JSON={self.k6_dir}/{self.load_params.load_type}_{self.load_params.out_file} " f"-e DELETERS={self.load_params.deleters} "
f"-e {self.load_params.load_type.upper()}_ENDPOINTS={self.load_params.endpoint} "
f"-e PREGEN_JSON={self.k6_dir}/"
f"{self.load_params.load_type}_{self.load_params.out_file} "
f"{self.k6_dir}/scenarios/{self.load_params.load_type}.js" f"{self.k6_dir}/scenarios/{self.load_params.load_type}.js"
) )
self._k6_process = RemoteProcess.create(command, self.host_client) self._k6_process = RemoteProcess.create(command, self.host_client)
@ -153,6 +173,30 @@ class K6:
def is_finished(self) -> bool: def is_finished(self) -> bool:
return not self._k6_process.running() return not self._k6_process.running()
def parsing_results(self) -> LoadResults:
output = self._k6_process.stdout(full=True).replace("\n", "")
metric_regex_map = {
"latency_write_min": r"neofs_obj_put_duration.*?min=(?P<latency_write_min>\d*\.\d*)",
"latency_write_max": r"neofs_obj_put_duration.*?max=(?P<latency_write_max>\d*\.\d*)",
"latency_write_med": r"neofs_obj_put_duration.*?med=(?P<latency_write_med>\d*\.\d*)",
"latency_write_avg": r"neofs_obj_put_duration.*?avg=(?P<latency_write_avg>\d*\.\d*)",
"write_ops": r"neofs_obj_put_total\W*\d*\W*(?P<write_ops>\d*\.\d*)",
"latency_read_min": r"neofs_obj_get_duration.*?min=(?P<latency_read_min>\d*\.\d*)",
"latency_read_max": r"neofs_obj_get_duration.*?max=(?P<latency_read_max>\d*\.\d*)",
"latency_read_med": r"neofs_obj_get_duration.*?med=(?P<latency_read_med>\d*\.\d*)",
"latency_read_avg": r"neofs_obj_get_duration.*?avg=(?P<latency_read_avg>\d*\.\d*)",
"read_ops": r"neofs_obj_get_total\W*\d*\W*(?P<read_ops>\d*\.\d*)",
}
metric_values = {}
for metric_name, metric_regex in metric_regex_map.items():
match = re.search(metric_regex, output)
if match:
metric_values[metric_name] = float(match.group(metric_name))
continue
metric_values[metric_name] = 0.0
load_result = LoadResults(**metric_values)
return load_result
@allure.step("Try to stop K6 with SIGTERM") @allure.step("Try to stop K6 with SIGTERM")
def _stop_k6(self) -> None: def _stop_k6(self) -> None:
for __attempt in range(self._k6_stop_attempts): for __attempt in range(self._k6_stop_attempts):