mirror of
https://github.com/ceph/ceph
synced 2025-01-10 21:20:46 +00:00
bdcc94a1d1
To be able to catch problems with python2 *and* python3, run flake8 with both versions. From the flake8 homepage: It is very important to install Flake8 on the correct version of Python for your needs. If you want Flake8 to properly parse new language features in Python 3.5 (for example), you need it to be installed on 3.5 for Flake8 to understand those features. In many ways, Flake8 is tied to the version of Python on which it runs. Also fix the problems with python3 on the way. Note: This requires now the six module for teuthology. But this is already an install_require in teuthology itself. Signed-off-by: Thomas Bechtold <tbechtold@suse.com>
224 lines
7.5 KiB
Python
224 lines
7.5 KiB
Python
"""
|
|
Rados benchmarking sweep
|
|
"""
|
|
import contextlib
|
|
import logging
|
|
import re
|
|
|
|
from cStringIO import StringIO
|
|
from itertools import product
|
|
|
|
from teuthology.orchestra import run
|
|
from teuthology import misc as teuthology
|
|
|
|
import six
|
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
|
|
@contextlib.contextmanager
|
|
def task(ctx, config):
|
|
"""
|
|
Execute a radosbench parameter sweep
|
|
|
|
Puts radosbench in a loop, taking values from the given config at each
|
|
iteration. If given, the min and max values below create a range, e.g.
|
|
min_replicas=1 and max_replicas=3 implies executing with 1-3 replicas.
|
|
|
|
Parameters:
|
|
|
|
clients: [client list]
|
|
time: seconds to run (default=120)
|
|
sizes: [list of object sizes] (default=[4M])
|
|
mode: <write|read|seq> (default=write)
|
|
repetitions: execute the same configuration multiple times (default=1)
|
|
min_num_replicas: minimum number of replicas to use (default = 3)
|
|
max_num_replicas: maximum number of replicas to use (default = 3)
|
|
min_num_osds: the minimum number of OSDs in a pool (default=all)
|
|
max_num_osds: the maximum number of OSDs in a pool (default=all)
|
|
file: name of CSV-formatted output file (default='radosbench.csv')
|
|
columns: columns to include (default=all)
|
|
- rep: execution number (takes values from 'repetitions')
|
|
- num_osd: number of osds for pool
|
|
- num_replica: number of replicas
|
|
- avg_throughput: throughput
|
|
- avg_latency: latency
|
|
- stdev_throughput:
|
|
- stdev_latency:
|
|
|
|
Example:
|
|
- radsobenchsweep:
|
|
columns: [rep, num_osd, num_replica, avg_throughput, stdev_throughput]
|
|
"""
|
|
log.info('Beginning radosbenchsweep...')
|
|
assert isinstance(config, dict), 'expecting dictionary for configuration'
|
|
|
|
# get and validate config values
|
|
# {
|
|
|
|
# only one client supported for now
|
|
if len(config.get('clients', [])) != 1:
|
|
raise Exception("Only one client can be specified")
|
|
|
|
# only write mode
|
|
if config.get('mode', 'write') != 'write':
|
|
raise Exception("Only 'write' mode supported for now.")
|
|
|
|
# OSDs
|
|
total_osds_in_cluster = teuthology.num_instances_of_type(ctx.cluster, 'osd')
|
|
min_num_osds = config.get('min_num_osds', total_osds_in_cluster)
|
|
max_num_osds = config.get('max_num_osds', total_osds_in_cluster)
|
|
|
|
if max_num_osds > total_osds_in_cluster:
|
|
raise Exception('max_num_osds cannot be greater than total in cluster')
|
|
if min_num_osds < 1:
|
|
raise Exception('min_num_osds cannot be less than 1')
|
|
if min_num_osds > max_num_osds:
|
|
raise Exception('min_num_osds cannot be greater than max_num_osd')
|
|
osds = range(0, (total_osds_in_cluster + 1))
|
|
|
|
# replicas
|
|
min_num_replicas = config.get('min_num_replicas', 3)
|
|
max_num_replicas = config.get('max_num_replicas', 3)
|
|
|
|
if min_num_replicas < 1:
|
|
raise Exception('min_num_replicas cannot be less than 1')
|
|
if min_num_replicas > max_num_replicas:
|
|
raise Exception('min_num_replicas cannot be greater than max_replicas')
|
|
if max_num_replicas > max_num_osds:
|
|
raise Exception('max_num_replicas cannot be greater than max_num_osds')
|
|
replicas = range(min_num_replicas, (max_num_replicas + 1))
|
|
|
|
# object size
|
|
sizes = config.get('size', [4 << 20])
|
|
|
|
# repetitions
|
|
reps = range(config.get('repetitions', 1))
|
|
|
|
# file
|
|
fname = config.get('file', 'radosbench.csv')
|
|
f = open('{}/{}'.format(ctx.archive, fname), 'w')
|
|
f.write(get_csv_header(config) + '\n')
|
|
# }
|
|
|
|
# set default pools size=1 to avoid 'unhealthy' issues
|
|
ctx.manager.set_pool_property('data', 'size', 1)
|
|
ctx.manager.set_pool_property('metadata', 'size', 1)
|
|
ctx.manager.set_pool_property('rbd', 'size', 1)
|
|
|
|
current_osds_out = 0
|
|
|
|
# sweep through all parameters
|
|
for osds_out, size, replica, rep in product(osds, sizes, replicas, reps):
|
|
|
|
osds_in = total_osds_in_cluster - osds_out
|
|
|
|
if osds_in == 0:
|
|
# we're done
|
|
break
|
|
|
|
if current_osds_out != osds_out:
|
|
# take an osd out
|
|
ctx.manager.raw_cluster_cmd(
|
|
'osd', 'reweight', str(osds_out-1), '0.0')
|
|
wait_until_healthy(ctx, config)
|
|
current_osds_out = osds_out
|
|
|
|
if osds_in not in range(min_num_osds, (max_num_osds + 1)):
|
|
# no need to execute with a number of osds that wasn't requested
|
|
continue
|
|
|
|
if osds_in < replica:
|
|
# cannot execute with more replicas than available osds
|
|
continue
|
|
|
|
run_radosbench(ctx, config, f, osds_in, size, replica, rep)
|
|
|
|
f.close()
|
|
|
|
yield
|
|
|
|
|
|
def get_csv_header(conf):
|
|
all_columns = [
|
|
'rep', 'num_osd', 'num_replica', 'avg_throughput',
|
|
'avg_latency', 'stdev_throughput', 'stdev_latency'
|
|
]
|
|
given_columns = conf.get('columns', None)
|
|
if given_columns and len(given_columns) != 0:
|
|
for column in given_columns:
|
|
if column not in all_columns:
|
|
raise Exception('Unknown column ' + column)
|
|
return ','.join(conf['columns'])
|
|
else:
|
|
conf['columns'] = all_columns
|
|
return ','.join(all_columns)
|
|
|
|
|
|
def run_radosbench(ctx, config, f, num_osds, size, replica, rep):
|
|
pool = ctx.manager.create_pool_with_unique_name()
|
|
|
|
ctx.manager.set_pool_property(pool, 'size', replica)
|
|
|
|
wait_until_healthy(ctx, config)
|
|
|
|
log.info('Executing with parameters: ')
|
|
log.info(' num_osd =' + str(num_osds))
|
|
log.info(' size =' + str(size))
|
|
log.info(' num_replicas =' + str(replica))
|
|
log.info(' repetition =' + str(rep))
|
|
|
|
for role in config.get('clients', ['client.0']):
|
|
assert isinstance(role, six.string_types)
|
|
PREFIX = 'client.'
|
|
assert role.startswith(PREFIX)
|
|
id_ = role[len(PREFIX):]
|
|
(remote,) = ctx.cluster.only(role).remotes.keys()
|
|
|
|
proc = remote.run(
|
|
args=[
|
|
'adjust-ulimits',
|
|
'ceph-coverage',
|
|
'{}/archive/coverage'.format(teuthology.get_testdir(ctx)),
|
|
'rados',
|
|
'--no-log-to-stderr',
|
|
'--name', role,
|
|
'-b', str(size),
|
|
'-p', pool,
|
|
'bench', str(config.get('time', 120)), 'write',
|
|
],
|
|
logger=log.getChild('radosbench.{id}'.format(id=id_)),
|
|
stdin=run.PIPE,
|
|
stdout=StringIO(),
|
|
wait=False
|
|
)
|
|
|
|
# parse output to get summary and format it as CSV
|
|
proc.wait()
|
|
out = proc.stdout.getvalue()
|
|
all_values = {
|
|
'stdev_throughput': re.sub(r'Stddev Bandwidth: ', '', re.search(
|
|
r'Stddev Bandwidth:.*', out).group(0)),
|
|
'stdev_latency': re.sub(r'Stddev Latency: ', '', re.search(
|
|
r'Stddev Latency:.*', out).group(0)),
|
|
'avg_throughput': re.sub(r'Bandwidth \(MB/sec\): ', '', re.search(
|
|
r'Bandwidth \(MB/sec\):.*', out).group(0)),
|
|
'avg_latency': re.sub(r'Average Latency: ', '', re.search(
|
|
r'Average Latency:.*', out).group(0)),
|
|
'rep': str(rep),
|
|
'num_osd': str(num_osds),
|
|
'num_replica': str(replica)
|
|
}
|
|
values_to_write = []
|
|
for column in config['columns']:
|
|
values_to_write.extend([all_values[column]])
|
|
f.write(','.join(values_to_write) + '\n')
|
|
|
|
ctx.manager.remove_pool(pool)
|
|
|
|
|
|
def wait_until_healthy(ctx, config):
|
|
first_mon = teuthology.get_first_mon(ctx, config)
|
|
(mon_remote,) = ctx.cluster.only(first_mon).remotes.keys()
|
|
teuthology.wait_until_healthy(ctx, mon_remote)
|