from cStringIO import StringIO import argparse import contextlib import errno import logging import os import shutil import subprocess import sys import tempfile from teuthology import misc as teuthology from teuthology import contextutil from teuthology.parallel import parallel from ..orchestra import run log = logging.getLogger(__name__) class DaemonState(object): def __init__(self, remote, role, id_, *command_args, **command_kwargs): self.remote = remote self.command_args = command_args self.command_kwargs = command_kwargs self.role = role self.id_ = id_ self.log = command_kwargs.get('logger', log) self.proc = None def stop(self): """ Note: this can raise a run.CommandFailedError, run.CommandCrashedError, or run.ConnectionLostError. """ if not self.running(): self.log.error('tried to stop a non-running daemon') return self.proc.stdin.close() self.log.debug('waiting for process to exit') run.wait([self.proc]) self.proc = None self.log.info('Stopped') def restart(self): self.log.info('Restarting') if self.proc is not None: self.log.debug('stopping old one...') self.stop() self.proc = self.remote.run(*self.command_args, **self.command_kwargs) self.log.info('Started') def running(self): return self.proc is not None class CephState(object): def __init__(self): self.daemons = {} def add_daemon(self, remote, role, id_, *args, **kwargs): if role not in self.daemons: self.daemons[role] = {} if id_ in self.daemons[role]: self.daemons[role][id_].stop() self.daemons[role][id_] = None self.daemons[role][id_] = DaemonState(remote, role, id_, *args, **kwargs) self.daemons[role][id_].restart() def get_daemon(self, role, id_): if role not in self.daemons: return None return self.daemons[role].get(str(id_), None) def iter_daemons_of_role(self, role): return self.daemons.get(role, {}).values() @contextlib.contextmanager def ceph_log(ctx, config): log.info('Creating log directories...') run.wait( ctx.cluster.run( args=[ 'install', '-d', '-m0755', '--', '/tmp/cephtest/archive/log', '/tmp/cephtest/archive/log/valgrind', '/tmp/cephtest/archive/profiling-logger', ], wait=False, ) ) try: yield finally: if ctx.archive is not None: log.info('Compressing logs...') run.wait( ctx.cluster.run( args=[ 'find', '/tmp/cephtest/archive/log', '-name', '*.log', '-print0', run.Raw('|'), 'xargs', '-0', '--no-run-if-empty', '--', 'gzip', '--', ], wait=False, ), ) # log file transfer is done by the generic archive data # handling @contextlib.contextmanager def ship_utilities(ctx, config): assert config is None FILES = ['daemon-helper', 'enable-coredump', 'chdir-coredump'] for filename in FILES: log.info('Shipping %r...', filename) src = os.path.join(os.path.dirname(__file__), filename) dst = os.path.join('/tmp/cephtest', filename) with file(src, 'rb') as f: for rem in ctx.cluster.remotes.iterkeys(): teuthology.write_file( remote=rem, path=dst, data=f, ) f.seek(0) rem.run( args=[ 'chmod', 'a=rx', '--', dst, ], ) try: yield finally: log.info('Removing shipped files: %s...', ' '.join(FILES)) filenames = ( os.path.join('/tmp/cephtest', filename) for filename in FILES ) run.wait( ctx.cluster.run( args=[ 'rm', '-rf', '--', ] + list(filenames), wait=False, ), ) def _download_binaries(remote, ceph_bindir_url): remote.run( args=[ 'install', '-d', '-m0755', '--', '/tmp/cephtest/binary', run.Raw('&&'), 'uname', '-m', run.Raw('|'), 'sed', '-e', 's/^/ceph./; s/$/.tgz/', run.Raw('|'), 'wget', '-nv', '-O-', '--base={url}'.format(url=ceph_bindir_url), # need to use --input-file to make wget respect --base '--input-file=-', run.Raw('|'), 'tar', '-xzf', '-', '-C', '/tmp/cephtest/binary', ], ) @contextlib.contextmanager def binaries(ctx, config): path = config.get('path') tmpdir = None if path is None: # fetch from gitbuilder gitbuilder log.info('Fetching and unpacking ceph binaries from gitbuilder...') sha1, ceph_bindir_url = teuthology.get_ceph_binary_url( package='ceph', branch=config.get('branch'), tag=config.get('tag'), sha1=config.get('sha1'), flavor=config.get('flavor'), format=config.get('format'), dist=config.get('dist'), arch=config.get('arch'), ) ctx.summary['ceph-sha1'] = sha1 if ctx.archive is not None: with file(os.path.join(ctx.archive, 'ceph-sha1'), 'w') as f: f.write(sha1 + '\n') with parallel() as p: for remote in ctx.cluster.remotes.iterkeys(): p.spawn(_download_binaries, remote, ceph_bindir_url) else: with tempfile.TemporaryFile(prefix='teuthology-tarball-', suffix='.tgz') as tar_fp: tmpdir = tempfile.mkdtemp(prefix='teuthology-tarball-') try: log.info('Installing %s to %s...' % (path, tmpdir)) subprocess.check_call( args=[ 'make', 'install', 'DESTDIR={tmpdir}'.format(tmpdir=tmpdir), ], cwd=path, ) try: os.symlink('.', os.path.join(tmpdir, 'usr', 'local')) except OSError as e: if e.errno == errno.EEXIST: pass else: raise log.info('Building ceph binary tarball from %s...', tmpdir) subprocess.check_call( args=[ 'tar', 'cz', '.', ], cwd=tmpdir, stdout=tar_fp, ) finally: shutil.rmtree(tmpdir, ignore_errors=True) log.info('Pushing tarball...') tar_fp.seek(0) writes = ctx.cluster.run( args=[ 'install', '-d', '-m0755', '--', '/tmp/cephtest/binary', run.Raw('&&'), 'tar', '-xzf', '-', '-C', '/tmp/cephtest/binary' ], stdin=run.PIPE, wait=False, ) teuthology.feed_many_stdins_and_close(tar_fp, writes) run.wait(writes) try: yield finally: log.info('Removing ceph binaries...') run.wait( ctx.cluster.run( args=[ 'rm', '-rf', '--', '/tmp/cephtest/binary', ], wait=False, ), ) def assign_devs(roles, devs): return dict(zip(roles, devs)) @contextlib.contextmanager def valgrind_post(ctx, config): try: yield finally: lookup_procs = list() val_path = '/tmp/cephtest/archive/log/valgrind' log.info('Checking for errors in any valgrind logs...'); for remote in ctx.cluster.remotes.iterkeys(): #look at valgrind logs for each node proc = remote.run( args=[ 'grep', "", run.Raw(val_path), run.Raw('|'), 'egrep', '-v', '-q', '(PossiblyLost|DefinitelyLost)'], wait = False, check_status=False ) lookup_procs.append((proc, remote)) valgrind_exception = None for (proc, remote) in lookup_procs: result = proc.exitstatus.get() if result != 1: valgrind_exception = Exception("saw valgrind issues in {node}".format(node=remote.name)) if valgrind_exception is not None: raise valgrind_exception @contextlib.contextmanager def cluster(ctx, config): log.info('Creating ceph cluster...') run.wait( ctx.cluster.run( args=[ 'install', '-d', '-m0755', '--', '/tmp/cephtest/data', ], wait=False, ) ) log.info('Generating config...') remotes_and_roles = ctx.cluster.remotes.items() roles = [roles for (remote, roles) in remotes_and_roles] ips = [host for (host, port) in (remote.ssh.get_transport().getpeername() for (remote, roles) in remotes_and_roles)] conf = teuthology.skeleton_config(roles=roles, ips=ips) for section, keys in config['conf'].iteritems(): for key, value in keys.iteritems(): log.info("[%s] %s = %s" % (section, key, value)) if section not in conf: conf[section] = {} conf[section][key] = value ctx.ceph = argparse.Namespace() ctx.ceph.conf = conf log.info('Writing configs...') conf_fp = StringIO() conf.write(conf_fp) conf_fp.seek(0) writes = ctx.cluster.run( args=[ 'python', '-c', 'import shutil, sys; shutil.copyfileobj(sys.stdin, file(sys.argv[1], "wb"))', '/tmp/cephtest/ceph.conf', ], stdin=run.PIPE, wait=False, ) teuthology.feed_many_stdins_and_close(conf_fp, writes) run.wait(writes) coverage_dir = '/tmp/cephtest/archive/coverage' firstmon = teuthology.get_first_mon(ctx, config) log.info('Setting up %s...' % firstmon) ctx.cluster.only(firstmon).run( args=[ '/tmp/cephtest/enable-coredump', '/tmp/cephtest/binary/usr/local/bin/ceph-coverage', coverage_dir, '/tmp/cephtest/binary/usr/local/bin/ceph-authtool', '--create-keyring', '/tmp/cephtest/ceph.keyring', ], ) ctx.cluster.only(firstmon).run( args=[ '/tmp/cephtest/enable-coredump', '/tmp/cephtest/binary/usr/local/bin/ceph-coverage', coverage_dir, '/tmp/cephtest/binary/usr/local/bin/ceph-authtool', '--gen-key', '--name=mon.', '/tmp/cephtest/ceph.keyring', ], ) (mon0_remote,) = ctx.cluster.only(firstmon).remotes.keys() teuthology.create_simple_monmap( remote=mon0_remote, conf=conf, ) log.info('Creating admin key on %s...' % firstmon) ctx.cluster.only(firstmon).run( args=[ '/tmp/cephtest/enable-coredump', '/tmp/cephtest/binary/usr/local/bin/ceph-coverage', coverage_dir, '/tmp/cephtest/binary/usr/local/bin/ceph-authtool', '--gen-key', '--name=client.admin', '--set-uid=0', '--cap', 'mon', 'allow *', '--cap', 'osd', 'allow *', '--cap', 'mds', 'allow', '/tmp/cephtest/ceph.keyring', ], ) log.info('Copying monmap to all nodes...') keyring = teuthology.get_file( remote=mon0_remote, path='/tmp/cephtest/ceph.keyring', ) monmap = teuthology.get_file( remote=mon0_remote, path='/tmp/cephtest/monmap', ) for rem in ctx.cluster.remotes.iterkeys(): # copy mon key and initial monmap log.info('Sending monmap to node {remote}'.format(remote=rem)) teuthology.write_file( remote=rem, path='/tmp/cephtest/ceph.keyring', data=keyring, ) teuthology.write_file( remote=rem, path='/tmp/cephtest/monmap', data=monmap, ) log.info('Setting up mon nodes...') mons = ctx.cluster.only(teuthology.is_type('mon')) run.wait( mons.run( args=[ '/tmp/cephtest/enable-coredump', '/tmp/cephtest/binary/usr/local/bin/ceph-coverage', coverage_dir, '/tmp/cephtest/binary/usr/local/bin/osdmaptool', '--clobber', '--createsimple', '{num:d}'.format( num=teuthology.num_instances_of_type(ctx.cluster, 'osd'), ), '/tmp/cephtest/osdmap', '--pg_bits', '2', '--pgp_bits', '4', ], wait=False, ), ) log.info('Setting up osd nodes...') osds = ctx.cluster.only(teuthology.is_type('osd')) for remote, roles_for_host in osds.remotes.iteritems(): for id_ in teuthology.roles_of_type(roles_for_host, 'osd'): remote.run( args=[ '/tmp/cephtest/enable-coredump', '/tmp/cephtest/binary/usr/local/bin/ceph-coverage', coverage_dir, '/tmp/cephtest/binary/usr/local/bin/ceph-authtool', '--create-keyring', '--gen-key', '--name=osd.{id}'.format(id=id_), '/tmp/cephtest/data/osd.{id}.keyring'.format(id=id_), ], ) log.info('Setting up mds nodes...') mdss = ctx.cluster.only(teuthology.is_type('mds')) for remote, roles_for_host in mdss.remotes.iteritems(): for id_ in teuthology.roles_of_type(roles_for_host, 'mds'): remote.run( args=[ '/tmp/cephtest/enable-coredump', '/tmp/cephtest/binary/usr/local/bin/ceph-coverage', coverage_dir, '/tmp/cephtest/binary/usr/local/bin/ceph-authtool', '--create-keyring', '--gen-key', '--name=mds.{id}'.format(id=id_), '/tmp/cephtest/data/mds.{id}.keyring'.format(id=id_), ], ) log.info('Setting up client nodes...') clients = ctx.cluster.only(teuthology.is_type('client')) for remote, roles_for_host in clients.remotes.iteritems(): for id_ in teuthology.roles_of_type(roles_for_host, 'client'): remote.run( args=[ '/tmp/cephtest/enable-coredump', '/tmp/cephtest/binary/usr/local/bin/ceph-coverage', coverage_dir, '/tmp/cephtest/binary/usr/local/bin/ceph-authtool', '--create-keyring', '--gen-key', # TODO this --name= is not really obeyed, all unknown "types" are munged to "client" '--name=client.{id}'.format(id=id_), '/tmp/cephtest/data/client.{id}.keyring'.format(id=id_), ], ) log.info('Reading keys from all nodes...') keys_fp = StringIO() keys = [] for remote, roles_for_host in ctx.cluster.remotes.iteritems(): for type_ in ['osd', 'mds', 'client']: for id_ in teuthology.roles_of_type(roles_for_host, type_): data = teuthology.get_file( remote=remote, path='/tmp/cephtest/data/{type}.{id}.keyring'.format( type=type_, id=id_, ), ) keys.append((type_, id_, data)) keys_fp.write(data) log.info('Adding keys to all mons...') writes = mons.run( args=[ 'cat', run.Raw('>>'), '/tmp/cephtest/ceph.keyring', ], stdin=run.PIPE, wait=False, ) keys_fp.seek(0) teuthology.feed_many_stdins_and_close(keys_fp, writes) run.wait(writes) for type_, id_, data in keys: run.wait( mons.run( args=[ '/tmp/cephtest/enable-coredump', '/tmp/cephtest/binary/usr/local/bin/ceph-coverage', coverage_dir, '/tmp/cephtest/binary/usr/local/bin/ceph-authtool', '/tmp/cephtest/ceph.keyring', '--name={type}.{id}'.format( type=type_, id=id_, ), ] + list(teuthology.generate_caps(type_)), wait=False, ), ) log.info('Running mkfs on mon nodes...') for remote, roles_for_host in mons.remotes.iteritems(): for id_ in teuthology.roles_of_type(roles_for_host, 'mon'): remote.run( args=[ '/tmp/cephtest/enable-coredump', '/tmp/cephtest/binary/usr/local/bin/ceph-coverage', coverage_dir, '/tmp/cephtest/binary/usr/local/bin/ceph-mon', '--mkfs', '-i', id_, '-c', '/tmp/cephtest/ceph.conf', '--monmap=/tmp/cephtest/monmap', '--osdmap=/tmp/cephtest/osdmap', '--keyring=/tmp/cephtest/ceph.keyring', ], ) log.info('Running mkfs on osd nodes...') devs_to_clean = {} for remote, roles_for_host in osds.remotes.iteritems(): roles_to_devs = {} if config.get('fs'): log.info('fs option selected, checkin for scratch devs') devs = teuthology.get_scratch_devices(remote) log.info('found devs: %s' % (str(devs),)) roles_to_devs = assign_devs( teuthology.roles_of_type(roles_for_host, 'osd'), devs ) log.info('dev map: %s' % (str(roles_to_devs),)) devs_to_clean[remote] = [] for id_ in teuthology.roles_of_type(roles_for_host, 'osd'): remote.run( args=[ 'mkdir', os.path.join('/tmp/cephtest/data', 'osd.{id}.data'.format(id=id_)), ], ) if roles_to_devs.get(id_): dev = roles_to_devs[id_] fs = config.get('fs') mkfs = ['mkfs.%s' % fs] package = None options = ['noatime'] if fs == 'btrfs': package = 'btrfs-tools' options.append('user_subvol_rm_allowed') if fs == 'xfs': package = 'xfsprogs' mkfs.append('-f') if fs == 'ext4' or fs == 'ext3': options.append('user_xattr') log.info('%s on %s on %s' % (mkfs, dev, remote)) if package is not None: remote.run( args=[ 'sudo', 'apt-get', 'install', '-y', package ] ) remote.run(args=['sudo'] + mkfs + [dev]) log.info('mount %s on %s -o %s' % (dev, remote, options)) remote.run( args=[ 'sudo', 'mount', '-t', fs, '-o', ','.join(options), dev, os.path.join('/tmp/cephtest/data', 'osd.{id}.data'.format(id=id_)), ] ) remote.run( args=[ 'sudo', 'chown', '-R', 'ubuntu.ubuntu', os.path.join('/tmp/cephtest/data', 'osd.{id}.data'.format(id=id_)) ] ) remote.run( args=[ 'sudo', 'chmod', '-R', '755', os.path.join('/tmp/cephtest/data', 'osd.{id}.data'.format(id=id_)) ] ) devs_to_clean[remote].append( os.path.join( '/tmp/cephtest/data', 'osd.{id}.data'.format(id=id_) ) ) for id_ in teuthology.roles_of_type(roles_for_host, 'osd'): remote.run( args=[ '/tmp/cephtest/enable-coredump', '/tmp/cephtest/binary/usr/local/bin/ceph-coverage', coverage_dir, '/tmp/cephtest/binary/usr/local/bin/ceph-osd', '--mkfs', '-i', id_, '-c', '/tmp/cephtest/ceph.conf', '--monmap', '/tmp/cephtest/monmap', ], ) run.wait( mons.run( args=[ 'rm', '--', '/tmp/cephtest/monmap', '/tmp/cephtest/osdmap', ], wait=False, ), ) try: yield finally: (mon0_remote,) = ctx.cluster.only(firstmon).remotes.keys() if ctx.archive is not None: log.info('Grabbing cluster log from %s %s...' % (mon0_remote, firstmon)) dest = os.path.join(ctx.archive, 'ceph.log') mon0_remote.run( args = [ 'cat', '--', '/tmp/cephtest/data/%s/log' % firstmon ], stdout=file(dest, 'wb'), ) log.info('Checking cluster ceph.log for badness...') def first_in_ceph_log(pattern, excludes): args = [ 'egrep', pattern, '/tmp/cephtest/data/%s/log' % firstmon, ] for exclude in excludes: args.extend([run.Raw('|'), 'egrep', '-v', exclude]) args.extend([ run.Raw('|'), 'head', '-n', '1', ]) r = mon0_remote.run( stdout=StringIO(), args=args, ) stdout = r.stdout.getvalue() if stdout != '': return stdout return None if first_in_ceph_log('\[ERR\]|\[WRN\]|\[SEC\]', config['log_whitelist']) is not None: log.warning('Found errors (ERR|WRN|SEC) in cluster log') ctx.summary['success'] = False # use the most severe problem as the failure reason if 'failure_reason' not in ctx.summary: for pattern in ['\[SEC\]', '\[ERR\]', '\[WRN\]']: match = first_in_ceph_log(pattern, config['log_whitelist']) if match is not None: ctx.summary['failure_reason'] = \ '"{match}" in cluster log'.format( match=match.rstrip('\n'), ) break for remote, dirs in devs_to_clean.iteritems(): for dir_ in dirs: log.info('Unmounting %s on %s' % (dir_, remote)) remote.run( args=[ 'sync', run.Raw('&&'), 'sudo', 'umount', '-f', dir_ ] ) log.info('Cleaning ceph cluster...') run.wait( ctx.cluster.run( args=[ 'rm', '-rf', '--', '/tmp/cephtest/ceph.conf', '/tmp/cephtest/ceph.keyring', '/tmp/cephtest/data', '/tmp/cephtest/monmap', run.Raw('/tmp/cephtest/asok.*') ], wait=False, ), ) @contextlib.contextmanager def run_daemon(ctx, config, type_): log.info('Starting %s daemons...' % type_) daemons = ctx.cluster.only(teuthology.is_type(type_)) coverage_dir = '/tmp/cephtest/archive/coverage' daemon_signal = 'kill' if config.get('coverage') or config.get('valgrind') is not None: daemon_signal = 'term' num_active = 0 for remote, roles_for_host in daemons.remotes.iteritems(): for id_ in teuthology.roles_of_type(roles_for_host, type_): name = '%s.%s' % (type_, id_) if not id_.endswith('-s'): num_active += 1 run_cmd = [ '/tmp/cephtest/enable-coredump', '/tmp/cephtest/binary/usr/local/bin/ceph-coverage', coverage_dir, '/tmp/cephtest/daemon-helper', daemon_signal, ] run_cmd_tail = [ '/tmp/cephtest/binary/usr/local/bin/ceph-%s' % type_, '-f', '-i', id_, '-c', '/tmp/cephtest/ceph.conf'] if config.get('valgrind') is not None: valgrind_args = None if type_ in config['valgrind']: valgrind_args = config['valgrind'][type_] if name in config['valgrind']: valgrind_args = config['valgrind'][name] run_cmd.extend(teuthology.get_valgrind_args(name, valgrind_args)) run_cmd.extend(run_cmd_tail) ctx.daemons.add_daemon(remote, type_, id_, args=run_cmd, logger=log.getChild(name), stdin=run.PIPE, wait=False, ) if type_ == 'mds': firstmon = teuthology.get_first_mon(ctx, config) (mon0_remote,) = ctx.cluster.only(firstmon).remotes.keys() mon0_remote.run(args=[ '/tmp/cephtest/enable-coredump', '/tmp/cephtest/binary/usr/local/bin/ceph-coverage', coverage_dir, '/tmp/cephtest/binary/usr/local/bin/ceph', '-c', '/tmp/cephtest/ceph.conf', 'mds', 'set_max_mds', str(num_active)]) try: yield finally: log.info('Shutting down %s daemons...' % type_) exc_info = (None, None, None) for daemon in ctx.daemons.iter_daemons_of_role(type_): try: daemon.stop() except (run.CommandFailedError, run.CommandCrashedError, run.ConnectionLostError): exc_info = sys.exc_info() log.exception('Saw exception from %s.%s', daemon.role, daemon.id_) if exc_info != (None, None, None): raise exc_info[0], exc_info[1], exc_info[2] def healthy(ctx, config): log.info('Waiting until ceph is healthy...') firstmon = teuthology.get_first_mon(ctx, config) (mon0_remote,) = ctx.cluster.only(firstmon).remotes.keys() teuthology.wait_until_osds_up( cluster=ctx.cluster, remote=mon0_remote ) teuthology.wait_until_healthy( remote=mon0_remote, ) @contextlib.contextmanager def task(ctx, config): """ Set up and tear down a Ceph cluster. For example:: tasks: - ceph: - interactive: You can also specify what branch to run:: tasks: - ceph: branch: foo Or a tag:: tasks: - ceph: tag: v0.42.13 Or a sha1:: tasks: - ceph: sha1: 1376a5ab0c89780eab39ffbbe436f6a6092314ed Or a local source dir:: tasks: - ceph: path: /home/sage/ceph To capture code coverage data, use:: tasks: - ceph: coverage: true To use btrfs, ext4, or xfs on the target's scratch disks, use:: tasks: - ceph: fs: btrfs Note, this will cause the task to check the /scratch_devs file on each node for available devices. If no such file is found, /dev/sdb will be used. To run some daemons under valgrind, include their names and the tool/args to use in a valgrind section:: tasks: - ceph: valgrind: mds.1: --tool=memcheck osd.1: [--tool=memcheck, --leak-check=no] Those nodes which are using memcheck or helgrind will get checked for bad results. To adjust or modify config options, use:: tasks: - ceph: conf: section: key: value For example:: tasks: - ceph: conf: mds.0: some option: value other key: other value client.0: debug client: 10 debug ms: 1 By default, the cluster log is checked for errors and warnings, and the run marked failed if any appear. You can ignore log entries by giving a list of egrep compatible regexes, i.e.: tasks: - ceph: log-whitelist: ['foo.*bar', 'bad message'] """ if config is None: config = {} assert isinstance(config, dict), \ "task ceph only supports a dictionary for configuration" overrides = ctx.config.get('overrides', {}) teuthology.deep_merge(config, overrides.get('ceph', {})) ctx.daemons = CephState() # Flavor tells us what gitbuilder to fetch the prebuilt software # from. It's a combination of possible keywords, in a specific # order, joined by dashes. It is used as a URL path name. If a # match is not found, the teuthology run fails. This is ugly, # and should be cleaned up at some point. dist = 'oneiric' format = 'tarball' arch = 'x86_64' flavor = 'basic' # First element: controlled by user (or not there, by default): # used to choose the right distribution, e.g. "oneiric". flavor = config.get('flavor', 'basic') if config.get('path'): # local dir precludes any other flavors flavor = 'local' else: if config.get('valgrind'): log.info('Using notcmalloc flavor and running some daemons under valgrind') flavor = 'notcmalloc' else: if config.get('coverage'): log.info('Recording coverage for this run.') flavor = 'gcov' ctx.summary['flavor'] = flavor if config.get('coverage'): coverage_dir = '/tmp/cephtest/archive/coverage' log.info('Creating coverage directory...') run.wait( ctx.cluster.run( args=[ 'install', '-d', '-m0755', '--', coverage_dir, ], wait=False, ) ) with contextutil.nested( lambda: ceph_log(ctx=ctx, config=None), lambda: ship_utilities(ctx=ctx, config=None), lambda: binaries(ctx=ctx, config=dict( branch=config.get('branch'), tag=config.get('tag'), sha1=config.get('sha1'), path=config.get('path'), flavor=flavor, dist=dist, format=format, arch=arch )), lambda: valgrind_post(ctx=ctx, config=config), lambda: cluster(ctx=ctx, config=dict( conf=config.get('conf', {}), fs=config.get('fs', None), log_whitelist=config.get('log-whitelist', []), )), lambda: run_daemon(ctx=ctx, config=config, type_='mon'), lambda: run_daemon(ctx=ctx, config=config, type_='osd'), lambda: run_daemon(ctx=ctx, config=config, type_='mds'), ): healthy(ctx=ctx, config=None) yield