mirror of
https://github.com/ceph/ceph
synced 2024-12-29 06:52:35 +00:00
ab1b78ae00
The helper gets a sequence number from the osd (or osds), and then polls the mon until that seq is reflected there. This is overkill in some cases, since many tests only require that the stats be reflected on the mgr (not the mon), but waiting for it to also reach the mon is sufficient! Signed-off-by: Sage Weil <sage@redhat.com>
177 lines
5.6 KiB
Python
177 lines
5.6 KiB
Python
"""
|
|
Lost_unfound
|
|
"""
|
|
import logging
|
|
import time
|
|
import ceph_manager
|
|
from teuthology import misc as teuthology
|
|
from teuthology.orchestra import run
|
|
from util.rados import rados
|
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
def task(ctx, config):
|
|
"""
|
|
Test handling of lost objects.
|
|
|
|
A pretty rigid cluseter is brought up andtested by this task
|
|
"""
|
|
POOL = 'unfound_pool'
|
|
if config is None:
|
|
config = {}
|
|
assert isinstance(config, dict), \
|
|
'lost_unfound task only accepts a dict for configuration'
|
|
first_mon = teuthology.get_first_mon(ctx, config)
|
|
(mon,) = ctx.cluster.only(first_mon).remotes.iterkeys()
|
|
|
|
manager = ceph_manager.CephManager(
|
|
mon,
|
|
ctx=ctx,
|
|
logger=log.getChild('ceph_manager'),
|
|
)
|
|
|
|
while len(manager.get_osd_status()['up']) < 3:
|
|
time.sleep(10)
|
|
|
|
manager.wait_for_clean()
|
|
|
|
manager.create_pool(POOL)
|
|
|
|
# something that is always there
|
|
dummyfile = '/etc/fstab'
|
|
|
|
# take an osd out until the very end
|
|
manager.kill_osd(2)
|
|
manager.mark_down_osd(2)
|
|
manager.mark_out_osd(2)
|
|
|
|
# kludge to make sure they get a map
|
|
rados(ctx, mon, ['-p', POOL, 'put', 'dummy', dummyfile])
|
|
|
|
manager.flush_pg_stats([0, 1])
|
|
manager.wait_for_recovery()
|
|
|
|
# create old objects
|
|
for f in range(1, 10):
|
|
rados(ctx, mon, ['-p', POOL, 'put', 'existing_%d' % f, dummyfile])
|
|
rados(ctx, mon, ['-p', POOL, 'put', 'existed_%d' % f, dummyfile])
|
|
rados(ctx, mon, ['-p', POOL, 'rm', 'existed_%d' % f])
|
|
|
|
# delay recovery, and make the pg log very long (to prevent backfill)
|
|
manager.raw_cluster_cmd(
|
|
'tell', 'osd.1',
|
|
'injectargs',
|
|
'--osd-recovery-delay-start 1000 --osd-min-pg-log-entries 100000000'
|
|
)
|
|
|
|
manager.kill_osd(0)
|
|
manager.mark_down_osd(0)
|
|
|
|
for f in range(1, 10):
|
|
rados(ctx, mon, ['-p', POOL, 'put', 'new_%d' % f, dummyfile])
|
|
rados(ctx, mon, ['-p', POOL, 'put', 'existed_%d' % f, dummyfile])
|
|
rados(ctx, mon, ['-p', POOL, 'put', 'existing_%d' % f, dummyfile])
|
|
|
|
# bring osd.0 back up, let it peer, but don't replicate the new
|
|
# objects...
|
|
log.info('osd.0 command_args is %s' % 'foo')
|
|
log.info(ctx.daemons.get_daemon('osd', 0).command_args)
|
|
ctx.daemons.get_daemon('osd', 0).command_kwargs['args'].extend([
|
|
'--osd-recovery-delay-start', '1000'
|
|
])
|
|
manager.revive_osd(0)
|
|
manager.mark_in_osd(0)
|
|
manager.wait_till_osd_is_up(0)
|
|
|
|
manager.flush_pg_stats([1, 0])
|
|
manager.wait_till_active()
|
|
|
|
# take out osd.1 and the only copy of those objects.
|
|
manager.kill_osd(1)
|
|
manager.mark_down_osd(1)
|
|
manager.mark_out_osd(1)
|
|
manager.raw_cluster_cmd('osd', 'lost', '1', '--yes-i-really-mean-it')
|
|
|
|
# bring up osd.2 so that things would otherwise, in theory, recovery fully
|
|
manager.revive_osd(2)
|
|
manager.mark_in_osd(2)
|
|
manager.wait_till_osd_is_up(2)
|
|
|
|
manager.flush_pg_stats([0, 2])
|
|
manager.wait_till_active()
|
|
manager.flush_pg_stats([0, 2])
|
|
|
|
# verify that there are unfound objects
|
|
unfound = manager.get_num_unfound_objects()
|
|
log.info("there are %d unfound objects" % unfound)
|
|
assert unfound
|
|
|
|
testdir = teuthology.get_testdir(ctx)
|
|
procs = []
|
|
if config.get('parallel_bench', True):
|
|
procs.append(mon.run(
|
|
args=[
|
|
"/bin/sh", "-c",
|
|
" ".join(['adjust-ulimits',
|
|
'ceph-coverage',
|
|
'{tdir}/archive/coverage',
|
|
'rados',
|
|
'--no-log-to-stderr',
|
|
'--name', 'client.admin',
|
|
'-b', str(4<<10),
|
|
'-p' , POOL,
|
|
'-t', '20',
|
|
'bench', '240', 'write',
|
|
]).format(tdir=testdir),
|
|
],
|
|
logger=log.getChild('radosbench.{id}'.format(id='client.admin')),
|
|
stdin=run.PIPE,
|
|
wait=False
|
|
))
|
|
time.sleep(10)
|
|
|
|
# mark stuff lost
|
|
pgs = manager.get_pg_stats()
|
|
for pg in pgs:
|
|
if pg['stat_sum']['num_objects_unfound'] > 0:
|
|
primary = 'osd.%d' % pg['acting'][0]
|
|
|
|
# verify that i can list them direct from the osd
|
|
log.info('listing missing/lost in %s state %s', pg['pgid'],
|
|
pg['state']);
|
|
m = manager.list_pg_missing(pg['pgid'])
|
|
#log.info('%s' % m)
|
|
assert m['num_unfound'] == pg['stat_sum']['num_objects_unfound']
|
|
num_unfound=0
|
|
for o in m['objects']:
|
|
if len(o['locations']) == 0:
|
|
num_unfound += 1
|
|
assert m['num_unfound'] == num_unfound
|
|
|
|
log.info("reverting unfound in %s on %s", pg['pgid'], primary)
|
|
manager.raw_cluster_cmd('pg', pg['pgid'],
|
|
'mark_unfound_lost', 'revert')
|
|
else:
|
|
log.info("no unfound in %s", pg['pgid'])
|
|
|
|
manager.raw_cluster_cmd('tell', 'osd.0', 'debug', 'kick_recovery_wq', '5')
|
|
manager.raw_cluster_cmd('tell', 'osd.2', 'debug', 'kick_recovery_wq', '5')
|
|
manager.flush_pg_stats([0, 2])
|
|
manager.wait_for_recovery()
|
|
|
|
# verify result
|
|
for f in range(1, 10):
|
|
err = rados(ctx, mon, ['-p', POOL, 'get', 'new_%d' % f, '-'])
|
|
assert err
|
|
err = rados(ctx, mon, ['-p', POOL, 'get', 'existed_%d' % f, '-'])
|
|
assert err
|
|
err = rados(ctx, mon, ['-p', POOL, 'get', 'existing_%d' % f, '-'])
|
|
assert not err
|
|
|
|
# see if osd.1 can cope
|
|
manager.revive_osd(1)
|
|
manager.mark_in_osd(1)
|
|
manager.wait_till_osd_is_up(1)
|
|
manager.wait_for_clean()
|
|
run.wait(procs)
|