mirror of
https://github.com/ceph/ceph
synced 2024-12-26 21:43:10 +00:00
2c2ef6d56b
Fixes: https://tracker.ceph.com/issues/51177 Signed-off-by: Kotresh HR <khiremat@redhat.com>
131 lines
4.6 KiB
Python
131 lines
4.6 KiB
Python
|
|
import logging
|
|
import json
|
|
|
|
from teuthology.task import Task
|
|
from teuthology import misc
|
|
|
|
from tasks import ceph_manager
|
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
|
|
class CheckCounter(Task):
|
|
"""
|
|
Use this task to validate that some daemon perf counters were
|
|
incremented by the nested tasks.
|
|
|
|
Config:
|
|
'cluster_name': optional, specify which cluster
|
|
'target': dictionary of daemon type to list of performance counters.
|
|
'dry_run': just log the value of the counters, don't fail if they
|
|
aren't nonzero.
|
|
|
|
Success condition is that for all of the named counters, at least
|
|
one of the daemons of that type has the counter nonzero.
|
|
|
|
Example to check cephfs dirfrag splits are happening:
|
|
- install:
|
|
- ceph:
|
|
- ceph-fuse:
|
|
- check-counter:
|
|
counters:
|
|
mds:
|
|
- "mds.dir_split"
|
|
-
|
|
name: "mds.dir_update"
|
|
min: 3
|
|
- workunit: ...
|
|
"""
|
|
@property
|
|
def admin_remote(self):
|
|
first_mon = misc.get_first_mon(self.ctx, None)
|
|
(result,) = self.ctx.cluster.only(first_mon).remotes.keys()
|
|
return result
|
|
|
|
def start(self):
|
|
log.info("START")
|
|
|
|
def end(self):
|
|
overrides = self.ctx.config.get('overrides', {})
|
|
misc.deep_merge(self.config, overrides.get('check-counter', {}))
|
|
|
|
cluster_name = self.config.get('cluster_name', None)
|
|
dry_run = self.config.get('dry_run', False)
|
|
targets = self.config.get('counters', {})
|
|
|
|
if cluster_name is None:
|
|
cluster_name = next(iter(self.ctx.managers.keys()))
|
|
|
|
|
|
mon_manager = ceph_manager.CephManager(self.admin_remote, ctx=self.ctx, logger=log.getChild('ceph_manager'))
|
|
active_mgr = json.loads(mon_manager.raw_cluster_cmd("mgr", "dump", "--format=json-pretty"))["active_name"]
|
|
|
|
for daemon_type, counters in targets.items():
|
|
# List of 'a', 'b', 'c'...
|
|
daemon_ids = list(misc.all_roles_of_type(self.ctx.cluster, daemon_type))
|
|
daemons = dict([(daemon_id,
|
|
self.ctx.daemons.get_daemon(daemon_type, daemon_id))
|
|
for daemon_id in daemon_ids])
|
|
|
|
expected = set()
|
|
seen = set()
|
|
|
|
for daemon_id, daemon in daemons.items():
|
|
if not daemon.running():
|
|
log.info("Ignoring daemon {0}, it isn't running".format(daemon_id))
|
|
continue
|
|
elif daemon_type == 'mgr' and daemon_id != active_mgr:
|
|
continue
|
|
else:
|
|
log.debug("Getting stats from {0}".format(daemon_id))
|
|
|
|
manager = self.ctx.managers[cluster_name]
|
|
proc = manager.admin_socket(daemon_type, daemon_id, ["perf", "dump"])
|
|
response_data = proc.stdout.getvalue().strip()
|
|
if response_data:
|
|
perf_dump = json.loads(response_data)
|
|
else:
|
|
log.warning("No admin socket response from {0}, skipping".format(daemon_id))
|
|
continue
|
|
|
|
minval = ''
|
|
expected_val = ''
|
|
for counter in counters:
|
|
if isinstance(counter, dict):
|
|
name = counter['name']
|
|
if 'min' in counter:
|
|
minval = counter['min']
|
|
if 'expected_val' in counter:
|
|
expected_val = counter['expected_val']
|
|
else:
|
|
name = counter
|
|
minval = 1
|
|
expected.add(name)
|
|
|
|
val = perf_dump
|
|
for key in name.split('.'):
|
|
if key not in val:
|
|
log.warning(f"Counter '{name}' not found on daemon {daemon_type}.{daemon_id}")
|
|
val = None
|
|
break
|
|
|
|
val = val[key]
|
|
|
|
if val is not None:
|
|
log.info(f"Daemon {daemon_type}.{daemon_id} {name}={val}")
|
|
if isinstance(minval, int) and val >= minval:
|
|
seen.add(name)
|
|
elif isinstance(expected_val, int) and val == expected_val:
|
|
seen.add(name)
|
|
|
|
if not dry_run:
|
|
unseen = set(expected) - set(seen)
|
|
if unseen:
|
|
raise RuntimeError("The following counters failed to be set "
|
|
"on {0} daemons: {1}".format(
|
|
daemon_type, unseen
|
|
))
|
|
|
|
task = CheckCounter
|