This test deals with enabling/disabling the modules. The assumption I
have is after enabling the
module test will wait for an active mgr but its not able to find it in
time and it fails. so taking inspiration from 6c7253be6f adding retries and logs to see if that's the case
Fixes: https://tracker.ceph.com/issues/62972
Signed-off-by: Nizamudeen A <nia@redhat.com>
In "nvmeof" task, change "client" config to "installer"
which allows to take inputs like "host.a".
nvmeof/basic: change 2-gateway-2-initiator to
4-gateway-2-inititator cluster
nvmeof/thrash: change 3-gateway-1-initiator to
4-gateway-1-inititaor cluster
Signed-off-by: Vallari Agrawal <val.agl002@gmail.com>
list-snaps is exempt from clipping in ImageDispatcher::PreprocessVisitor
because it's considered to be an internal API. Further, reads issued
by ObjectCopyRequest based on list-snaps results may also be exempt
because of READ_FLAG_DISABLE_CLIPPING.
Since RawFormat allows specifying a set of snapshots (possibly of
varying size!) to be imported, it needs to compensate for that in its
list-snaps implementation. Otherwise, an out-of-bounds read will
eventually be submitted to the stream.
Fixes: https://tracker.ceph.com/issues/67845
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
in the rgw/multisite suite, jobs fail on user creation:
> radosgw-admin --cluster c1 account create --account-id RGW11111111111111111
> radosgw-admin --cluster c1 user create --uid rgw-multisite-test-user --account-id RGW11111111111111111 --account-root --rgw-zone test-zone1 --rgw-zonegroup test-zonegroup --rgw-realm test-realm --display-name TestUser --gen-access-key --gen-secret
> could not create user: unable to create user, Failed to load account by id
realms/two-zones.yaml misspells `is_default` as `is default` for the
realm, so it doesn't get set as default. the `account create` command
doesn't specify a realm/zonegroup/zone, so operates on the "default"
zone and zonegroup
use `zone_args()` to add the explicit realm/zonegroup/zone arguments
Fixes: https://tracker.ceph.com/issues/67839
Signed-off-by: Casey Bodley <cbodley@redhat.com>
smb: cluster public ip addresses support
Reviewed-by: Adam King <adking@redhat.com>
Reviewed-by: Anoop C S <anoopcs@cryptolab.net>
Reviewed-by: Avan Thakkar <athakkar@redhat.com>
Reviewed-by: Michael Adam <obnox@samba.org>
1. qa/tasks/nvmeof.py: add other methods to stop nvmeof daemons
2. add qa/workunits/rbd/nvmeof_namespace_test.sh which adds and
deletes new namespaces. It is run in nvmeof_namespaces.yaml
job where fio happens to other namespaces in background.
Signed-off-by: Vallari Agrawal <val.agl002@gmail.com>
* refs/pull/58419/head:
mds: generate correct path for unlinked snapped files
qa: add test for cephx path check on unlinked snapped dir tree
mds: add debugging for stray_prior_path
Reviewed-by: Milind Changire <mchangir@redhat.com>
Groups was made a required parameter to be
`ceph orch apply nvmeof <pool> <group>` in
https://github.com/ceph/ceph/pull/58860.
That broke the `nvmeof` suite so this PR fixes that.
Right now, all gateway are deployed in a single group.
Later, this would be changed to have multi groups for a better test.
Signed-off-by: Vallari Agrawal <val.agl002@gmail.com>
this commit adds support for the 'bench' admin command in the OSD,
allowing administrators to perform benchmark tests on the OSD. The
'bench' command accepts 4 optional parameters with the following
default values:
1. count - Total number of bytes to write (default: 1GB).
2. size - Block size for each write operation (default: 4MB).
3. object_size - Size of each object to write (default: 0).
4. object_num - Number of objects to write (default: 0).
The results of the benchmark are returned in a JSON formatted output,
which includes the following fields:
1. bytes_written - Total number of bytes written during the benchmark.
2. blocksize - Block size used for each write operation.
3. elapsed_sec - Total time taken to complete the benchmark in seconds.
4. bytes_per_sec - Write throughput in bytes per second.
5. iops - Number of input/output operations per second.
Example JSON output:
```json
{
"osd_bench_results": {
"bytes_written": 1073741824,
"blocksize": 4194304,
"elapsed_sec": 0.5,
"bytes_per_sec": 2147483648,
"iops": 512
}
}
Fixes: https://tracker.ceph.com/issues/66380
Signed-off-by: Nitzan Mordechai <nmordech@redhat.com>
The scrub scheduler no longer "upgrades" shallow scrubs into
deep ones on error, so the tests that check this functionality
are no longer valid.
Signed-off-by: Ronen Friedman <rfriedma@redhat.com>
The conditions for auto-repair scrubs should have been changed
when need_auto lost some of its setters.
Also fix the rescheduling of repair scrubs
when the last scrub ended with errors.
Signed-off-by: Ronen Friedman <rfriedma@redhat.com>
Disabling osd-scrub-test.sh::TEST_scrub_extended_sleep,
as the test is no longer valid (updated code no longer
produces the same logs or the same behavior).
Signed-off-by: Ronen Friedman <rfriedma@redhat.com>
The variable was added in commit 9b6b7c35d0 ("Handle
differently-named xmlstarlet binary for *suse") but this
compatibility business is long outdated:
Mon Oct 13 08:52:37 UTC 2014 - toms@opensuse.org
- SPEC file changes
- Added link from /usr/bin/xml to /usr/bin/xmlstarlet as other
distributions do the same
- Did the same for the manpage
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
The cephadm_from_container allows one to do a single container build
and then point teuthology at that image as the "single source of truth".
I find this extremely convenient when running teuthology locally and
I keep carrying this patch around - I figure having it upstream will
simplify my workflow. Maybe someday it'll benefit others too.
To use it I set up a yaml overrides file with the following content:
```yaml
overrides:
cephadm:
image: "quay.io/phlogistonjohn/ceph:dev"
cephadm_from_container: true
verify_ceph_hash: false
verify_ceph_hash: false
```
This let's me test my custom builds fairly easily!
Signed-off-by: John Mulligan <phlogistonjohn@asynchrono.us>
During teuthology tests, the tearing down of the cluster between two
tests causes the config to be reset and a config_notify generated. This
leads to a race to create a new mount using the old fscid. But by the
time the mount is attempted the new fs gets created with a new fscid.
This situation leads to the client mount waiting for a connection
completion notification from the mds for 5 minutes (default timeout)
and eventually giving up.
However, the default teuthology command timeout is 2 minutes. So,
teuthology fails the command and declares the job as failed way before
the mount can timeout.
The resolution to this case is to lower the client mount timeout to 30
seconds so that the config_notify fails fast paving the way for
successive commands to get executed with the new fs.
An unhandled cluster warning about an unresponsive client also gets
emitted later during qa job termination which leads to teuthology
declaring the job as failed. As of now this warning seems harmless since
it is emitted during cluster cleanup phase.
So, this warning is added to the log-ignorelist section in the
snap-schedule YAML.
Fixes: https://tracker.ceph.com/issues/66009
Signed-off-by: Milind Changire <mchangir@redhat.com>
To work around a potential linking issue between
nvme-cli ad libnvme that prevents nvme-cli from
correctly generating a hostnqn, causing
nvme_fabrics: found same hostid edb4e426-766f-44c6-b127-da2a5b7446ef but different hostnqn hostnqn
messages in dmesg and the inability to setup nvme
loop devices
Fixes: https://tracker.ceph.com/issues/67684
Signed-off-by: Adam King <adking@redhat.com>