Found using mypy 0.990. Newer versions of mypy see that the callback
function passed to `addFilter` are expected to return a bool. Update
`cherrypy_filter` to do have the correct type.
I did not try to fix the issue that this same function seems copy-pasted
into three different files. >:-}
Signed-off-by: John Mulligan <jmulligan@redhat.com>
When run locally mypy complains about the `name` field, possibly due to
us using an old mypy version on newer versions of python (we're stuck on
a Sept. 2022 version if you weren't aware). However, this could still be
an issue in mypy as there are open issues related to the `.name` field
of an enum. Regardless, we wouldn't know - so just work around it.
Signed-off-by: John Mulligan <jmulligan@redhat.com>
Configure mypy not to use namespace packages as the cython source dirs
in src/pybind/{cephfs,rados,rbd}, etc. confuse mypy into thinking
it's an empty namespace package rather than a missing module.
Found using mypy 0.990.
Signed-off-by: John Mulligan <jmulligan@redhat.com>
When a scrub job fails, update its NB to the current time plus a
fixed delay. This prevents the job from being scheduled again
immediately.
Signed-off-by: Ronen Friedman <rfriedma@redhat.com>
The call should update last_scrub_stamp, too, without
requiring an extra call to on_scrub_schedule_input_change()
Signed-off-by: Ronen Friedman <rfriedma@redhat.com>
The NB enables the OSD to delay the next attempt to schedule a specific
scrub job. This is useful for jobs that have failed for whatever
reason, especially if the primary has failed to acquire the replicas.
Signed-off-by: Ronen Friedman <rfriedma@redhat.com>
The 'penalized jobs' queue was used to track scrub jobs that had failed
to acquire their replicas, and to prevent those jobs from being retried
too quickly. This functionality will be replaced by a
simple 'not before' delay (see the next commits).
Signed-off-by: Ronen Friedman <rfriedma@redhat.com>
crimson: drop store from ECBackend to not shadow PGBackend::store
Reviewed-by: Samuel Just <sjust@redhat.com>
Reviewed-by: Matan Breizman <mbreizma@redhat.com>
src/crimson: Add support for the OSD to receive config changes
Reviewed-by: Samuel Just <sjust@redhat.com>
Reviewed-by: Matan Breizman <mbreizma@redhat.com>
As an optimization, try to ensure that the object map for the end
version is preloaded through the acquisition of exclusive lock and
as a consequence remains around until exclusive lock is released.
If it's not around, DiffRequest would (re)load it on each call.
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
If the object map for the end version is around (already loaded in
memory, either due to the end version being a snapshot or due to
exclusive lock being held), use it to run diff-iterate against the
beginning of time. Since it's the only object map needed in that
case, such calls would be satisfied locally.
Fixes: https://tracker.ceph.com/issues/63341
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
In preparation for potentially using in-memory object map, decouple
object map processing from loading object maps and place the logic in
prepare_for_object_map() and process_object_map().
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
T (ConstIterator or Iterator) is confused with const T here:
IteratorImpl dereference operator is wrongly overloaded on const
and returns Reference instead of ConstReference for ConstIterator.
This then fails inside bufferlist bowels because Reference is
incompatible with bufferlist::const_iterator.
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
Currently it's done in two cases:
- if the loaded object map is larger than expected based on byte size,
it's truncated to expected number of objects
- in case of deep-copy, if the loaded object map is smaller than diff
state, it's expanded to get "track the largest of all versions in the
set" semantics
Both of these cases can be easily dealt with without modifying the
object map. Being able to process a const object map is needed for
working on in-memory object map which is external to DiffRequest.
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
It's totally broken: instead of returning the current position and
moving to the next position, it returns the next position and doesn't
move anywhere. Luckily it hasn't been used until now.
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
This is a leftover from commit 2b3a46801d ("librbd: switch
diff-iterate API to use new object-map diff helper").
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
When getting parent diff, pass the overlap-reduced image extent instead
of the entire 0..overlap range to avoid a similar quadratic slowdown on
cloned images.
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
Currently diff-iterate in fast-diff mode is performed on the entire
image no matter what image extent is passed to the API. Then, unused
diff just gets discarded as DiffIterate ends up querying only objects
that the passed image extent maps to. This hasn't been an issue for
internal consumers ("rbd du", "rbd diff", etc) because they work on the
entire image, but turns out to lead to quadratic slowdown in some QEMU
use cases.
0..UINT64_MAX range is carved out for deep-copy which is unranged by
definition. To get effectively unranged diff-iterate, 0..UINT64_MAX-1
range can be used.
Fixes: https://tracker.ceph.com/issues/63341
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
For each covered edge case or error, run through the following
scenarios:
- where the edge case concerns snap_id_start
- where the edge case concerns snap_id_end
- where the edge case concerns intermediate snapshot and
snap_id_start == 0 (diff against the beginning of time)
- where the edge case concerns intermediate snapshot and
snap_id_start != 0 (diff from snapshot)
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
In case of diff-iterate against the beginning of time, the result
depends only on the end version. Loading and processing object maps
or intermediate snapshots is redundant and can be skipped.
This optimization is made possible by commit be507aaed1 ("librbd:
diff-iterate shouldn't ever report "new hole" against a hole") and, to
a lesser extent, the previous commit.
Getting FastDiffInvalid, LoadObjectMapError and ObjectMapTooSmall to
pass required tweaking not just expectations, but also start/end snap
ids and thus also the meaning of these tests. This is addressed in the
next commit.
Fixes: https://tracker.ceph.com/issues/63341
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
The new "track the largest of all versions in the set, diff state is
only ever grown" semantics introduced in commit 330f2a7bb9 ("librbd:
helper state machine for computing diffs between object-maps") don't
make sense for diff-iterate. It's a waste because DiffIterate won't
query beyond the end version size -- this is baked into the API.
Limit this behavior to deep-copy and resurrect the original behavior
from 2015 for diff-iterate.
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
Commit 399a45e113 ("librbd/object_map: rbd diff between two
snapshots lists entire image content") fixed most of the damage caused
by commit b81cd2460d ("librbd/object_map: diff state machine should
track object existence"), but the case of a "resize diff" when diffing
from snapshot was missed. An area that was freshly allocated in image
resize is the same in principle as a freshly created image and objects
marked OBJECT_EXISTS_CLEAN are no exception. Diff for such objects in
such an area should be set to DIFF_STATE_DATA_UPDATED, however
currently when diffing from snapshot, it's set to DIFF_STATE_DATA.
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
It became redundant with commit b81cd2460d ("librbd/object_map: diff
state machine should track object existence") -- it != end_it condition
in the loop is sufficient.
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
In preparation for multiple similarly configured MockTestImageCtx
objects being used in a single test, centralize their creation and add
a couple of helpers for setting expectations from a callback.
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
- Remove static_snap_oi name -- left over from old classic implementation,
misleading.
- OpsExecutor::prepare_clone can only be invoked on a primary, remove
the branch.
- Create the obs directly and pass to obc via set_clone_state, which
sets fully_loaded.
Signed-off-by: Samuel Just <sjust@redhat.com>
We can't guarantee that all continuations will complete before on_flush,
so we unhook and invalidate all live obcs from the registry upon
interval_change. As long as any lingering continuations simply note the
interval change and terminate, this should be safe.
Signed-off-by: Samuel Just <sjust@redhat.com>
We need to be able to clear the lru without waiting for all outstanding
references to be released. Items with such references enter an
invalidated state where references can still be added and removed with
deletion at use_count == 0, but can't be accessed any longer via the lru.
Signed-off-by: Samuel Just <sjust@redhat.com>