rgw/notifications: fetch object state to get size, in rgw_lc.cc
Reviewed-by: Daniel Gryniewicz <dang@redhat.com>
Reviewed-by: Yuval Lifshitz <ylifshit@redhat.com>
kv/RocksDBStore: don't use real wholespace iterator for prefixed access
Reviewed-by: Adam Kupczyk <akupczyk@redhat.com>
Reviewed-by: Cory Snyder <csnyder@iland.com>
crush: fix compilation guard around specialization for libfmt => 9
Reviewed-by: Samuel Just <sjust@redhat.com>
Reviewed-by: Matan Breizman <Matan.Brz@gmail.com>
ceph-volume tests are failing, OSDs never get up and running.
For some reason, updating the OS early in the testing workflow
addresses that issue in the CI.
-- to be continued ... --
Signed-off-by: Guillaume Abrioux <gabrioux@ibm.com>
* refs/pull/48280/head:
qa: test new client with old cluster
Reviewed-by: Xiubo Li <xiubli@redhat.com>
Reviewed-by: Rishabh Dave <ridave@redhat.com>
Reviewed-by: Venky Shankar <vshankar@redhat.com>
* refs/pull/49102/head:
qa: add test for schedule auto deactivation
doc/cephfs: add note about path usage in snap-schedule
mgr/snap_schedule: add debug log for missing or wrong path
Reviewed-by: Venky Shankar <vshankar@redhat.com>
Reviewed-by: Rishabh Dave <ridave@redhat.com>
Now, the current code does not allow qat to exert
its acceleration advantage, but leads to poor performance
of qat. The QAT batch mode is implemented here, so that
the QAT performance can be displayed.
When we scale up the number of concurrent requests and worry about
QAT instance will be bottleneck, we can fall back to CPU.
And there add a parameter to set the times of the number of QAT
instance to wait for free instance, which can avoid QAT to be
not busy and make sure QAT full of utilization as much as possible.
max_queue_size is up to max_requests
Add optional_yield in RGWPutObj_BlockEncrypt and
RGWGetObj_BlockDecrypt. Make it with coroutine and non-coroutine mode.
Signed-off-by: Feng Hualong <hualong.feng@intel.com>
Bucket notification supports reading the rule configuration from the cache first, rather than directly from the bottom layer
Signed-off-by: liuhong <liuhong_yewu@cmss.chinamobile.com>
* refs/pull/45224/head:
mds: add support for snapshots alternate_name
Reviewed-by: Xiubo Li <xiubli@redhat.com>
Reviewed-by: David Disseldorp <ddiss@suse.de>
Reviewed-by: Jeff Layton <jlayton@redhat.com>
Reviewed-by: Venky Shankar <vshankar@redhat.com>
Adds an experimental script that allows a bucket index of a
non-versioned bucket to be restored by applying `radosgw-admin object
reindex ...` to all objects in the specified bucket. The objects in
the bucket are determined by scanning the data pool for head objects
containing the bucket's marker.
Signed-off-by: J. Eric Ivancich <ivancich@redhat.com>
This subcommand will take a bucket and an object (or a list of objects
in a file) and add those objects to the bucket's index. It does so by
preparing the bucket index entry update and then allowing the
so-called "dir suggest" mechansim, which is triggered by listing the
bucket, to completing the bucket index entry. This mechanism is
triggered by listing the bucket. Using this mechanism has the
advantage of doing this lazily that both allows the reindex operation
to run more quickly and distributes the workload over time.
This subcommand can help restore lost bucket index entries.
Signed-off-by: J. Eric Ivancich <ivancich@redhat.com>
the stable branches like quincy pulls from the quay.io/ceph/ceph:v17 to
bootstrap the ceph cluster in test environments. This will cause issues
because the branches are changing constantly but the image is not. So
using the quay.ceph.io repo to bring the cluster in test environment.
Signed-off-by: Nizamudeen A <nia@redhat.com>