Merge branch 'next'

This commit is contained in:
Gary Lowell 2013-12-10 21:00:14 +00:00
commit e5a02c33e2
11 changed files with 69 additions and 29 deletions

View File

@ -8,7 +8,7 @@ AC_PREREQ(2.59)
# VERSION define is not used by the code. It gets a version string
# from 'git describe'; see src/ceph_ver.[ch]
AC_INIT([ceph], [0.72], [ceph-devel@vger.kernel.org])
AC_INIT([ceph], [0.73], [ceph-devel@vger.kernel.org])
# Create release string. Used with VERSION for RPMs.
RPM_RELEASE=0

6
debian/changelog vendored
View File

@ -1,3 +1,9 @@
ceph (0.73-1) precise; urgency=low
* New upstream release
-- Gary Lowell <gary.lowell@inktank.com> Tue, 10 Dec 2013 04:55:06 +0000
ceph (0.72-1) stable; urgency=low
* New upstream release

View File

@ -148,6 +148,7 @@ OPTION(mon_osd_down_out_subtree_limit, OPT_STR, "rack") // smallest crush unit
OPTION(mon_osd_min_up_ratio, OPT_DOUBLE, .3) // min osds required to be up to mark things down
OPTION(mon_osd_min_in_ratio, OPT_DOUBLE, .3) // min osds required to be in to mark things out
OPTION(mon_osd_max_op_age, OPT_DOUBLE, 32) // max op age before we get concerned (make it a power of 2)
OPTION(mon_osd_max_split_count, OPT_INT, 32) // largest number of PGs per "involved" OSD to let split create
OPTION(mon_stat_smooth_intervals, OPT_INT, 2) // smooth stats over last N PGMap maps
OPTION(mon_lease, OPT_FLOAT, 5) // lease interval
OPTION(mon_lease_renew_interval, OPT_FLOAT, 3) // on leader, to renew the lease

View File

@ -2870,7 +2870,21 @@ int OSDMonitor::prepare_command_pool_set(map<string,cmd_vartype> &cmdmap,
}
if (n <= (int)p.get_pg_num()) {
ss << "specified pg_num " << n << " <= current " << p.get_pg_num();
if (n < (int)p.get_pg_num())
return -EEXIST;
else
return 0;
} else {
int expected_osds = MIN(p.get_pg_num(), osdmap.get_num_osds());
int64_t new_pgs = n - p.get_pg_num();
int64_t pgs_per_osd = new_pgs / expected_osds;
if (pgs_per_osd > g_conf->mon_osd_max_split_count) {
ss << "specified pg_num " << n << " is too large (creating "
<< new_pgs << " new PGs on ~" << expected_osds
<< " OSDs exceeds per-OSD max of" << g_conf->mon_osd_max_split_count
<< ')';
return -E2BIG;
}
for(set<pg_t>::iterator i = mon->pgmon()->pg_map.creating_pgs.begin();
i != mon->pgmon()->pg_map.creating_pgs.end();
++i) {

View File

@ -1574,7 +1574,7 @@ void FileStore::queue_op(OpSequencer *osr, Op *o)
op_wq.queue(osr);
}
void FileStore::op_queue_reserve_throttle(Op *o)
void FileStore::op_queue_reserve_throttle(Op *o, ThreadPool::TPHandle *handle)
{
// Do not call while holding the journal lock!
uint64_t max_ops = m_filestore_queue_max_ops;
@ -1596,7 +1596,11 @@ void FileStore::op_queue_reserve_throttle(Op *o)
&& (op_queue_bytes + o->bytes) > max_bytes)) {
dout(2) << "waiting " << op_queue_len + 1 << " > " << max_ops << " ops || "
<< op_queue_bytes + o->bytes << " > " << max_bytes << dendl;
if (handle)
handle->suspend_tp_timeout();
op_throttle_cond.Wait(op_throttle_lock);
if (handle)
handle->reset_tp_timeout();
}
op_queue_len++;
@ -1681,7 +1685,8 @@ struct C_JournaledAhead : public Context {
};
int FileStore::queue_transactions(Sequencer *posr, list<Transaction*> &tls,
TrackedOpRef osd_op)
TrackedOpRef osd_op,
ThreadPool::TPHandle *handle)
{
Context *onreadable;
Context *ondisk;
@ -1709,7 +1714,7 @@ int FileStore::queue_transactions(Sequencer *posr, list<Transaction*> &tls,
if (journal && journal->is_writeable() && !m_filestore_journal_trailing) {
Op *o = build_op(tls, onreadable, onreadable_sync, osd_op);
op_queue_reserve_throttle(o);
op_queue_reserve_throttle(o, handle);
journal->throttle();
uint64_t op_num = submit_manager.op_submit_start();
o->op = op_num;

View File

@ -310,7 +310,7 @@ private:
Context *onreadable, Context *onreadable_sync,
TrackedOpRef osd_op);
void queue_op(OpSequencer *osr, Op *o);
void op_queue_reserve_throttle(Op *o);
void op_queue_reserve_throttle(Op *o, ThreadPool::TPHandle *handle = NULL);
void op_queue_release_throttle(Op *o);
void _journaled_ahead(OpSequencer *osr, Op *o, Context *ondisk);
friend struct C_JournaledAhead;
@ -383,7 +383,8 @@ public:
ThreadPool::TPHandle *handle);
int queue_transactions(Sequencer *osr, list<Transaction*>& tls,
TrackedOpRef op = TrackedOpRef());
TrackedOpRef op = TrackedOpRef(),
ThreadPool::TPHandle *handle = NULL);
/**
* set replay guard xattr on given file

View File

@ -19,6 +19,7 @@
#include "include/types.h"
#include "osd/osd_types.h"
#include "common/TrackedOp.h"
#include "common/WorkQueue.h"
#include "ObjectMap.h"
#include <errno.h>
@ -809,34 +810,40 @@ public:
}
unsigned apply_transactions(Sequencer *osr, list<Transaction*>& tls, Context *ondisk=0);
int queue_transaction(Sequencer *osr, Transaction* t) {
int queue_transaction(Sequencer *osr, Transaction* t,
ThreadPool::TPHandle *handle = NULL) {
list<Transaction *> tls;
tls.push_back(t);
return queue_transactions(osr, tls, new C_DeleteTransaction(t));
return queue_transactions(osr, tls, new C_DeleteTransaction(t),
NULL, NULL, TrackedOpRef(), handle);
}
int queue_transaction(Sequencer *osr, Transaction *t, Context *onreadable, Context *ondisk=0,
Context *onreadable_sync=0,
TrackedOpRef op = TrackedOpRef()) {
TrackedOpRef op = TrackedOpRef(),
ThreadPool::TPHandle *handle = NULL) {
list<Transaction*> tls;
tls.push_back(t);
return queue_transactions(osr, tls, onreadable, ondisk, onreadable_sync, op);
return queue_transactions(osr, tls, onreadable, ondisk, onreadable_sync,
op, handle);
}
int queue_transactions(Sequencer *osr, list<Transaction*>& tls,
Context *onreadable, Context *ondisk=0,
Context *onreadable_sync=0,
TrackedOpRef op = TrackedOpRef()) {
TrackedOpRef op = TrackedOpRef(),
ThreadPool::TPHandle *handle = NULL) {
assert(!tls.empty());
tls.back()->register_on_applied(onreadable);
tls.back()->register_on_commit(ondisk);
tls.back()->register_on_applied_sync(onreadable_sync);
return queue_transactions(osr, tls, op);
return queue_transactions(osr, tls, op, handle);
}
virtual int queue_transactions(
Sequencer *osr, list<Transaction*>& tls,
TrackedOpRef op = TrackedOpRef()) = 0;
TrackedOpRef op = TrackedOpRef(),
ThreadPool::TPHandle *handle = NULL) = 0;
int queue_transactions(

View File

@ -6023,13 +6023,15 @@ PG::RecoveryCtx OSD::create_context()
return rctx;
}
void OSD::dispatch_context_transaction(PG::RecoveryCtx &ctx, PG *pg)
void OSD::dispatch_context_transaction(PG::RecoveryCtx &ctx, PG *pg,
ThreadPool::TPHandle *handle)
{
if (!ctx.transaction->empty()) {
ctx.on_applied->add(new ObjectStore::C_DeleteTransaction(ctx.transaction));
int tr = store->queue_transaction(
pg->osr.get(),
ctx.transaction, ctx.on_applied, ctx.on_safe);
ctx.transaction, ctx.on_applied, ctx.on_safe, NULL,
TrackedOpRef(), handle);
assert(tr == 0);
ctx.transaction = new ObjectStore::Transaction;
ctx.on_applied = new C_Contexts(cct);
@ -6057,7 +6059,8 @@ bool OSD::compat_must_dispatch_immediately(PG *pg)
return false;
}
void OSD::dispatch_context(PG::RecoveryCtx &ctx, PG *pg, OSDMapRef curmap)
void OSD::dispatch_context(PG::RecoveryCtx &ctx, PG *pg, OSDMapRef curmap,
ThreadPool::TPHandle *handle)
{
if (service.get_osdmap()->is_up(whoami)) {
do_notifies(*ctx.notify_list, curmap);
@ -6077,7 +6080,8 @@ void OSD::dispatch_context(PG::RecoveryCtx &ctx, PG *pg, OSDMapRef curmap)
ctx.on_applied->add(new ObjectStore::C_DeleteTransaction(ctx.transaction));
int tr = store->queue_transaction(
pg->osr.get(),
ctx.transaction, ctx.on_applied, ctx.on_safe);
ctx.transaction, ctx.on_applied, ctx.on_safe, NULL, TrackedOpRef(),
handle);
assert(tr == 0);
}
}
@ -7282,17 +7286,17 @@ void OSD::process_peering_events(
split_pgs.clear();
}
if (compat_must_dispatch_immediately(pg)) {
dispatch_context(rctx, pg, curmap);
dispatch_context(rctx, pg, curmap, &handle);
rctx = create_context();
} else {
dispatch_context_transaction(rctx, pg);
dispatch_context_transaction(rctx, pg, &handle);
}
pg->unlock();
handle.reset_tp_timeout();
}
if (need_up_thru)
queue_want_up_thru(same_interval_since);
dispatch_context(rctx, 0, curmap);
dispatch_context(rctx, 0, curmap, &handle);
service.send_pg_temp();
}

View File

@ -1319,8 +1319,10 @@ protected:
// -- generic pg peering --
PG::RecoveryCtx create_context();
bool compat_must_dispatch_immediately(PG *pg);
void dispatch_context(PG::RecoveryCtx &ctx, PG *pg, OSDMapRef curmap);
void dispatch_context_transaction(PG::RecoveryCtx &ctx, PG *pg);
void dispatch_context(PG::RecoveryCtx &ctx, PG *pg, OSDMapRef curmap,
ThreadPool::TPHandle *handle = NULL);
void dispatch_context_transaction(PG::RecoveryCtx &ctx, PG *pg,
ThreadPool::TPHandle *handle = NULL);
void do_notifies(map< int,vector<pair<pg_notify_t, pg_interval_map_t> > >& notify_list,
OSDMapRef map);
void do_queries(map< int, map<pg_t,pg_query_t> >& query_map,

View File

@ -673,18 +673,18 @@ void OSDMap::get_all_osds(set<int32_t>& ls) const
ls.insert(i);
}
int OSDMap::get_num_up_osds() const
unsigned OSDMap::get_num_up_osds() const
{
int n = 0;
unsigned n = 0;
for (int i=0; i<max_osd; i++)
if ((osd_state[i] & CEPH_OSD_EXISTS) &&
(osd_state[i] & CEPH_OSD_UP)) n++;
return n;
}
int OSDMap::get_num_in_osds() const
unsigned OSDMap::get_num_in_osds() const
{
int n = 0;
unsigned n = 0;
for (int i=0; i<max_osd; i++)
if ((osd_state[i] & CEPH_OSD_EXISTS) &&
get_weight(i) != CEPH_OSD_OUT) n++;

View File

@ -259,14 +259,14 @@ private:
int get_max_osd() const { return max_osd; }
void set_max_osd(int m);
int get_num_osds() const {
unsigned get_num_osds() const {
return num_osd;
}
int calc_num_osds();
void get_all_osds(set<int32_t>& ls) const;
int get_num_up_osds() const;
int get_num_in_osds() const;
unsigned get_num_up_osds() const;
unsigned get_num_in_osds() const;
int get_flags() const { return flags; }
int test_flag(int f) const { return flags & f; }