ceph/src/mds/Locker.h
Yan, Zheng 72d249ab0a mds: properly distinguish cap update from snap flush
Both Locker::_do_cap_update() and Locker::_do_snap_update() can update
snap inode. But only the second function is called in the case that a
client has finished snap data writeback,

Signed-off-by: "Yan, Zheng" <zyan@redhat.com>
2018-04-27 19:10:02 +08:00

293 lines
9.7 KiB
C++

// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
/*
* Ceph - scalable distributed file system
*
* Copyright (C) 2004-2006 Sage Weil <sage@newdream.net>
*
* This is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License version 2.1, as published by the Free Software
* Foundation. See file COPYING.
*
*/
#ifndef CEPH_MDS_LOCKER_H
#define CEPH_MDS_LOCKER_H
#include <string_view>
#include "include/types.h"
#include <map>
#include <list>
#include <set>
using std::map;
using std::list;
using std::set;
class MDSRank;
class Session;
class CDentry;
struct SnapRealm;
class Message;
class MLock;
class Capability;
class SimpleLock;
class ScatterLock;
class LocalLock;
#include "CInode.h"
#include "SimpleLock.h"
#include "Mutation.h"
class Locker {
private:
MDSRank *mds;
MDCache *mdcache;
public:
Locker(MDSRank *m, MDCache *c);
SimpleLock *get_lock(int lock_type, MDSCacheObjectInfo &info);
void dispatch(Message *m);
void handle_lock(MLock *m);
void tick();
void nudge_log(SimpleLock *lock);
protected:
void send_lock_message(SimpleLock *lock, int msg);
void send_lock_message(SimpleLock *lock, int msg, const bufferlist &data);
// -- locks --
void _drop_rdlocks(MutationImpl *mut, set<CInode*> *pneed_issue);
void _drop_non_rdlocks(MutationImpl *mut, set<CInode*> *pneed_issue);
public:
void include_snap_rdlocks(set<SimpleLock*>& rdlocks, CInode *in);
void include_snap_rdlocks_wlayout(set<SimpleLock*>& rdlocks, CInode *in,
file_layout_t **layout);
bool acquire_locks(MDRequestRef& mdr,
set<SimpleLock*> &rdlocks,
set<SimpleLock*> &wrlocks,
set<SimpleLock*> &xlocks,
map<SimpleLock*,mds_rank_t> *remote_wrlocks=NULL,
CInode *auth_pin_freeze=NULL,
bool auth_pin_nonblock=false);
void notify_freeze_waiter(MDSCacheObject *o);
void cancel_locking(MutationImpl *mut, set<CInode*> *pneed_issue);
void drop_locks(MutationImpl *mut, set<CInode*> *pneed_issue=0);
void set_xlocks_done(MutationImpl *mut, bool skip_dentry=false);
void drop_non_rdlocks(MutationImpl *mut, set<CInode*> *pneed_issue=0);
void drop_rdlocks_for_early_reply(MutationImpl *mut);
void eval_gather(SimpleLock *lock, bool first=false, bool *need_issue=0, list<MDSInternalContextBase*> *pfinishers=0);
void eval(SimpleLock *lock, bool *need_issue);
void eval_any(SimpleLock *lock, bool *need_issue, list<MDSInternalContextBase*> *pfinishers=0, bool first=false) {
if (!lock->is_stable())
eval_gather(lock, first, need_issue, pfinishers);
else if (lock->get_parent()->is_auth())
eval(lock, need_issue);
}
void eval_scatter_gathers(CInode *in);
void eval_cap_gather(CInode *in, set<CInode*> *issue_set=0);
bool eval(CInode *in, int mask, bool caps_imported=false);
void try_eval(MDSCacheObject *p, int mask);
void try_eval(SimpleLock *lock, bool *pneed_issue);
bool _rdlock_kick(SimpleLock *lock, bool as_anon);
bool rdlock_try(SimpleLock *lock, client_t client, MDSInternalContextBase *c);
bool rdlock_start(SimpleLock *lock, MDRequestRef& mut, bool as_anon=false);
void rdlock_finish(SimpleLock *lock, MutationImpl *mut, bool *pneed_issue);
bool can_rdlock_set(set<SimpleLock*>& locks);
bool rdlock_try_set(set<SimpleLock*>& locks);
void rdlock_take_set(set<SimpleLock*>& locks, MutationRef& mut);
void wrlock_force(SimpleLock *lock, MutationRef& mut);
bool wrlock_start(SimpleLock *lock, MDRequestRef& mut, bool nowait=false);
void wrlock_finish(SimpleLock *lock, MutationImpl *mut, bool *pneed_issue);
void remote_wrlock_start(SimpleLock *lock, mds_rank_t target, MDRequestRef& mut);
void remote_wrlock_finish(SimpleLock *lock, mds_rank_t target, MutationImpl *mut);
bool xlock_start(SimpleLock *lock, MDRequestRef& mut);
void _finish_xlock(SimpleLock *lock, client_t xlocker, bool *pneed_issue);
void xlock_finish(SimpleLock *lock, MutationImpl *mut, bool *pneed_issue);
void xlock_export(SimpleLock *lock, MutationImpl *mut);
void xlock_import(SimpleLock *lock);
// simple
public:
void try_simple_eval(SimpleLock *lock);
bool simple_rdlock_try(SimpleLock *lock, MDSInternalContextBase *con);
protected:
void simple_eval(SimpleLock *lock, bool *need_issue);
void handle_simple_lock(SimpleLock *lock, MLock *m);
public:
bool simple_sync(SimpleLock *lock, bool *need_issue=0);
protected:
void simple_lock(SimpleLock *lock, bool *need_issue=0);
void simple_excl(SimpleLock *lock, bool *need_issue=0);
void simple_xlock(SimpleLock *lock);
// scatter
public:
void scatter_eval(ScatterLock *lock, bool *need_issue); // public for MDCache::adjust_subtree_auth()
void scatter_tick();
void scatter_nudge(ScatterLock *lock, MDSInternalContextBase *c, bool forcelockchange=false);
protected:
void handle_scatter_lock(ScatterLock *lock, MLock *m);
bool scatter_scatter_fastpath(ScatterLock *lock);
void scatter_scatter(ScatterLock *lock, bool nowait=false);
void scatter_tempsync(ScatterLock *lock, bool *need_issue=0);
void scatter_writebehind(ScatterLock *lock);
void scatter_writebehind_finish(ScatterLock *lock, MutationRef& mut);
xlist<ScatterLock*> updated_scatterlocks;
public:
void mark_updated_scatterlock(ScatterLock *lock);
void handle_reqrdlock(SimpleLock *lock, MLock *m);
// caps
// when to defer processing client cap release or writeback due to being
// frozen. the condition must be consistent across handle_client_caps and
// process_request_cap_release to preserve ordering.
bool should_defer_client_cap_frozen(CInode *in);
void process_request_cap_release(MDRequestRef& mdr, client_t client, const ceph_mds_request_release& r,
std::string_view dname);
void kick_cap_releases(MDRequestRef& mdr);
void kick_issue_caps(CInode *in, client_t client, ceph_seq_t seq);
void remove_client_cap(CInode *in, client_t client);
void get_late_revoking_clients(std::list<client_t> *result) const;
bool any_late_revoking_caps(xlist<Capability*> const &revoking) const;
protected:
bool _need_flush_mdlog(CInode *in, int wanted_caps);
void adjust_cap_wanted(Capability *cap, int wanted, int issue_seq);
void handle_client_caps(class MClientCaps *m);
void _update_cap_fields(CInode *in, int dirty, MClientCaps *m, CInode::mempool_inode *pi);
void _do_snap_update(CInode *in, snapid_t snap, int dirty, snapid_t follows, client_t client, MClientCaps *m, MClientCaps *ack);
void _do_null_snapflush(CInode *head_in, client_t client, snapid_t last=CEPH_NOSNAP);
bool _do_cap_update(CInode *in, Capability *cap, int dirty, snapid_t follows, MClientCaps *m,
MClientCaps *ack=0, bool *need_flush=NULL);
void handle_client_cap_release(class MClientCapRelease *m);
void _do_cap_release(client_t client, inodeno_t ino, uint64_t cap_id, ceph_seq_t mseq, ceph_seq_t seq);
void caps_tick();
// Maintain a global list to quickly find if any caps are late revoking
xlist<Capability*> revoking_caps;
// Maintain a per-client list to find clients responsible for late ones quickly
std::map<client_t, xlist<Capability*> > revoking_caps_by_client;
elist<CInode*> need_snapflush_inodes;
public:
void snapflush_nudge(CInode *in);
void mark_need_snapflush_inode(CInode *in);
// local
public:
void local_wrlock_grab(LocalLock *lock, MutationRef& mut);
protected:
bool local_wrlock_start(LocalLock *lock, MDRequestRef& mut);
void local_wrlock_finish(LocalLock *lock, MutationImpl *mut);
bool local_xlock_start(LocalLock *lock, MDRequestRef& mut);
void local_xlock_finish(LocalLock *lock, MutationImpl *mut);
// file
public:
void file_eval(ScatterLock *lock, bool *need_issue);
protected:
void handle_file_lock(ScatterLock *lock, MLock *m);
void scatter_mix(ScatterLock *lock, bool *need_issue=0);
void file_excl(ScatterLock *lock, bool *need_issue=0);
void file_xsyn(SimpleLock *lock, bool *need_issue=0);
public:
void file_recover(ScatterLock *lock);
private:
xlist<ScatterLock*> updated_filelocks;
public:
void mark_updated_Filelock(ScatterLock *lock);
// -- file i/o --
public:
version_t issue_file_data_version(CInode *in);
Capability* issue_new_caps(CInode *in, int mode, Session *session, SnapRealm *conrealm, bool is_replay);
bool issue_caps(CInode *in, Capability *only_cap=0);
void issue_caps_set(set<CInode*>& inset);
void issue_truncate(CInode *in);
void revoke_stale_caps(Session *session);
void revoke_stale_caps(Capability *cap);
void resume_stale_caps(Session *session);
void remove_stale_leases(Session *session);
public:
void request_inode_file_caps(CInode *in);
protected:
void handle_inode_file_caps(class MInodeFileCaps *m);
void file_update_finish(CInode *in, MutationRef& mut, unsigned flags,
client_t client, MClientCaps *ack);
private:
uint64_t calc_new_max_size(CInode::mempool_inode *pi, uint64_t size);
public:
void calc_new_client_ranges(CInode *in, uint64_t size,
CInode::mempool_inode::client_range_map* new_ranges,
bool *max_increased);
bool check_inode_max_size(CInode *in, bool force_wrlock=false,
uint64_t newmax=0, uint64_t newsize=0,
utime_t mtime=utime_t());
void share_inode_max_size(CInode *in, Capability *only_cap=0);
private:
friend class C_MDL_CheckMaxSize;
friend class C_MDL_RequestInodeFileCaps;
friend class C_Locker_FileUpdate_finish;
friend class C_Locker_RetryCapRelease;
friend class C_Locker_Eval;
friend class C_Locker_ScatterWB;
friend class LockerContext;
friend class LockerLogContext;
// -- client leases --
public:
void handle_client_lease(struct MClientLease *m);
void issue_client_lease(CDentry *dn, client_t client, bufferlist &bl, utime_t now, Session *session);
void revoke_client_leases(SimpleLock *lock);
};
#endif