ceph/branches/ebofs/newsyn.cc
sageweil dc48f25847 branch for ebofs changes
git-svn-id: https://ceph.svn.sf.net/svnroot/ceph@2100 29311d96-e01e-0410-9327-a35deaab8ce9
2007-11-21 00:32:00 +00:00

435 lines
12 KiB
C++

// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
/*
* Ceph - scalable distributed file system
*
* Copyright (C) 2004-2006 Sage Weil <sage@newdream.net>
*
* This is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License version 2.1, as published by the Free Software
* Foundation. See file COPYING.
*
*/
#define intabs(x) ((x) >= 0 ? (x):(-(x)))
#include <mpi.h>
#include <sys/stat.h>
#include <iostream>
#include <string>
using namespace std;
#include <fcntl.h>
#include "config.h"
#include "mds/MDS.h"
#include "osd/OSD.h"
#include "mon/Monitor.h"
#include "client/Client.h"
#include "client/SyntheticClient.h"
#include "msg/SimpleMessenger.h"
#include "common/Timer.h"
class C_Test : public Context {
public:
void finish(int r) {
cout << "C_Test->finish(" << r << ")" << std::endl;
}
};
extern std::map<entity_name_t,float> g_fake_kill_after;
/*
* start up NewMessenger via MPI.
*/
pair<int,int> mpi_bootstrap_new(int& argc, char**& argv, MonMap *monmap)
{
MPI_Init(&argc, &argv);
int mpi_world;
int mpi_rank;
MPI_Comm_size(MPI_COMM_WORLD, &mpi_world);
MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank);
// first, synchronize clocks.
if (g_conf.clock_tare) {
if (1) {
// use an MPI barrier. probably not terribly precise.
MPI_Barrier(MPI_COMM_WORLD);
g_clock.tare();
} else {
// use wall clock; assume NTP has all nodes synchronized already.
// FIXME someday: this hangs for some reason. whatever.
utime_t z = g_clock.now();
MPI_Bcast( &z, sizeof(z), MPI_CHAR,
0, MPI_COMM_WORLD);
cout << "z is " << z << std::endl;
g_clock.tare(z);
}
}
// start up all monitors at known addresses.
entity_inst_t moninst[mpi_world]; // only care about first g_conf.num_mon of these.
rank.start_rank(); // bind and listen
if (mpi_rank < g_conf.num_mon) {
moninst[mpi_rank].addr = rank.rank_addr;
moninst[mpi_rank].name = entity_name_t(entity_name_t::TYPE_MON, mpi_rank);
//cerr << mpi_rank << " at " << rank.get_listen_addr() << std::endl;
}
MPI_Gather( &moninst[mpi_rank], sizeof(entity_inst_t), MPI_CHAR,
moninst, sizeof(entity_inst_t), MPI_CHAR,
0, MPI_COMM_WORLD);
if (mpi_rank == 0) {
for (int i=0; i<g_conf.num_mon; i++) {
cerr << "mon" << i << " is at " << moninst[i] << std::endl;
monmap->mon_inst[i] = moninst[i];
}
}
// distribute monmap
bufferlist bl;
if (mpi_rank == 0) {
monmap->encode(bl);
monmap->write(".ceph_monmap");
} else {
int l = g_conf.num_mon * 1000; // nice'n big.
bufferptr bp(l);
bl.append(bp);
}
MPI_Bcast(bl.c_str(), bl.length(), MPI_CHAR,
0, MPI_COMM_WORLD);
if (mpi_rank > 0) {
monmap->decode(bl);
}
// wait for everyone!
MPI_Barrier(MPI_COMM_WORLD);
return pair<int,int>(mpi_rank, mpi_world);
}
utime_t tick_start;
int tick_count = 0;
class C_Tick : public Context {
public:
void finish(int) {
utime_t now = g_clock.now() - tick_start;
cout << "tick +" << g_conf.tick << " -> " << now << " (" << tick_count << ")" << std::endl;
tick_count += g_conf.tick;
utime_t next = tick_start;
next.sec_ref() += tick_count;
g_timer.add_event_at(next, new C_Tick);
}
};
class C_Die : public Context {
public:
void finish(int) {
cerr << "die" << std::endl;
_exit(1);
}
};
class C_Debug : public Context {
public:
void finish(int) {
int size = (long)&g_conf.debug_after - (long)&g_conf.debug;
memcpy((char*)&g_conf.debug, (char*)&g_debug_after_conf.debug, size);
cout << "debug_after flipping debug settings" << std::endl;
//g_conf.debug_ms = 1;
}
};
int main(int argc, char **argv)
{
vector<char*> args;
argv_to_vec(argc, argv, args);
map<int,int> kill_osd_after;
if (1) {
vector<char*> nargs;
for (unsigned i=0; i<args.size(); i++) {
if (strcmp(args[i],"--kill_osd_after") == 0) {
int o = atoi(args[++i]);
int w = atoi(args[++i]);
kill_osd_after[o] = w;
}
else {
nargs.push_back( args[i] );
}
}
args.swap(nargs);
}
// stop on our own (by default)
g_conf.mon_stop_on_last_unmount = true;
g_conf.mon_stop_with_last_mds = true;
parse_config_options(args);
parse_syn_options(args);
//int start_mon = g_conf.num_mon > 0 ? g_conf.num_mon:0;
int start_mds = g_conf.num_mds > 0 ? g_conf.num_mds:0;
int start_osd = g_conf.num_osd > 0 ? g_conf.num_osd:0;
int start_client = g_conf.num_client > 0 ? g_conf.num_client:0;
//g_conf.num_mon = intabs(g_conf.num_mon);
g_conf.num_mds = intabs(g_conf.num_mds);
g_conf.num_client = intabs(g_conf.num_client);
g_conf.num_osd = intabs(g_conf.num_osd);
if (g_conf.kill_after)
g_timer.add_event_after(g_conf.kill_after, new C_Die);
if (g_conf.debug_after)
g_timer.add_event_after(g_conf.debug_after, new C_Debug);
if (g_conf.tick) {
tick_start = g_clock.now();
g_timer.add_event_after(g_conf.tick, new C_Tick);
}
vector<char*> nargs;
for (unsigned i=0; i<args.size(); i++) {
//cout << "a " << args[i] << std::endl;
// unknown arg, pass it on.
nargs.push_back(args[i]);
}
args = nargs;
if (!args.empty()) {
for (unsigned i=0; i<args.size(); i++)
cerr << "stray arg " << args[i] << std::endl;
}
assert(args.empty());
// start up messenger via MPI
MonMap *monmap = new MonMap(g_conf.num_mon);
pair<int,int> mpiwho = mpi_bootstrap_new(argc, argv, monmap);
int mpirank = mpiwho.first;
int world = mpiwho.second;
int need = 0;
if (g_conf.ms_skip_rank0) need++;
need += start_mds;
if (g_conf.ms_stripe_osds)
need++;
else
need += start_osd;
if (start_client) {
if (!g_conf.ms_overlay_clients)
need += 1;
}
assert(need <= world);
if (mpirank == 0)
cerr << "nummds " << start_mds << " numosd " << start_osd << " numclient " << start_client << " .. need " << need << ", have " << world << std::endl;
char hostname[100];
gethostname(hostname,100);
int pid = getpid();
int started = 0;
//if (mpirank == 0) g_conf.debug = 20;
// courtesy symlinks
char ffrom[100];
char fto[100];
sprintf(fto, "%s.%d", hostname, pid);
// create mon
if (mpirank < g_conf.num_mon) {
Monitor *mon = new Monitor(mpirank, rank.register_entity(entity_name_t(entity_name_t::TYPE_MON, mpirank)), monmap);
mon->init();
if (g_conf.dout_dir) {
sprintf(ffrom, "%s/mon%d", g_conf.dout_dir, mpirank);
::symlink(fto, ffrom);
}
}
// wait for monitors to start.
MPI_Barrier(MPI_COMM_WORLD);
// okay, home free!
MPI_Finalize();
// create mds
map<int,MDS*> mds;
map<int,OSD*> mdsosd;
for (int i=0; i<start_mds; i++) {
if (mpirank != g_conf.ms_skip_rank0+i) continue;
Messenger *m = rank.register_entity(entity_name_t(entity_name_t::TYPE_MDS, i));
cerr << "mds" << i << " at " << m->get_myaddr() << " " << hostname << "." << pid << std::endl;
if (g_conf.dout_dir) {
sprintf(ffrom, "%s/mds%d", g_conf.dout_dir, i);
::symlink(fto, ffrom);
}
mds[i] = new MDS(i, m, monmap);
mds[i]->init();
started++;
if (g_conf.mds_local_osd) {
int n = i+g_conf.num_osd;
mdsosd[i] = new OSD(n, rank.register_entity(entity_name_t(entity_name_t::TYPE_OSD, n)), monmap);
mdsosd[i]->init();
}
if (g_fake_kill_after.count(entity_name_t::MDS(i))) {
cerr << "mds" << i << " will die after " << g_fake_kill_after[entity_name_t::MDS(i)] << std::endl;
g_timer.add_event_after(g_fake_kill_after[entity_name_t::MDS(i)], new C_Die);
}
}
// create osd
map<int,OSD*> osd;
int max_osd_nodes = world - start_mds - g_conf.ms_skip_rank0; // assumes 0 clients, if we stripe.
int osds_per_node = (start_osd-1)/max_osd_nodes + 1;
for (int i=0; i<start_osd; i++) {
if (g_conf.ms_stripe_osds) {
if (mpirank != g_conf.ms_skip_rank0+start_mds + i / osds_per_node) continue;
} else {
if (mpirank != g_conf.ms_skip_rank0+start_mds + i) continue;
}
if (kill_osd_after.count(i))
g_timer.add_event_after(kill_osd_after[i], new C_Die);
Messenger *m = rank.register_entity(entity_name_t(entity_name_t::TYPE_OSD, i));
cerr << "osd" << i << " at " << m->get_myaddr() << " " << hostname << "." << pid << std::endl;
if (g_conf.dout_dir) {
sprintf(ffrom, "%s/osd%d", g_conf.dout_dir, i);
::symlink(fto, ffrom);
}
osd[i] = new OSD(i, m, monmap);
if (osd[i]->init() < 0)
return 1;
started++;
}
if (g_conf.ms_overlay_clients) sleep(5);
// create client
int skip_osd = start_osd;
if (g_conf.ms_overlay_clients)
skip_osd = 0; // put clients with osds too!
int client_nodes = world - start_mds - skip_osd - g_conf.ms_skip_rank0;
int clients_per_node = 1;
if (start_client && client_nodes > 0) clients_per_node = (start_client-1) / client_nodes + 1;
set<int> clientlist;
map<int,Client *> client;//[start_client];
map<int,SyntheticClient *> syn;//[start_client];
int nclients = 0;
for (int i=0; i<start_client; i++) {
int node = g_conf.ms_skip_rank0+start_mds + skip_osd + i % client_nodes;
if (mpirank != node) continue;
clientlist.insert(i);
client[i] = new Client(rank.register_entity(entity_name_t(entity_name_t::TYPE_CLIENT, -1)), monmap);
syn[i] = new SyntheticClient(client[i]);
started++;
nclients++;
}
if (!clientlist.empty()) generic_dout(2) << "i have " << clientlist << dendl;
for (set<int>::iterator it = clientlist.begin();
it != clientlist.end();
it++) {
int i = *it;
//cerr << "starting synthetic client" << i << " on rank " << mpirank << std::endl;
syn[i]->start_thread();
}
if (nclients) {
cerr << nclients << " clients at " << rank.rank_addr << " " << hostname << "." << pid << std::endl;
}
for (set<int>::iterator it = clientlist.begin();
it != clientlist.end();
it++) {
int i = *it;
// cout << "waiting for synthetic client" << i << " to finish" << std::endl;
syn[i]->join_thread();
// fix simpelmeessenger race first!
//delete syn[i];
//delete client[i];
}
if (mpirank && !started) {
//dout(1) << "IDLE" << dendl;
cerr << "idle at " << rank.rank_addr << " mpirank " << mpirank << " " << hostname << "." << pid << std::endl;
}
// wait for everything to finish
rank.wait();
cerr << "newsyn done on " << hostname << "." << pid << std::endl;
// cd on exit, so that gmon.out (if any) goes into a separate directory for each node.
char s[20];
sprintf(s, "gmon/%d", mpirank);
mkdir(s, 0755);
chdir(s);
return 0; // whatever, cleanup hangs sometimes (stopping ebofs threads?).
// cleanup
for (map<int,MDS*>::iterator i = mds.begin(); i != mds.end(); i++)
delete i->second;
for (map<int,OSD*>::iterator i = mdsosd.begin(); i != mdsosd.end(); i++)
delete i->second;
for (map<int,OSD*>::iterator i = osd.begin(); i != osd.end(); i++)
delete i->second;
/*
for (map<int,Client*>::iterator i = client.begin(); i != client.end(); i++)
delete i->second;
for (map<int,SyntheticClient*>::iterator i = syn.begin(); i != syn.end(); i++)
delete i->second;
*/
/*
for (int i=0; i<start_mds; i++) {
if (mpirank != MPI_DEST_TO_RANK(MSG_ADDR_MDS(i),world)) continue;
delete mds[i];
}
for (int i=0; i<start_osd; i++) {
if (mpirank != MPI_DEST_TO_RANK(MSG_ADDR_OSD(i),world)) continue;
delete osd[i];
}
for (int i=0; i<start_client; i++) {
if (mpirank != MPI_DEST_TO_RANK(MSG_ADDR_CLIENT(i),world)) continue;
delete client[i];
}
*/
return 0;
}