mars/mars_aio.c

699 lines
16 KiB
C
Raw Normal View History

2010-11-26 13:45:10 +00:00
// (c) 2010 Thomas Schoebel-Theuer / 1&1 Internet AG
//#define BRICK_DEBUGGING
//#define MARS_DEBUGGING
2011-02-23 20:48:06 +00:00
//#define IO_DEBUGGING
2010-11-26 13:45:10 +00:00
#include <linux/kernel.h>
#include <linux/module.h>
#include <linux/string.h>
#include <linux/list.h>
#include <linux/types.h>
#include <linux/blkdev.h>
#include <linux/kthread.h>
#include <linux/spinlock.h>
#include <linux/wait.h>
#include <linux/file.h>
#include "mars.h"
#define MARS_MAX_AIO 1024
#define MARS_MAX_AIO_READ 32
2011-03-02 09:30:56 +00:00
#define STRONG_MM
2010-11-26 13:45:10 +00:00
///////////////////////// own type definitions ////////////////////////
2011-03-07 10:27:38 +00:00
#include "mars_aio.h"
////////////////// some helpers //////////////////
2011-03-10 11:40:06 +00:00
static inline
void _enqueue(struct aio_threadinfo *tinfo, struct aio_mref_aspect *mref_a, int prio, bool at_end)
2011-03-07 10:27:38 +00:00
{
unsigned long flags;
2011-03-10 11:40:06 +00:00
if (prio > MARS_PRIO_LOW)
prio = MARS_PRIO_LOW;
if (prio < MARS_PRIO_HIGH)
prio = MARS_PRIO_HIGH;
2011-03-07 10:27:38 +00:00
traced_lock(&tinfo->lock, flags);
2011-03-10 11:40:06 +00:00
if (at_end) {
list_add_tail(&mref_a->io_head, &tinfo->mref_list[prio]);
} else {
list_add(&mref_a->io_head, &tinfo->mref_list[prio]);
}
2011-03-07 10:27:38 +00:00
traced_unlock(&tinfo->lock, flags);
}
2011-03-10 11:40:06 +00:00
static inline
struct aio_mref_aspect *_dequeue(struct aio_threadinfo *tinfo, bool do_remove)
2011-03-07 10:27:38 +00:00
{
2011-03-10 11:40:06 +00:00
struct aio_mref_aspect *mref_a = NULL;
int prio;
unsigned long flags = 0;
2011-03-07 10:27:38 +00:00
2011-03-10 11:40:06 +00:00
if (do_remove)
traced_lock(&tinfo->lock, flags);
2011-03-07 10:27:38 +00:00
2011-03-10 11:40:06 +00:00
for (prio = MARS_PRIO_HIGH; prio <= MARS_PRIO_LOW; prio++) {
struct list_head *tmp = tinfo->mref_list[prio].next;
if (tmp != &tinfo->mref_list[prio]) {
if (do_remove) {
list_del_init(tmp);
}
mref_a = container_of(tmp, struct aio_mref_aspect, io_head);
goto done;
}
2011-03-07 10:27:38 +00:00
}
2011-03-10 11:40:06 +00:00
done:
if (do_remove)
traced_unlock(&tinfo->lock, flags);
2011-03-08 16:45:52 +00:00
return mref_a;
}
2010-11-26 13:45:10 +00:00
////////////////// own brick / input / output operations //////////////////
2011-03-07 10:27:38 +00:00
static int aio_ref_get(struct aio_output *output, struct mref_object *mref)
2010-11-26 13:45:10 +00:00
{
2011-03-07 10:27:38 +00:00
struct file *file = output->filp;
2010-11-26 13:45:10 +00:00
_CHECK_ATOMIC(&mref->ref_count, !=, 0);
2011-03-07 10:27:38 +00:00
if (file) {
mref->ref_total_size = i_size_read(file->f_mapping->host);
}
2010-12-10 17:40:20 +00:00
/* Buffered IO is implemented, but should not be used
* except for testing.
* Always precede this with a buf brick -- otherwise you
* can get bad performance!
2010-11-26 13:45:10 +00:00
*/
2010-12-10 17:40:20 +00:00
if (!mref->ref_data) {
2011-03-07 10:27:38 +00:00
struct aio_mref_aspect *mref_a = aio_mref_get_aspect(output, mref);
2010-12-10 17:40:20 +00:00
if (!mref_a)
return -EILSEQ;
mref->ref_data = kmalloc(mref->ref_len, GFP_MARS);
if (!mref->ref_data)
return -ENOMEM;
mref->ref_flags = 0;
mref_a->do_dealloc = true;
2011-02-23 20:48:06 +00:00
#if 0 // litter flags for testing
2010-12-10 17:40:20 +00:00
if (mref->ref_rw) {
static int random = 0;
if (!(random++ % 2))
2010-12-15 12:13:18 +00:00
mref->ref_flags |= MREF_UPTODATE;
2010-12-10 17:40:20 +00:00
}
#endif
}
2010-11-26 13:45:10 +00:00
atomic_inc(&mref->ref_count);
return 0;
}
2011-03-07 10:27:38 +00:00
static void aio_ref_put(struct aio_output *output, struct mref_object *mref)
2010-11-26 13:45:10 +00:00
{
2011-03-07 10:27:38 +00:00
struct file *file = output->filp;
struct aio_mref_aspect *mref_a;
2010-11-26 13:45:10 +00:00
CHECK_ATOMIC(&mref->ref_count, 1);
2011-03-07 10:27:38 +00:00
if (file) {
mref->ref_total_size = i_size_read(file->f_mapping->host);
}
if (!atomic_dec_and_test(&mref->ref_count)) {
goto done;
}
mref_a = aio_mref_get_aspect(output, mref);
2010-12-10 17:40:20 +00:00
if (mref_a && mref_a->do_dealloc) {
kfree(mref->ref_data);
}
2011-03-07 10:27:38 +00:00
aio_free_mref(mref);
done:;
2010-11-26 13:45:10 +00:00
}
2011-03-07 10:27:38 +00:00
static
void _complete(struct aio_output *output, struct mref_object *mref, int err)
{
struct generic_callback *cb;
cb = mref->ref_cb;
cb->cb_error = err;
if (err < 0) {
MARS_ERR("IO error %d\n", err);
} else {
mref->ref_flags |= MREF_UPTODATE;
}
cb->cb_fn(cb);
aio_ref_put(output, mref);
}
static void aio_ref_io(struct aio_output *output, struct mref_object *mref)
2010-11-26 13:45:10 +00:00
{
struct aio_threadinfo *tinfo = &output->tinfo[0];
2011-03-07 10:27:38 +00:00
struct aio_mref_aspect *mref_a;
int err = -EINVAL;
2010-11-26 13:45:10 +00:00
atomic_inc(&mref->ref_count);
if (unlikely(!output->filp)) {
goto done;
}
2011-02-23 20:48:06 +00:00
MARS_IO("AIO rw=%d pos=%lld len=%d data=%p\n", mref->ref_rw, mref->ref_pos, mref->ref_len, mref->ref_data);
2010-11-26 13:45:10 +00:00
2011-03-07 10:27:38 +00:00
mref_a = aio_mref_get_aspect(output, mref);
if (unlikely(!mref_a)) {
goto done;
}
2011-03-10 11:40:06 +00:00
_enqueue(tinfo, mref_a, mref->ref_prio, true);
2010-11-26 13:45:10 +00:00
2011-02-23 20:48:06 +00:00
wake_up_interruptible(&tinfo->event);
2010-11-26 13:45:10 +00:00
return;
done:
2011-03-07 10:27:38 +00:00
_complete(output, mref, err);
2010-11-26 13:45:10 +00:00
}
2011-03-07 10:27:38 +00:00
static int aio_submit(struct aio_output *output, struct aio_mref_aspect *mref_a, bool use_fdsync)
2010-11-26 13:45:10 +00:00
{
2010-12-15 12:13:18 +00:00
struct mref_object *mref = mref_a->object;
2010-11-26 13:45:10 +00:00
mm_segment_t oldfs;
int res;
struct iocb iocb = {
.aio_data = (__u64)mref_a,
2010-12-10 17:40:20 +00:00
.aio_lio_opcode = use_fdsync ? IOCB_CMD_FDSYNC : (mref->ref_rw != 0 ? IOCB_CMD_PWRITE : IOCB_CMD_PREAD),
2010-11-26 13:45:10 +00:00
.aio_fildes = output->fd,
.aio_buf = (unsigned long)mref->ref_data,
.aio_nbytes = mref->ref_len,
.aio_offset = mref->ref_pos,
};
struct iocb *iocbp = &iocb;
oldfs = get_fs();
set_fs(get_ds());
res = sys_io_submit(output->ctxp, 1, &iocbp);
set_fs(oldfs);
if (res < 0 && res != -EAGAIN)
MARS_ERR("error = %d\n", res);
return res;
}
2011-03-07 10:27:38 +00:00
static int aio_submit_dummy(struct aio_output *output)
2011-02-23 20:48:06 +00:00
{
mm_segment_t oldfs;
int res;
struct iocb iocb = {
};
struct iocb *iocbp = &iocb;
oldfs = get_fs();
set_fs(get_ds());
res = sys_io_submit(output->ctxp, 1, &iocbp);
set_fs(oldfs);
return res;
}
2011-03-07 10:27:38 +00:00
static int aio_submit_thread(void *data)
2010-11-26 13:45:10 +00:00
{
struct aio_threadinfo *tinfo = data;
2011-03-07 10:27:38 +00:00
struct aio_output *output = tinfo->output;
struct file *file = output->filp;
2011-03-02 09:30:56 +00:00
struct mm_struct *old_mm;
2010-11-26 13:45:10 +00:00
int err;
/* TODO: this is provisionary. We only need it for sys_io_submit().
* The latter should be accompanied by a future vfs_submit() or
2011-02-23 20:48:06 +00:00
* do_submit() which currently does not exist :(
2010-11-26 13:45:10 +00:00
* FIXME: corresponding cleanup NYI
*/
err = get_unused_fd();
MARS_INF("fd = %d\n", err);
if (unlikely(err < 0))
return err;
output->fd = err;
fd_install(err, output->filp);
MARS_INF("kthread has started.\n");
//set_user_nice(current, -20);
2011-03-02 09:30:56 +00:00
old_mm = fake_mm();
2010-11-26 13:45:10 +00:00
if (!current->mm)
2011-03-02 09:30:56 +00:00
return -ENOMEM;
2010-11-26 13:45:10 +00:00
while (!kthread_should_stop()) {
2011-03-07 10:27:38 +00:00
struct aio_mref_aspect *mref_a;
struct mref_object *mref;
2010-11-26 13:45:10 +00:00
int err;
wait_event_interruptible_timeout(
tinfo->event,
2011-03-10 11:40:06 +00:00
kthread_should_stop() ||
_dequeue(tinfo, false),
2010-11-26 13:45:10 +00:00
HZ);
2011-03-10 11:40:06 +00:00
mref_a = _dequeue(tinfo, true);
if (!mref_a) {
2010-11-26 13:45:10 +00:00
continue;
2011-03-10 11:40:06 +00:00
}
2010-11-26 13:45:10 +00:00
2011-03-07 10:27:38 +00:00
// check for reads behind EOF
mref = mref_a->object;
if (!mref->ref_rw && mref->ref_pos + mref->ref_len > i_size_read(file->f_mapping->host)) {
2011-03-10 11:40:06 +00:00
if (mref->ref_timeout > 0 &&
((!mref_a->start_jiffies && (mref_a->start_jiffies = jiffies, true)) ||
mref_a->start_jiffies + mref->ref_timeout >= (long long)jiffies)) {
msleep(50);
_enqueue(tinfo, mref_a, mref->ref_prio, true);
2011-03-07 10:27:38 +00:00
continue;
}
2011-03-10 11:40:06 +00:00
_complete(output, mref, -ENODATA);
2011-03-07 10:27:38 +00:00
continue;
}
2010-11-26 13:45:10 +00:00
2011-03-07 10:27:38 +00:00
err = aio_submit(output, mref_a, false);
2010-11-26 13:45:10 +00:00
if (err == -EAGAIN) {
2011-03-10 11:40:06 +00:00
_enqueue(tinfo, mref_a, mref->ref_prio, false);
msleep(20);
2010-11-26 13:45:10 +00:00
continue;
}
2011-03-07 10:27:38 +00:00
if (unlikely(err < 0)) {
_complete(output, mref, err);
2010-11-26 13:45:10 +00:00
}
}
MARS_INF("kthread has stopped.\n");
2011-02-23 20:48:06 +00:00
tinfo->terminated = true;
2011-03-02 09:30:56 +00:00
cleanup_mm(old_mm);
2010-11-26 13:45:10 +00:00
return 0;
}
2011-03-07 10:27:38 +00:00
static int aio_event_thread(void *data)
2010-11-26 13:45:10 +00:00
{
struct aio_threadinfo *tinfo = data;
2011-03-07 10:27:38 +00:00
struct aio_output *output = tinfo->output;
2010-11-26 13:45:10 +00:00
struct aio_threadinfo *other = &output->tinfo[2];
2011-03-02 09:30:56 +00:00
struct mm_struct *old_mm;
int err = -ENOMEM;
2010-11-26 13:45:10 +00:00
MARS_INF("kthread has started.\n");
//set_user_nice(current, -20);
2011-03-02 09:30:56 +00:00
old_mm = fake_mm();
2010-11-26 13:45:10 +00:00
if (!current->mm)
2011-03-02 09:30:56 +00:00
goto err;
#if 1
if (!output->ctxp) {
mm_segment_t oldfs;
if (!current->mm) {
MARS_ERR("mm = %p\n", current->mm);
err = -EINVAL;
goto err;
}
oldfs = get_fs();
set_fs(get_ds());
err = sys_io_setup(MARS_MAX_AIO, &output->ctxp);
set_fs(oldfs);
if (unlikely(err))
goto err;
}
#endif
2010-11-26 13:45:10 +00:00
while (!kthread_should_stop()) {
mm_segment_t oldfs;
int count;
int bounced;
int i;
struct timespec timeout = {
2011-02-23 20:48:06 +00:00
.tv_sec = 10,
2010-11-26 13:45:10 +00:00
};
struct io_event events[MARS_MAX_AIO_READ];
oldfs = get_fs();
set_fs(get_ds());
2011-02-23 20:48:06 +00:00
/* TODO: don't timeout upon termination.
* Probably we should submit a dummy request.
*/
2010-11-26 13:45:10 +00:00
count = sys_io_getevents(output->ctxp, 1, MARS_MAX_AIO_READ, events, &timeout);
set_fs(oldfs);
//MARS_INF("count = %d\n", count);
bounced = 0;
for (i = 0; i < count; i++) {
2011-03-07 10:27:38 +00:00
struct aio_mref_aspect *mref_a = (void*)events[i].data;
2011-02-23 20:48:06 +00:00
struct mref_object *mref;
2010-11-26 13:45:10 +00:00
int err = events[i].res;
2011-02-23 20:48:06 +00:00
if (!mref_a) {
continue; // this was a dummy request
}
mref = mref_a->object;
MARS_IO("AIO done %p pos = %lld len = %d rw = %d\n", mref, mref->ref_pos, mref->ref_len, mref->ref_rw);
2010-11-26 13:45:10 +00:00
2011-03-11 13:57:54 +00:00
if (output->brick->o_fdsync
2010-11-26 13:45:10 +00:00
&& err >= 0
2011-03-10 11:40:06 +00:00
&& mref->ref_rw != READ
2010-11-26 13:45:10 +00:00
&& !mref_a->resubmit++) {
2011-03-10 11:40:06 +00:00
// workaround for non-implemented AIO FSYNC operation
2010-11-26 13:45:10 +00:00
if (!output->filp->f_op->aio_fsync) {
2011-03-10 11:40:06 +00:00
_enqueue(other, mref_a, mref->ref_prio, true);
2010-11-26 13:45:10 +00:00
bounced++;
continue;
}
2011-03-07 10:27:38 +00:00
err = aio_submit(output, mref_a, true);
2010-11-26 13:45:10 +00:00
if (likely(err >= 0))
continue;
}
2011-03-07 10:27:38 +00:00
_complete(output, mref, err);
2010-11-26 13:45:10 +00:00
}
if (bounced)
2011-02-23 20:48:06 +00:00
wake_up_interruptible(&other->event);
2010-11-26 13:45:10 +00:00
}
2011-03-02 09:30:56 +00:00
err = 0;
2010-11-26 13:45:10 +00:00
2011-03-02 09:30:56 +00:00
err:
MARS_INF("kthread has stopped, err = %d\n", err);
2011-02-23 20:48:06 +00:00
tinfo->terminated = true;
2011-03-02 09:30:56 +00:00
cleanup_mm(old_mm);
return err;
2010-11-26 13:45:10 +00:00
}
/* Workaround for non-implemented aio_fsync()
*/
2011-03-07 10:27:38 +00:00
static int aio_sync_thread(void *data)
2010-11-26 13:45:10 +00:00
{
struct aio_threadinfo *tinfo = data;
2011-03-07 10:27:38 +00:00
struct aio_output *output = tinfo->output;
2010-11-26 13:45:10 +00:00
struct file *file = output->filp;
2011-02-23 20:48:06 +00:00
MARS_INF("kthread has started on '%s'.\n", output->brick->brick_name);
2010-11-26 13:45:10 +00:00
//set_user_nice(current, -20);
while (!kthread_should_stop()) {
LIST_HEAD(tmp_list);
unsigned long flags;
2011-03-10 11:40:06 +00:00
int i;
2010-11-26 13:45:10 +00:00
int err;
wait_event_interruptible_timeout(
tinfo->event,
2011-03-10 11:40:06 +00:00
kthread_should_stop() ||
_dequeue(tinfo, false),
2010-11-26 13:45:10 +00:00
HZ);
traced_lock(&tinfo->lock, flags);
2011-03-10 11:40:06 +00:00
for (i = MARS_PRIO_HIGH; i <= MARS_PRIO_LOW; i++) {
if (!list_empty(&tinfo->mref_list[i])) {
// move over the whole list
list_replace_init(&tinfo->mref_list[i], &tmp_list);
break;
}
2010-11-26 13:45:10 +00:00
}
traced_unlock(&tinfo->lock, flags);
if (list_empty(&tmp_list))
continue;
err = vfs_fsync(file, file->f_path.dentry, 1);
if (err < 0)
MARS_ERR("FDSYNC error %d\n", err);
/* Signal completion for the whole list.
* No locking needed, it's on the stack.
*/
while (!list_empty(&tmp_list)) {
struct list_head *tmp = tmp_list.next;
2011-03-07 10:27:38 +00:00
struct aio_mref_aspect *mref_a = container_of(tmp, struct aio_mref_aspect, io_head);
2010-11-26 13:45:10 +00:00
list_del_init(tmp);
2011-03-07 10:27:38 +00:00
_complete(output, mref_a->object, err);
2010-11-26 13:45:10 +00:00
}
}
MARS_INF("kthread has stopped.\n");
2011-02-23 20:48:06 +00:00
tinfo->terminated = true;
2010-11-26 13:45:10 +00:00
return 0;
}
2011-03-07 10:27:38 +00:00
static int aio_get_info(struct aio_output *output, struct mars_info *info)
2010-11-26 13:45:10 +00:00
{
struct file *file = output->filp;
2011-02-23 20:48:06 +00:00
if (unlikely(!file || !file->f_mapping || !file->f_mapping->host))
return -EINVAL;
2010-11-26 13:45:10 +00:00
info->current_size = i_size_read(file->f_mapping->host);
2011-02-23 20:48:06 +00:00
MARS_DBG("determined file size = %lld\n", info->current_size);
2010-11-26 13:45:10 +00:00
info->backing_file = file;
return 0;
}
//////////////// object / aspect constructors / destructors ///////////////
2011-03-07 10:27:38 +00:00
static int aio_mref_aspect_init_fn(struct generic_aspect *_ini, void *_init_data)
2010-11-26 13:45:10 +00:00
{
2011-03-07 10:27:38 +00:00
struct aio_mref_aspect *ini = (void*)_ini;
2010-11-26 13:45:10 +00:00
INIT_LIST_HEAD(&ini->io_head);
return 0;
}
2011-03-07 10:27:38 +00:00
static void aio_mref_aspect_exit_fn(struct generic_aspect *_ini, void *_init_data)
2010-11-26 13:45:10 +00:00
{
2011-03-07 10:27:38 +00:00
struct aio_mref_aspect *ini = (void*)_ini;
2010-11-26 13:45:10 +00:00
(void)ini;
}
2011-03-07 10:27:38 +00:00
MARS_MAKE_STATICS(aio);
2010-11-26 13:45:10 +00:00
////////////////////// brick constructors / destructors ////////////////////
2011-03-07 10:27:38 +00:00
static int aio_brick_construct(struct aio_brick *brick)
2010-11-26 13:45:10 +00:00
{
return 0;
}
2011-03-07 10:27:38 +00:00
static int aio_switch(struct aio_brick *brick)
2010-11-26 13:45:10 +00:00
{
static int index = 0;
2011-03-07 10:27:38 +00:00
struct aio_output *output = brick->outputs[0];
2011-02-28 18:00:32 +00:00
const char *path = output->brick->brick_name;
2010-11-26 13:45:10 +00:00
int flags = O_CREAT | O_RDWR | O_LARGEFILE;
int prot = 0600;
mm_segment_t oldfs;
int i;
int err = 0;
2011-02-23 20:48:06 +00:00
MARS_DBG("power.button = %d\n", brick->power.button);
if (!brick->power.button)
goto cleanup;
2011-03-01 18:00:14 +00:00
if (brick->power.led_on)
goto done;
2011-02-23 20:48:06 +00:00
mars_power_led_off((void*)brick, false);
2011-03-11 13:57:54 +00:00
if (brick->o_direct) {
2010-11-26 13:45:10 +00:00
flags |= O_DIRECT;
MARS_INF("using O_DIRECT on %s\n", path);
}
oldfs = get_fs();
set_fs(get_ds());
output->filp = filp_open(path, flags, prot);
set_fs(oldfs);
if (unlikely(IS_ERR(output->filp))) {
err = PTR_ERR(output->filp);
MARS_ERR("can't open file '%s' status=%d\n", path, err);
output->filp = NULL;
return err;
}
2011-02-23 20:48:06 +00:00
MARS_DBG("opened file '%s'\n", path);
2011-03-11 13:57:54 +00:00
#if 1
{
struct inode *inode = output->filp->f_mapping->host;
if (S_ISBLK(inode->i_mode)) {
MARS_INF("changing readahead from %lu to %d\n", inode->i_bdev->bd_disk->queue->backing_dev_info.ra_pages, brick->readahead);
inode->i_bdev->bd_disk->queue->backing_dev_info.ra_pages = brick->readahead;
}
}
#endif
2011-03-02 09:30:56 +00:00
#if 0 // not here
2010-11-26 13:45:10 +00:00
if (!output->ctxp) {
2011-02-23 20:48:06 +00:00
if (!current->mm) {
MARS_ERR("mm = %p\n", current->mm);
err = -EINVAL;
goto err;
}
2010-11-26 13:45:10 +00:00
oldfs = get_fs();
set_fs(get_ds());
err = sys_io_setup(MARS_MAX_AIO, &output->ctxp);
set_fs(oldfs);
if (unlikely(err))
goto err;
}
2011-03-02 09:30:56 +00:00
#endif
2010-11-26 13:45:10 +00:00
for (i = 0; i < 3; i++) {
static int (*fn[])(void*) = {
2011-03-07 10:27:38 +00:00
aio_submit_thread,
aio_event_thread,
aio_sync_thread,
2010-11-26 13:45:10 +00:00
};
struct aio_threadinfo *tinfo = &output->tinfo[i];
2011-03-10 11:40:06 +00:00
int j;
for (j = MARS_PRIO_HIGH; j <= MARS_PRIO_LOW; j++) {
INIT_LIST_HEAD(&tinfo->mref_list[j]);
}
2010-11-26 13:45:10 +00:00
tinfo->output = output;
spin_lock_init(&tinfo->lock);
init_waitqueue_head(&tinfo->event);
2011-02-23 20:48:06 +00:00
tinfo->terminated = false;
2010-11-26 13:45:10 +00:00
tinfo->thread = kthread_create(fn[i], tinfo, "mars_aio%d", index++);
if (IS_ERR(tinfo->thread)) {
err = PTR_ERR(tinfo->thread);
MARS_ERR("cannot create thread\n");
tinfo->thread = NULL;
goto err;
}
wake_up_process(tinfo->thread);
}
MARS_INF("opened file '%s'\n", path);
2011-02-23 20:48:06 +00:00
mars_power_led_on((void*)brick, true);
MARS_DBG("successfully switched on.\n");
2011-03-01 18:00:14 +00:00
done:
2010-11-26 13:45:10 +00:00
return 0;
err:
MARS_ERR("status = %d\n", err);
cleanup:
2011-03-01 18:00:14 +00:00
if (brick->power.led_off) {
goto done;
}
2011-02-23 20:48:06 +00:00
mars_power_led_on((void*)brick, false);
for (i = 0; i < 3; i++) {
2010-11-26 13:45:10 +00:00
struct aio_threadinfo *tinfo = &output->tinfo[i];
if (tinfo->thread) {
kthread_stop(tinfo->thread);
tinfo->thread = NULL;
}
}
2011-03-07 10:27:38 +00:00
aio_submit_dummy(output);
2011-02-23 20:48:06 +00:00
for (i = 0; i < 3; i++) {
struct aio_threadinfo *tinfo = &output->tinfo[i];
if (tinfo->thread) {
// wait for termination
wait_event_interruptible_timeout(
tinfo->event,
tinfo->terminated, 30 * HZ);
if (tinfo->terminated)
tinfo->thread = NULL;
}
2010-11-26 13:45:10 +00:00
}
2011-02-23 20:48:06 +00:00
mars_power_led_off((void*)brick,
(output->tinfo[0].thread == NULL &&
output->tinfo[1].thread == NULL &&
output->tinfo[2].thread == NULL));
if (brick->power.led_off) {
if (output->filp) {
filp_close(output->filp, NULL);
output->filp = NULL;
}
if (output->ctxp) {
#if 0 // FIXME this crashes
sys_io_destroy(output->ctxp);
#endif
output->ctxp = 0;
}
2010-11-26 13:45:10 +00:00
}
2011-02-23 20:48:06 +00:00
MARS_DBG("switch off status = %d\n", err);
2010-11-26 13:45:10 +00:00
return err;
}
2011-03-07 10:27:38 +00:00
static int aio_output_construct(struct aio_output *output)
2010-11-26 13:45:10 +00:00
{
return 0;
}
2011-03-07 10:27:38 +00:00
static int aio_output_destruct(struct aio_output *output)
2010-11-26 13:45:10 +00:00
{
2011-02-28 18:00:32 +00:00
return mars_power_button((void*)output->brick, false);
2010-11-26 13:45:10 +00:00
}
///////////////////////// static structs ////////////////////////
2011-03-07 10:27:38 +00:00
static struct aio_brick_ops aio_brick_ops = {
.brick_switch = aio_switch,
2010-11-26 13:45:10 +00:00
};
2011-03-07 10:27:38 +00:00
static struct aio_output_ops aio_output_ops = {
.make_object_layout = aio_make_object_layout,
.mref_get = aio_ref_get,
.mref_put = aio_ref_put,
.mref_io = aio_ref_io,
.mars_get_info = aio_get_info,
2010-11-26 13:45:10 +00:00
};
2011-03-07 10:27:38 +00:00
const struct aio_output_type aio_output_type = {
.type_name = "aio_output",
.output_size = sizeof(struct aio_output),
.master_ops = &aio_output_ops,
.output_construct = &aio_output_construct,
.output_destruct = &aio_output_destruct,
.aspect_types = aio_aspect_types,
2010-11-26 13:45:10 +00:00
.layout_code = {
2010-12-15 12:13:18 +00:00
[BRICK_OBJ_MREF] = LAYOUT_NONE,
2010-11-26 13:45:10 +00:00
}
};
2011-03-07 10:27:38 +00:00
static const struct aio_output_type *aio_output_types[] = {
&aio_output_type,
2010-11-26 13:45:10 +00:00
};
2011-03-07 10:27:38 +00:00
const struct aio_brick_type aio_brick_type = {
.type_name = "aio_brick",
.brick_size = sizeof(struct aio_brick),
2010-11-26 13:45:10 +00:00
.max_inputs = 0,
.max_outputs = 1,
2011-03-07 10:27:38 +00:00
.master_ops = &aio_brick_ops,
.default_output_types = aio_output_types,
.brick_construct = &aio_brick_construct,
2010-11-26 13:45:10 +00:00
};
2011-03-07 10:27:38 +00:00
EXPORT_SYMBOL_GPL(aio_brick_type);
2010-11-26 13:45:10 +00:00
////////////////// module init stuff /////////////////////////
2011-03-07 10:27:38 +00:00
static int __init _init_aio(void)
2010-11-26 13:45:10 +00:00
{
2011-03-07 10:27:38 +00:00
MARS_INF("init_aio()\n");
_aio_brick_type = (void*)&aio_brick_type;
return aio_register_brick_type();
2010-11-26 13:45:10 +00:00
}
2011-03-07 10:27:38 +00:00
static void __exit _exit_aio(void)
2010-11-26 13:45:10 +00:00
{
2011-03-07 10:27:38 +00:00
MARS_INF("exit_aio()\n");
aio_unregister_brick_type();
2010-11-26 13:45:10 +00:00
}
2011-03-07 10:27:38 +00:00
MODULE_DESCRIPTION("MARS aio brick");
2010-11-26 13:45:10 +00:00
MODULE_AUTHOR("Thomas Schoebel-Theuer <tst@1und1.de>");
MODULE_LICENSE("GPL");
2011-03-07 10:27:38 +00:00
module_init(_init_aio);
module_exit(_exit_aio);