blob: f188d1c665fac1a7240504c3957364e036a2a451 [file] [log] [blame]
/*
* read/write() engine that uses syslet to be async
*
*/
#include <stdio.h>
#include <stdlib.h>
#include <unistd.h>
#include <errno.h>
#include <assert.h>
#include "../fio.h"
#include "../os.h"
#ifdef FIO_HAVE_SYSLET
struct syslet_data {
struct io_u **events;
unsigned int nr_events;
struct async_head_user ahu;
struct syslet_uatom **ring;
};
/*
* Inspect the ring to see if we have completed events
*/
static void fio_syslet_complete(struct thread_data *td)
{
struct syslet_data *sd = td->io_ops->data;
do {
struct syslet_uatom *atom;
struct io_u *io_u;
long ret;
atom = sd->ring[sd->ahu.user_ring_idx];
if (!atom)
break;
sd->ring[sd->ahu.user_ring_idx] = NULL;
if (++sd->ahu.user_ring_idx == td->iodepth)
sd->ahu.user_ring_idx = 0;
io_u = atom->private;
ret = *atom->ret_ptr;
if (ret > 0)
io_u->resid = io_u->xfer_buflen - ret;
else if (ret < 0)
io_u->error = ret;
sd->events[sd->nr_events++] = io_u;
} while (1);
}
static int fio_syslet_getevents(struct thread_data *td, int min,
int fio_unused max,
struct timespec fio_unused *t)
{
struct syslet_data *sd = td->io_ops->data;
int get_events;
long ret;
do {
fio_syslet_complete(td);
/*
* do we have enough immediate completions?
*/
if (sd->nr_events >= (unsigned int) min)
break;
/*
* OK, we need to wait for some events...
*/
get_events = min - sd->nr_events;
ret = async_wait(get_events, sd->ahu.user_ring_idx, &sd->ahu);
if (ret < 0)
return -errno;
} while (1);
ret = sd->nr_events;
sd->nr_events = 0;
return ret;
}
static struct io_u *fio_syslet_event(struct thread_data *td, int event)
{
struct syslet_data *sd = td->io_ops->data;
return sd->events[event];
}
static void init_atom(struct syslet_uatom *atom, int nr, void *arg0,
void *arg1, void *arg2, void *arg3, void *ret_ptr,
unsigned long flags, void *priv)
{
atom->flags = flags;
atom->nr = nr;
atom->ret_ptr = ret_ptr;
atom->next = NULL;
atom->arg_ptr[0] = arg0;
atom->arg_ptr[1] = arg1;
atom->arg_ptr[2] = arg2;
atom->arg_ptr[3] = arg3;
atom->arg_ptr[4] = atom->arg_ptr[5] = NULL;
atom->private = priv;
}
/*
* Use seek atom for sync
*/
static void fio_syslet_prep_sync(struct io_u *io_u, struct fio_file *f)
{
init_atom(&io_u->req.atom, __NR_fsync, &f->fd, NULL, NULL, NULL,
&io_u->req.ret, 0, io_u);
}
static void fio_syslet_prep_rw(struct io_u *io_u, struct fio_file *f)
{
int nr;
/*
* prepare rw
*/
if (io_u->ddir == DDIR_READ)
nr = __NR_pread64;
else
nr = __NR_pwrite64;
init_atom(&io_u->req.atom, nr, &f->fd, &io_u->xfer_buf,
&io_u->xfer_buflen, &io_u->offset, &io_u->req.ret, 0, io_u);
}
static int fio_syslet_prep(struct thread_data fio_unused *td, struct io_u *io_u)
{
struct fio_file *f = io_u->file;
if (io_u->ddir == DDIR_SYNC)
fio_syslet_prep_sync(io_u, f);
else
fio_syslet_prep_rw(io_u, f);
return 0;
}
static void cachemiss_thread_start(void)
{
while (1)
async_thread();
}
#define THREAD_STACK_SIZE (16384)
static unsigned long thread_stack_alloc()
{
return (unsigned long)malloc(THREAD_STACK_SIZE) + THREAD_STACK_SIZE;
}
static int fio_syslet_queue(struct thread_data *td, struct io_u *io_u)
{
struct syslet_data *sd = td->io_ops->data;
struct syslet_uatom *done;
long ret;
if (!sd->ahu.new_thread_stack)
sd->ahu.new_thread_stack = thread_stack_alloc();
/*
* On sync completion, the atom is returned. So on NULL return
* it's queued asynchronously.
*/
done = async_exec(&io_u->req.atom, &sd->ahu);
if (!done)
return FIO_Q_QUEUED;
/*
* completed sync
*/
ret = io_u->req.ret;
if (ret != (long) io_u->xfer_buflen) {
if (ret > 0) {
io_u->resid = io_u->xfer_buflen - ret;
io_u->error = 0;
return FIO_Q_COMPLETED;
} else
io_u->error = errno;
}
assert(sd->nr_events < td->iodepth);
if (io_u->error)
td_verror(td, io_u->error);
return FIO_Q_COMPLETED;
}
static int async_head_init(struct syslet_data *sd, unsigned int depth)
{
unsigned long ring_size;
memset(&sd->ahu, 0, sizeof(struct async_head_user));
ring_size = sizeof(struct syslet_uatom *) * depth;
sd->ring = malloc(ring_size);
memset(sd->ring, 0, ring_size);
sd->ahu.user_ring_idx = 0;
sd->ahu.completion_ring = sd->ring;
sd->ahu.ring_size_bytes = ring_size;
sd->ahu.head_stack = thread_stack_alloc();
sd->ahu.head_eip = (unsigned long)cachemiss_thread_start;
sd->ahu.new_thread_eip = (unsigned long)cachemiss_thread_start;
return 0;
}
static void async_head_exit(struct syslet_data *sd)
{
free(sd->ring);
}
static void fio_syslet_cleanup(struct thread_data *td)
{
struct syslet_data *sd = td->io_ops->data;
if (sd) {
async_head_exit(sd);
free(sd->events);
free(sd);
td->io_ops->data = NULL;
}
}
static int fio_syslet_init(struct thread_data *td)
{
struct syslet_data *sd;
sd = malloc(sizeof(*sd));
memset(sd, 0, sizeof(*sd));
sd->events = malloc(sizeof(struct io_u *) * td->iodepth);
memset(sd->events, 0, sizeof(struct io_u *) * td->iodepth);
/*
* This will handily fail for kernels where syslet isn't available
*/
if (async_head_init(sd, td->iodepth)) {
free(sd->events);
free(sd);
return 1;
}
td->io_ops->data = sd;
return 0;
}
static struct ioengine_ops ioengine = {
.name = "syslet-rw",
.version = FIO_IOOPS_VERSION,
.init = fio_syslet_init,
.prep = fio_syslet_prep,
.queue = fio_syslet_queue,
.getevents = fio_syslet_getevents,
.event = fio_syslet_event,
.cleanup = fio_syslet_cleanup,
};
#else /* FIO_HAVE_SYSLET */
/*
* When we have a proper configure system in place, we simply wont build
* and install this io engine. For now install a crippled version that
* just complains and fails to load.
*/
static int fio_syslet_init(struct thread_data fio_unused *td)
{
fprintf(stderr, "fio: syslet not available\n");
return 1;
}
static struct ioengine_ops ioengine = {
.name = "syslet-rw",
.version = FIO_IOOPS_VERSION,
.init = fio_syslet_init,
};
#endif /* FIO_HAVE_SYSLET */
static void fio_init fio_syslet_register(void)
{
register_ioengine(&ioengine);
}
static void fio_exit fio_syslet_unregister(void)
{
unregister_ioengine(&ioengine);
}