blob: f8e89f381bbd171fbed1877066f33c744dd7a6e1 [file] [log] [blame]
#include <stdio.h>
#include <stdlib.h>
#include <unistd.h>
#include <ctype.h>
#include <string.h>
#include <getopt.h>
#include <assert.h>
#include "fio.h"
#include "parse.h"
#define td_var_offset(var) ((size_t) &((struct thread_options *)0)->var)
/*
* Check if mmap/mmaphuge has a :/foo/bar/file at the end. If so, return that.
*/
static char *get_opt_postfix(const char *str)
{
char *p = strstr(str, ":");
if (!p)
return NULL;
p++;
strip_blank_front(&p);
strip_blank_end(p);
return strdup(p);
}
static int str_rw_cb(void *data, const char *str)
{
struct thread_data *td = data;
char *nr = get_opt_postfix(str);
td->o.ddir_nr = 1;
if (nr)
td->o.ddir_nr = atoi(nr);
return 0;
}
static int str_mem_cb(void *data, const char *mem)
{
struct thread_data *td = data;
if (td->o.mem_type == MEM_MMAPHUGE || td->o.mem_type == MEM_MMAP) {
td->mmapfile = get_opt_postfix(mem);
if (td->o.mem_type == MEM_MMAPHUGE && !td->mmapfile) {
log_err("fio: mmaphuge:/path/to/file\n");
return 1;
}
}
return 0;
}
static int str_lockmem_cb(void fio_unused *data, unsigned long *val)
{
mlock_size = *val;
return 0;
}
#ifdef FIO_HAVE_IOPRIO
static int str_prioclass_cb(void *data, unsigned int *val)
{
struct thread_data *td = data;
td->ioprio |= *val << IOPRIO_CLASS_SHIFT;
return 0;
}
static int str_prio_cb(void *data, unsigned int *val)
{
struct thread_data *td = data;
td->ioprio |= *val;
return 0;
}
#endif
static int str_exitall_cb(void)
{
exitall_on_terminate = 1;
return 0;
}
static void fill_cpu_mask(os_cpu_mask_t cpumask, int cpu)
{
#ifdef FIO_HAVE_CPU_AFFINITY
unsigned int i;
CPU_ZERO(&cpumask);
for (i = 0; i < sizeof(int) * 8; i++) {
if ((1 << i) & cpu)
CPU_SET(i, &cpumask);
}
#endif
}
static int str_cpumask_cb(void *data, unsigned int *val)
{
struct thread_data *td = data;
fill_cpu_mask(td->o.cpumask, *val);
return 0;
}
static int str_fst_cb(void *data, const char *str)
{
struct thread_data *td = data;
char *nr = get_opt_postfix(str);
td->file_service_nr = 1;
if (nr)
td->file_service_nr = atoi(nr);
return 0;
}
static int str_filename_cb(void *data, const char *input)
{
struct thread_data *td = data;
char *fname, *str, *p;
p = str = strdup(input);
strip_blank_front(&str);
strip_blank_end(str);
if (!td->files_index)
td->o.nr_files = 0;
while ((fname = strsep(&str, ":")) != NULL) {
if (!strlen(fname))
break;
add_file(td, fname);
td->o.nr_files++;
}
free(p);
return 0;
}
static int str_directory_cb(void *data, const char fio_unused *str)
{
struct thread_data *td = data;
struct stat sb;
if (lstat(td->o.directory, &sb) < 0) {
log_err("fio: %s is not a directory\n", td->o.directory);
td_verror(td, errno, "lstat");
return 1;
}
if (!S_ISDIR(sb.st_mode)) {
log_err("fio: %s is not a directory\n", td->o.directory);
return 1;
}
return 0;
}
static int str_opendir_cb(void *data, const char fio_unused *str)
{
struct thread_data *td = data;
if (!td->files_index)
td->o.nr_files = 0;
return add_dir_files(td, td->o.opendir);
}
#define __stringify_1(x) #x
#define __stringify(x) __stringify_1(x)
/*
* Map of job/command line options
*/
static struct fio_option options[] = {
{
.name = "description",
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(description),
.help = "Text job description",
},
{
.name = "name",
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(name),
.help = "Name of this job",
},
{
.name = "directory",
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(directory),
.cb = str_directory_cb,
.help = "Directory to store files in",
},
{
.name = "filename",
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(filename),
.cb = str_filename_cb,
.help = "File(s) to use for the workload",
},
{
.name = "opendir",
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(opendir),
.cb = str_opendir_cb,
.help = "Recursively add files from this directory and down",
},
{
.name = "rw",
.alias = "readwrite",
.type = FIO_OPT_STR,
.cb = str_rw_cb,
.off1 = td_var_offset(td_ddir),
.help = "IO direction",
.def = "read",
.posval = {
{ .ival = "read",
.oval = TD_DDIR_READ,
.help = "Sequential read",
},
{ .ival = "write",
.oval = TD_DDIR_WRITE,
.help = "Sequential write",
},
{ .ival = "randread",
.oval = TD_DDIR_RANDREAD,
.help = "Random read",
},
{ .ival = "randwrite",
.oval = TD_DDIR_RANDWRITE,
.help = "Random write",
},
{ .ival = "rw",
.oval = TD_DDIR_RW,
.help = "Sequential read and write mix",
},
{ .ival = "randrw",
.oval = TD_DDIR_RANDRW,
.help = "Random read and write mix"
},
},
},
{
.name = "fadvise_hint",
.type = FIO_OPT_BOOL,
.off1 = td_var_offset(fadvise_hint),
.help = "Use fadvise() to advise the kernel on IO pattern",
.def = "1",
},
{
.name = "ioengine",
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(ioengine),
.help = "IO engine to use",
.def = "sync",
.posval = {
{ .ival = "sync",
.help = "Use read/write",
},
#ifdef FIO_HAVE_LIBAIO
{ .ival = "libaio",
.help = "Linux native asynchronous IO",
},
#endif
#ifdef FIO_HAVE_POSIXAIO
{ .ival = "posixaio",
.help = "POSIX asynchronous IO",
},
#endif
{ .ival = "mmap",
.help = "Memory mapped IO",
},
#ifdef FIO_HAVE_SPLICE
{ .ival = "splice",
.help = "splice/vmsplice based IO",
},
#endif
#ifdef FIO_HAVE_SGIO
{ .ival = "sg",
.help = "SCSI generic v3 IO",
},
#endif
{ .ival = "null",
.help = "Testing engine (no data transfer)",
},
{ .ival = "net",
.help = "Network IO",
},
#ifdef FIO_HAVE_SYSLET
{ .ival = "syslet-rw",
.help = "syslet enabled async pread/pwrite IO",
},
#endif
{ .ival = "cpuio",
.help = "CPU cycler burner engine",
},
#ifdef FIO_HAVE_GUASI
{ .ival = "guasi",
.help = "GUASI IO engine",
},
#endif
{ .ival = "external",
.help = "Load external engine (append name)",
},
},
},
{
.name = "iodepth",
.type = FIO_OPT_INT,
.off1 = td_var_offset(iodepth),
.help = "Amount of IO buffers to keep in flight",
.def = "1",
},
{
.name = "iodepth_batch",
.type = FIO_OPT_INT,
.off1 = td_var_offset(iodepth_batch),
.help = "Number of IO to submit in one go",
},
{
.name = "iodepth_low",
.type = FIO_OPT_INT,
.off1 = td_var_offset(iodepth_low),
.help = "Low water mark for queuing depth",
},
{
.name = "size",
.type = FIO_OPT_STR_VAL,
.off1 = td_var_offset(size),
.help = "Total size of device or files",
},
{
.name = "filesize",
.type = FIO_OPT_STR_VAL,
.off1 = td_var_offset(file_size_low),
.off2 = td_var_offset(file_size_high),
.help = "Size of individual files",
},
{
.name = "bs",
.alias = "blocksize",
.type = FIO_OPT_STR_VAL_INT,
.off1 = td_var_offset(bs[DDIR_READ]),
.off2 = td_var_offset(bs[DDIR_WRITE]),
.help = "Block size unit",
.def = "4k",
},
{
.name = "bsrange",
.alias = "blocksize_range",
.type = FIO_OPT_RANGE,
.off1 = td_var_offset(min_bs[DDIR_READ]),
.off2 = td_var_offset(max_bs[DDIR_READ]),
.off3 = td_var_offset(min_bs[DDIR_WRITE]),
.off4 = td_var_offset(max_bs[DDIR_WRITE]),
.help = "Set block size range (in more detail than bs)",
},
{
.name = "bs_unaligned",
.alias = "blocksize_unaligned",
.type = FIO_OPT_STR_SET,
.off1 = td_var_offset(bs_unaligned),
.help = "Don't sector align IO buffer sizes",
},
{
.name = "offset",
.type = FIO_OPT_STR_VAL,
.off1 = td_var_offset(start_offset),
.help = "Start IO from this offset",
.def = "0",
},
{
.name = "randrepeat",
.type = FIO_OPT_BOOL,
.off1 = td_var_offset(rand_repeatable),
.help = "Use repeatable random IO pattern",
.def = "1",
},
{
.name = "norandommap",
.type = FIO_OPT_STR_SET,
.off1 = td_var_offset(norandommap),
.help = "Accept potential duplicate random blocks",
},
{
.name = "nrfiles",
.type = FIO_OPT_INT,
.off1 = td_var_offset(nr_files),
.help = "Split job workload between this number of files",
.def = "1",
},
{
.name = "openfiles",
.type = FIO_OPT_INT,
.off1 = td_var_offset(open_files),
.help = "Number of files to keep open at the same time",
},
{
.name = "file_service_type",
.type = FIO_OPT_STR,
.cb = str_fst_cb,
.off1 = td_var_offset(file_service_type),
.help = "How to select which file to service next",
.def = "roundrobin",
.posval = {
{ .ival = "random",
.oval = FIO_FSERVICE_RANDOM,
.help = "Choose a file at random",
},
{ .ival = "roundrobin",
.oval = FIO_FSERVICE_RR,
.help = "Round robin select files",
},
},
},
{
.name = "fsync",
.type = FIO_OPT_INT,
.off1 = td_var_offset(fsync_blocks),
.help = "Issue fsync for writes every given number of blocks",
.def = "0",
},
{
.name = "direct",
.type = FIO_OPT_BOOL,
.off1 = td_var_offset(odirect),
.help = "Use O_DIRECT IO (negates buffered)",
.def = "0",
},
{
.name = "buffered",
.type = FIO_OPT_BOOL,
.off1 = td_var_offset(odirect),
.neg = 1,
.help = "Use buffered IO (negates direct)",
.def = "1",
},
{
.name = "overwrite",
.type = FIO_OPT_BOOL,
.off1 = td_var_offset(overwrite),
.help = "When writing, set whether to overwrite current data",
.def = "0",
},
{
.name = "loops",
.type = FIO_OPT_INT,
.off1 = td_var_offset(loops),
.help = "Number of times to run the job",
.def = "1",
},
{
.name = "numjobs",
.type = FIO_OPT_INT,
.off1 = td_var_offset(numjobs),
.help = "Duplicate this job this many times",
.def = "1",
},
{
.name = "startdelay",
.type = FIO_OPT_INT,
.off1 = td_var_offset(start_delay),
.help = "Only start job when this period has passed",
.def = "0",
},
{
.name = "runtime",
.alias = "timeout",
.type = FIO_OPT_STR_VAL_TIME,
.off1 = td_var_offset(timeout),
.help = "Stop workload when this amount of time has passed",
.def = "0",
},
{
.name = "mem",
.alias = "iomem",
.type = FIO_OPT_STR,
.cb = str_mem_cb,
.off1 = td_var_offset(mem_type),
.help = "Backing type for IO buffers",
.def = "malloc",
.posval = {
{ .ival = "malloc",
.oval = MEM_MALLOC,
.help = "Use malloc(3) for IO buffers",
},
{ .ival = "shm",
.oval = MEM_SHM,
.help = "Use shared memory segments for IO buffers",
},
#ifdef FIO_HAVE_HUGETLB
{ .ival = "shmhuge",
.oval = MEM_SHMHUGE,
.help = "Like shm, but use huge pages",
},
#endif
{ .ival = "mmap",
.oval = MEM_MMAP,
.help = "Use mmap(2) (file or anon) for IO buffers",
},
#ifdef FIO_HAVE_HUGETLB
{ .ival = "mmaphuge",
.oval = MEM_MMAPHUGE,
.help = "Like mmap, but use huge pages",
},
#endif
},
},
{
.name = "verify",
.type = FIO_OPT_STR,
.off1 = td_var_offset(verify),
.help = "Verify data written",
.def = "0",
.posval = {
{ .ival = "0",
.oval = VERIFY_NONE,
.help = "Don't do IO verification",
},
{ .ival = "crc32",
.oval = VERIFY_CRC32,
.help = "Use crc32 checksums for verification",
},
{ .ival = "md5",
.oval = VERIFY_MD5,
.help = "Use md5 checksums for verification",
},
{
.ival = "null",
.oval = VERIFY_NULL,
.help = "Pretend to verify",
},
},
},
{
.name = "verifysort",
.type = FIO_OPT_BOOL,
.off1 = td_var_offset(verifysort),
.help = "Sort written verify blocks for read back",
.def = "1",
},
{
.name = "write_iolog",
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(write_iolog_file),
.help = "Store IO pattern to file",
},
{
.name = "read_iolog",
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(read_iolog_file),
.help = "Playback IO pattern from file",
},
{
.name = "exec_prerun",
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(exec_prerun),
.help = "Execute this file prior to running job",
},
{
.name = "exec_postrun",
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(exec_postrun),
.help = "Execute this file after running job",
},
#ifdef FIO_HAVE_IOSCHED_SWITCH
{
.name = "ioscheduler",
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(ioscheduler),
.help = "Use this IO scheduler on the backing device",
},
#endif
{
.name = "zonesize",
.type = FIO_OPT_STR_VAL,
.off1 = td_var_offset(zone_size),
.help = "Give size of an IO zone",
.def = "0",
},
{
.name = "zoneskip",
.type = FIO_OPT_STR_VAL,
.off1 = td_var_offset(zone_skip),
.help = "Space between IO zones",
.def = "0",
},
{
.name = "lockmem",
.type = FIO_OPT_STR_VAL,
.cb = str_lockmem_cb,
.help = "Lock down this amount of memory",
.def = "0",
},
{
.name = "rwmixcycle",
.type = FIO_OPT_INT,
.off1 = td_var_offset(rwmixcycle),
.help = "Cycle period for mixed read/write workloads (msec)",
.def = "500",
},
{
.name = "rwmixread",
.type = FIO_OPT_INT,
.off1 = td_var_offset(rwmix[DDIR_READ]),
.maxval = 100,
.help = "Percentage of mixed workload that is reads",
.def = "50",
},
{
.name = "rwmixwrite",
.type = FIO_OPT_INT,
.off1 = td_var_offset(rwmix[DDIR_WRITE]),
.maxval = 100,
.help = "Percentage of mixed workload that is writes",
.def = "50",
},
{
.name = "nice",
.type = FIO_OPT_INT,
.off1 = td_var_offset(nice),
.help = "Set job CPU nice value",
.minval = -19,
.maxval = 20,
.def = "0",
},
#ifdef FIO_HAVE_IOPRIO
{
.name = "prio",
.type = FIO_OPT_INT,
.cb = str_prio_cb,
.help = "Set job IO priority value",
.minval = 0,
.maxval = 7,
},
{
.name = "prioclass",
.type = FIO_OPT_INT,
.cb = str_prioclass_cb,
.help = "Set job IO priority class",
.minval = 0,
.maxval = 3,
},
#endif
{
.name = "thinktime",
.type = FIO_OPT_INT,
.off1 = td_var_offset(thinktime),
.help = "Idle time between IO buffers (usec)",
.def = "0",
},
{
.name = "thinktime_spin",
.type = FIO_OPT_INT,
.off1 = td_var_offset(thinktime_spin),
.help = "Start think time by spinning this amount (usec)",
.def = "0",
},
{
.name = "thinktime_blocks",
.type = FIO_OPT_INT,
.off1 = td_var_offset(thinktime_blocks),
.help = "IO buffer period between 'thinktime'",
.def = "1",
},
{
.name = "rate",
.type = FIO_OPT_INT,
.off1 = td_var_offset(rate),
.help = "Set bandwidth rate",
},
{
.name = "ratemin",
.type = FIO_OPT_INT,
.off1 = td_var_offset(ratemin),
.help = "Job must meet this rate or it will be shutdown",
},
{
.name = "rate_iops",
.type = FIO_OPT_INT,
.off1 = td_var_offset(rate_iops),
.help = "Limit IO used to this number of IO operations/sec",
},
{
.name = "rate_iops_min",
.type = FIO_OPT_INT,
.off1 = td_var_offset(rate_iops_min),
.help = "Job must meet this rate or it will be shutdown",
},
{
.name = "ratecycle",
.type = FIO_OPT_INT,
.off1 = td_var_offset(ratecycle),
.help = "Window average for rate limits (msec)",
.def = "1000",
},
{
.name = "invalidate",
.type = FIO_OPT_BOOL,
.off1 = td_var_offset(invalidate_cache),
.help = "Invalidate buffer/page cache prior to running job",
.def = "1",
},
{
.name = "sync",
.type = FIO_OPT_BOOL,
.off1 = td_var_offset(sync_io),
.help = "Use O_SYNC for buffered writes",
.def = "0",
},
{
.name = "bwavgtime",
.type = FIO_OPT_INT,
.off1 = td_var_offset(bw_avg_time),
.help = "Time window over which to calculate bandwidth (msec)",
.def = "500",
},
{
.name = "create_serialize",
.type = FIO_OPT_BOOL,
.off1 = td_var_offset(create_serialize),
.help = "Serialize creating of job files",
.def = "1",
},
{
.name = "create_fsync",
.type = FIO_OPT_BOOL,
.off1 = td_var_offset(create_fsync),
.help = "Fsync file after creation",
.def = "1",
},
{
.name = "cpuload",
.type = FIO_OPT_INT,
.off1 = td_var_offset(cpuload),
.help = "Use this percentage of CPU",
},
{
.name = "cpuchunks",
.type = FIO_OPT_INT,
.off1 = td_var_offset(cpucycle),
.help = "Length of the CPU burn cycles (usecs)",
.def = "50000",
},
#ifdef FIO_HAVE_CPU_AFFINITY
{
.name = "cpumask",
.type = FIO_OPT_INT,
.cb = str_cpumask_cb,
.help = "CPU affinity mask",
},
#endif
{
.name = "end_fsync",
.type = FIO_OPT_BOOL,
.off1 = td_var_offset(end_fsync),
.help = "Include fsync at the end of job",
.def = "0",
},
{
.name = "fsync_on_close",
.type = FIO_OPT_BOOL,
.off1 = td_var_offset(fsync_on_close),
.help = "fsync files on close",
.def = "0",
},
{
.name = "unlink",
.type = FIO_OPT_BOOL,
.off1 = td_var_offset(unlink),
.help = "Unlink created files after job has completed",
.def = "0",
},
{
.name = "exitall",
.type = FIO_OPT_STR_SET,
.cb = str_exitall_cb,
.help = "Terminate all jobs when one exits",
},
{
.name = "stonewall",
.type = FIO_OPT_STR_SET,
.off1 = td_var_offset(stonewall),
.help = "Insert a hard barrier between this job and previous",
},
{
.name = "new_group",
.type = FIO_OPT_STR_SET,
.off1 = td_var_offset(new_group),
.help = "Mark the start of a new group (for reporting)",
},
{
.name = "thread",
.type = FIO_OPT_STR_SET,
.off1 = td_var_offset(use_thread),
.help = "Use threads instead of forks",
},
{
.name = "write_bw_log",
.type = FIO_OPT_STR_SET,
.off1 = td_var_offset(write_bw_log),
.help = "Write log of bandwidth during run",
},
{
.name = "write_lat_log",
.type = FIO_OPT_STR_SET,
.off1 = td_var_offset(write_lat_log),
.help = "Write log of latency during run",
},
{
.name = "hugepage-size",
.type = FIO_OPT_STR_VAL,
.off1 = td_var_offset(hugepage_size),
.help = "When using hugepages, specify size of each page",
.def = __stringify(FIO_HUGE_PAGE),
},
{
.name = "group_reporting",
.type = FIO_OPT_STR_SET,
.off1 = td_var_offset(group_reporting),
.help = "Do reporting on a per-group basis",
},
{
.name = NULL,
},
};
void fio_options_dup_and_init(struct option *long_options)
{
struct fio_option *o;
unsigned int i;
options_init(options);
i = 0;
while (long_options[i].name)
i++;
o = &options[0];
while (o->name) {
long_options[i].name = o->name;
long_options[i].val = FIO_GETOPT_JOB;
if (o->type == FIO_OPT_STR_SET)
long_options[i].has_arg = no_argument;
else
long_options[i].has_arg = required_argument;
i++;
o++;
assert(i < FIO_NR_OPTIONS);
}
}
int fio_option_parse(struct thread_data *td, const char *opt)
{
return parse_option(opt, options, td);
}
int fio_cmd_option_parse(struct thread_data *td, const char *opt, char *val)
{
return parse_cmd_option(opt, val, options, td);
}
void fio_fill_default_options(struct thread_data *td)
{
fill_default_options(td, options);
}
int fio_show_option_help(const char *opt)
{
return show_cmd_help(options, opt);
}
static void __options_mem(struct thread_data *td, int alloc)
{
struct thread_options *o = &td->o;
struct fio_option *opt;
char **ptr;
int i;
for (i = 0, opt = &options[0]; opt->name; i++, opt = &options[i]) {
if (opt->type != FIO_OPT_STR_STORE)
continue;
ptr = (void *) o + opt->off1;
if (*ptr) {
if (alloc)
*ptr = strdup(*ptr);
else {
free(*ptr);
*ptr = NULL;
}
}
}
}
/*
* dupe FIO_OPT_STR_STORE options
*/
void options_mem_dupe(struct thread_data *td)
{
__options_mem(td, 1);
}
void options_mem_free(struct thread_data fio_unused *td)
{
#if 0
__options_mem(td, 0);
#endif
}