blob: 23c95810c76e368c4744f6358cab183caa3bbb66 [file] [log] [blame]
/* Copyright (c) 2012-2013, The Linux Foundation. All rights reserved.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License version 2 and
* only version 2 as published by the Free Software Foundation.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* The test scheduler allows to test the block device by dispatching
* specific requests according to the test case and declare PASS/FAIL
* according to the requests completion error code.
* Each test is exposed via debugfs and can be triggered by writing to
* the debugfs file.
*
*/
/* elevator test iosched */
#include <linux/blkdev.h>
#include <linux/elevator.h>
#include <linux/bio.h>
#include <linux/module.h>
#include <linux/slab.h>
#include <linux/init.h>
#include <linux/debugfs.h>
#include <linux/test-iosched.h>
#include <linux/delay.h>
#include "blk.h"
#define MODULE_NAME "test-iosched"
#define WR_RD_START_REQ_ID 1234
#define UNIQUE_START_REQ_ID 5678
#define TIMEOUT_TIMER_MS 40000
#define TEST_MAX_TESTCASE_ROUNDS 15
#define test_pr_debug(fmt, args...) pr_debug("%s: "fmt"\n", MODULE_NAME, args)
#define test_pr_info(fmt, args...) pr_info("%s: "fmt"\n", MODULE_NAME, args)
#define test_pr_err(fmt, args...) pr_err("%s: "fmt"\n", MODULE_NAME, args)
static DEFINE_SPINLOCK(blk_dev_test_list_lock);
static LIST_HEAD(blk_dev_test_list);
static struct test_data *ptd;
/**
* test_iosched_get_req_queue() - returns the request queue
* served by the scheduler
*/
struct request_queue *test_iosched_get_req_queue(void)
{
if (!ptd)
return NULL;
return ptd->req_q;
}
EXPORT_SYMBOL(test_iosched_get_req_queue);
/**
* test_iosched_mark_test_completion() - Wakeup the debugfs
* thread, waiting on the test completion
*/
void test_iosched_mark_test_completion(void)
{
if (!ptd)
return;
test_pr_info("%s: mark test is completed, test_count=%d,",
__func__, ptd->test_count);
test_pr_info("%s: reinsert_count=%d, dispatched_count=%d",
__func__, ptd->reinsert_count, ptd->dispatched_count);
ptd->test_state = TEST_COMPLETED;
wake_up(&ptd->wait_q);
}
EXPORT_SYMBOL(test_iosched_mark_test_completion);
/**
* check_test_completion() - Check if all the queued test
* requests were completed
*/
void check_test_completion(void)
{
struct test_request *test_rq;
if (!ptd)
goto exit;
list_for_each_entry(test_rq, &ptd->dispatched_queue, queuelist)
if (!test_rq->req_completed)
goto exit;
if (!list_empty(&ptd->test_queue)
|| !list_empty(&ptd->reinsert_queue)
|| !list_empty(&ptd->urgent_queue)) {
test_pr_info("%s: Test still not completed,", __func__);
test_pr_info("%s: test_count=%d, reinsert_count=%d",
__func__, ptd->test_count, ptd->reinsert_count);
test_pr_info("%s: dispatched_count=%d, urgent_count=%d",
__func__, ptd->dispatched_count, ptd->urgent_count);
goto exit;
}
ptd->test_info.test_duration = jiffies -
ptd->test_info.test_duration;
test_pr_info("%s: Test is completed, test_count=%d, reinsert_count=%d,",
__func__, ptd->test_count, ptd->reinsert_count);
test_pr_info("%s: dispatched_count=%d",
__func__, ptd->dispatched_count);
test_iosched_mark_test_completion();
exit:
return;
}
EXPORT_SYMBOL(check_test_completion);
/*
* A callback to be called per bio completion.
* Frees the bio memory.
*/
static void end_test_bio(struct bio *bio, int err)
{
if (err)
clear_bit(BIO_UPTODATE, &bio->bi_flags);
bio_put(bio);
}
/*
* A callback to be called per request completion.
* the request memory is not freed here, will be freed later after the test
* results checking.
*/
static void end_test_req(struct request *rq, int err)
{
struct test_request *test_rq;
test_rq = (struct test_request *)rq->elv.priv[0];
BUG_ON(!test_rq);
test_pr_debug("%s: request %d completed, err=%d",
__func__, test_rq->req_id, err);
test_rq->req_completed = true;
test_rq->req_result = err;
check_test_completion();
}
/**
* test_iosched_add_unique_test_req() - Create and queue a non
* read/write request (such as FLUSH/DISCRAD/SANITIZE).
* @is_err_expcted: A flag to indicate if this request
* should succeed or not
* @req_unique: The type of request to add
* @start_sec: start address of the first bio
* @nr_sects: number of sectors in the request
* @end_req_io: specific completion callback. When not
* set, the defaulcallback will be used
*/
int test_iosched_add_unique_test_req(int is_err_expcted,
enum req_unique_type req_unique,
int start_sec, int nr_sects, rq_end_io_fn *end_req_io)
{
struct bio *bio;
struct request *rq;
int rw_flags;
struct test_request *test_rq;
if (!ptd)
return -ENODEV;
bio = bio_alloc(GFP_KERNEL, 0);
if (!bio) {
test_pr_err("%s: Failed to allocate a bio", __func__);
return -ENODEV;
}
bio_get(bio);
bio->bi_end_io = end_test_bio;
switch (req_unique) {
case REQ_UNIQUE_FLUSH:
bio->bi_rw = WRITE_FLUSH;
break;
case REQ_UNIQUE_DISCARD:
bio->bi_rw = REQ_WRITE | REQ_DISCARD;
bio->bi_size = nr_sects << 9;
bio->bi_sector = start_sec;
break;
case REQ_UNIQUE_SANITIZE:
bio->bi_rw = REQ_WRITE | REQ_SANITIZE;
break;
default:
test_pr_err("%s: Invalid request type %d", __func__,
req_unique);
bio_put(bio);
return -ENODEV;
}
rw_flags = bio_data_dir(bio);
if (bio->bi_rw & REQ_SYNC)
rw_flags |= REQ_SYNC;
rq = blk_get_request(ptd->req_q, rw_flags, GFP_KERNEL);
if (!rq) {
test_pr_err("%s: Failed to allocate a request", __func__);
bio_put(bio);
return -ENODEV;
}
init_request_from_bio(rq, bio);
if (end_req_io)
rq->end_io = end_req_io;
else
rq->end_io = end_test_req;
test_rq = kzalloc(sizeof(struct test_request), GFP_KERNEL);
if (!test_rq) {
test_pr_err("%s: Failed to allocate a test request", __func__);
bio_put(bio);
blk_put_request(rq);
return -ENODEV;
}
test_rq->req_completed = false;
test_rq->req_result = -EINVAL;
test_rq->rq = rq;
test_rq->is_err_expected = is_err_expcted;
rq->elv.priv[0] = (void *)test_rq;
test_rq->req_id = ptd->unique_next_req_id++;
test_pr_debug(
"%s: added request %d to the test requests list, type = %d",
__func__, test_rq->req_id, req_unique);
spin_lock_irq(ptd->req_q->queue_lock);
list_add_tail(&test_rq->queuelist, &ptd->test_queue);
ptd->test_count++;
spin_unlock_irq(ptd->req_q->queue_lock);
return 0;
}
EXPORT_SYMBOL(test_iosched_add_unique_test_req);
/*
* Get a pattern to be filled in the request data buffer.
* If the pattern used is (-1) the buffer will be filled with sequential
* numbers
*/
static void fill_buf_with_pattern(int *buf, int num_bytes, int pattern)
{
int i = 0;
int num_of_dwords = num_bytes/sizeof(int);
if (pattern == TEST_NO_PATTERN)
return;
/* num_bytes should be aligned to sizeof(int) */
BUG_ON((num_bytes % sizeof(int)) != 0);
if (pattern == TEST_PATTERN_SEQUENTIAL) {
for (i = 0; i < num_of_dwords; i++)
buf[i] = i;
} else {
for (i = 0; i < num_of_dwords; i++)
buf[i] = pattern;
}
}
/**
* test_iosched_create_test_req() - Create a read/write request.
* @is_err_expcted: A flag to indicate if this request
* should succeed or not
* @direction: READ/WRITE
* @start_sec: start address of the first bio
* @num_bios: number of BIOs to be allocated for the
* request
* @pattern: A pattern, to be written into the write
* requests data buffer. In case of READ
* request, the given pattern is kept as
* the expected pattern. The expected
* pattern will be compared in the test
* check result function. If no comparisson
* is required, set pattern to
* TEST_NO_PATTERN.
* @end_req_io: specific completion callback. When not
* set,the default callback will be used
*
* This function allocates the test request and the block
* request and calls blk_rq_map_kern which allocates the
* required BIO. The allocated test request and the block
* request memory is freed at the end of the test and the
* allocated BIO memory is freed by end_test_bio.
*/
struct test_request *test_iosched_create_test_req(int is_err_expcted,
int direction, int start_sec,
int num_bios, int pattern, rq_end_io_fn *end_req_io)
{
struct request *rq;
struct test_request *test_rq;
int rw_flags, buf_size;
int ret = 0, i;
unsigned int *bio_ptr = NULL;
struct bio *bio = NULL;
if (!ptd)
return NULL;
rw_flags = direction;
rq = blk_get_request(ptd->req_q, rw_flags, GFP_KERNEL);
if (!rq) {
test_pr_err("%s: Failed to allocate a request", __func__);
return NULL;
}
test_rq = kzalloc(sizeof(struct test_request), GFP_KERNEL);
if (!test_rq) {
test_pr_err("%s: Failed to allocate test request", __func__);
blk_put_request(rq);
return NULL;
}
buf_size = sizeof(unsigned int) * BIO_U32_SIZE * num_bios;
test_rq->bios_buffer = kzalloc(buf_size, GFP_KERNEL);
if (!test_rq->bios_buffer) {
test_pr_err("%s: Failed to allocate the data buf", __func__);
goto err;
}
test_rq->buf_size = buf_size;
if (direction == WRITE)
fill_buf_with_pattern(test_rq->bios_buffer,
buf_size, pattern);
test_rq->wr_rd_data_pattern = pattern;
bio_ptr = test_rq->bios_buffer;
for (i = 0; i < num_bios; ++i) {
ret = blk_rq_map_kern(ptd->req_q, rq,
(void *)bio_ptr,
sizeof(unsigned int)*BIO_U32_SIZE,
GFP_KERNEL);
if (ret) {
test_pr_err("%s: blk_rq_map_kern returned error %d",
__func__, ret);
goto err;
}
bio_ptr += BIO_U32_SIZE;
}
if (end_req_io)
rq->end_io = end_req_io;
else
rq->end_io = end_test_req;
rq->__sector = start_sec;
rq->cmd_type |= REQ_TYPE_FS;
rq->cmd_flags |= REQ_SORTED;
if (rq->bio) {
rq->bio->bi_sector = start_sec;
rq->bio->bi_end_io = end_test_bio;
bio = rq->bio;
while ((bio = bio->bi_next) != NULL)
bio->bi_end_io = end_test_bio;
}
ptd->num_of_write_bios += num_bios;
test_rq->req_id = ptd->wr_rd_next_req_id++;
test_rq->req_completed = false;
test_rq->req_result = -EINVAL;
test_rq->rq = rq;
if (ptd->test_info.get_rq_disk_fn)
test_rq->rq->rq_disk = ptd->test_info.get_rq_disk_fn();
test_rq->is_err_expected = is_err_expcted;
rq->elv.priv[0] = (void *)test_rq;
test_pr_debug("%s: created test request %d, buf_size=%d",
__func__, test_rq->req_id, buf_size);
return test_rq;
err:
blk_put_request(rq);
kfree(test_rq->bios_buffer);
return NULL;
}
EXPORT_SYMBOL(test_iosched_create_test_req);
/**
* test_iosched_add_wr_rd_test_req() - Create and queue a
* read/write request.
* @is_err_expcted: A flag to indicate if this request
* should succeed or not
* @direction: READ/WRITE
* @start_sec: start address of the first bio
* @num_bios: number of BIOs to be allocated for the
* request
* @pattern: A pattern, to be written into the write
* requests data buffer. In case of READ
* request, the given pattern is kept as
* the expected pattern. The expected
* pattern will be compared in the test
* check result function. If no comparisson
* is required, set pattern to
* TEST_NO_PATTERN.
* @end_req_io: specific completion callback. When not
* set,the default callback will be used
*
* This function allocates the test request and the block
* request and calls blk_rq_map_kern which allocates the
* required BIO. Upon success the new request is added to the
* test_queue. The allocated test request and the block request
* memory is freed at the end of the test and the allocated BIO
* memory is freed by end_test_bio.
*/
int test_iosched_add_wr_rd_test_req(int is_err_expcted,
int direction, int start_sec,
int num_bios, int pattern, rq_end_io_fn *end_req_io)
{
struct test_request *test_rq = NULL;
test_rq = test_iosched_create_test_req(is_err_expcted,
direction, start_sec,
num_bios, pattern, end_req_io);
if (test_rq) {
spin_lock_irq(ptd->req_q->queue_lock);
list_add_tail(&test_rq->queuelist, &ptd->test_queue);
ptd->test_count++;
spin_unlock_irq(ptd->req_q->queue_lock);
return 0;
}
return -ENODEV;
}
EXPORT_SYMBOL(test_iosched_add_wr_rd_test_req);
/* Converts the testcase number into a string */
static char *get_test_case_str(struct test_data *td)
{
if (td->test_info.get_test_case_str_fn)
return td->test_info.get_test_case_str_fn(td);
return "Unknown testcase";
}
/*
* Verify that the test request data buffer includes the expected
* pattern
*/
static int compare_buffer_to_pattern(struct test_request *test_rq)
{
int i = 0;
int num_of_dwords = test_rq->buf_size/sizeof(int);
/* num_bytes should be aligned to sizeof(int) */
BUG_ON((test_rq->buf_size % sizeof(int)) != 0);
BUG_ON(test_rq->bios_buffer == NULL);
if (test_rq->wr_rd_data_pattern == TEST_NO_PATTERN)
return 0;
if (test_rq->wr_rd_data_pattern == TEST_PATTERN_SEQUENTIAL) {
for (i = 0; i < num_of_dwords; i++) {
if (test_rq->bios_buffer[i] != i) {
test_pr_err(
"%s: wrong pattern 0x%x in index %d",
__func__, test_rq->bios_buffer[i], i);
return -EINVAL;
}
}
} else {
for (i = 0; i < num_of_dwords; i++) {
if (test_rq->bios_buffer[i] !=
test_rq->wr_rd_data_pattern) {
test_pr_err(
"%s: wrong pattern 0x%x in index %d",
__func__, test_rq->bios_buffer[i], i);
return -EINVAL;
}
}
}
return 0;
}
/*
* Determine if the test passed or failed.
* The function checks the test request completion value and calls
* check_testcase_result for result checking that are specific
* to a test case.
*/
static int check_test_result(struct test_data *td)
{
struct test_request *test_rq;
int res = 0;
static int run;
if (!ptd)
goto err;
list_for_each_entry(test_rq, &ptd->dispatched_queue, queuelist) {
if (!test_rq->rq) {
test_pr_info("%s: req_id %d is contains empty req",
__func__, test_rq->req_id);
continue;
}
if (!test_rq->req_completed) {
test_pr_err("%s: rq %d not completed", __func__,
test_rq->req_id);
res = -EINVAL;
goto err;
}
if ((test_rq->req_result < 0) && !test_rq->is_err_expected) {
test_pr_err(
"%s: rq %d completed with err, not as expected",
__func__, test_rq->req_id);
res = -EINVAL;
goto err;
}
if ((test_rq->req_result == 0) && test_rq->is_err_expected) {
test_pr_err("%s: rq %d succeeded, not as expected",
__func__, test_rq->req_id);
res = -EINVAL;
goto err;
}
if (rq_data_dir(test_rq->rq) == READ) {
res = compare_buffer_to_pattern(test_rq);
if (res) {
test_pr_err("%s: read pattern not as expected",
__func__);
res = -EINVAL;
goto err;
}
}
}
if (td->test_info.check_test_result_fn) {
res = td->test_info.check_test_result_fn(td);
if (res)
goto err;
}
test_pr_info("%s: %s, run# %03d, PASSED",
__func__, get_test_case_str(td), ++run);
td->test_result = TEST_PASSED;
return 0;
err:
test_pr_err("%s: %s, run# %03d, FAILED",
__func__, get_test_case_str(td), ++run);
td->test_result = TEST_FAILED;
return res;
}
/* Create and queue the required requests according to the test case */
static int prepare_test(struct test_data *td)
{
int ret = 0;
if (td->test_info.prepare_test_fn) {
ret = td->test_info.prepare_test_fn(td);
return ret;
}
return 0;
}
/* Run the test */
static int run_test(struct test_data *td)
{
int ret = 0;
if (td->test_info.run_test_fn) {
ret = td->test_info.run_test_fn(td);
return ret;
}
__blk_run_queue(td->req_q);
return 0;
}
/*
* free_test_queue() - Free all allocated test requests in the given test_queue:
* free their requests and BIOs buffer
* @test_queue the test queue to be freed
*/
static void free_test_queue(struct list_head *test_queue)
{
struct test_request *test_rq;
struct bio *bio;
while (!list_empty(test_queue)) {
test_rq = list_entry(test_queue->next, struct test_request,
queuelist);
list_del_init(&test_rq->queuelist);
/*
* If the request was not completed we need to free its BIOs
* and remove it from the packed list
*/
if (!test_rq->req_completed) {
test_pr_info(
"%s: Freeing memory of an uncompleted request",
__func__);
list_del_init(&test_rq->rq->queuelist);
while ((bio = test_rq->rq->bio) != NULL) {
test_rq->rq->bio = bio->bi_next;
bio_put(bio);
}
}
blk_put_request(test_rq->rq);
kfree(test_rq->bios_buffer);
kfree(test_rq);
}
}
/*
* free_test_requests() - Free all allocated test requests in
* all test queues in given test_data.
* @td The test_data struct whos test requests will be
* freed.
*/
static void free_test_requests(struct test_data *td)
{
if (!td)
return;
if (td->urgent_count) {
free_test_queue(&td->urgent_queue);
td->urgent_count = 0;
}
if (td->test_count) {
free_test_queue(&td->test_queue);
td->test_count = 0;
}
if (td->dispatched_count) {
free_test_queue(&td->dispatched_queue);
td->dispatched_count = 0;
}
if (td->reinsert_count) {
free_test_queue(&td->reinsert_queue);
td->reinsert_count = 0;
}
}
/*
* post_test() - Do post test operations. Free the allocated
* test requests, their requests and BIOs buffer.
* @td The test_data struct for the test that has
* ended.
*/
static int post_test(struct test_data *td)
{
int ret = 0;
if (td->test_info.post_test_fn)
ret = td->test_info.post_test_fn(td);
ptd->test_info.testcase = 0;
ptd->test_state = TEST_IDLE;
free_test_requests(td);
return ret;
}
/*
* The timer verifies that the test will be completed even if we don't get
* the completion callback for all the requests.
*/
static void test_timeout_handler(unsigned long data)
{
struct test_data *td = (struct test_data *)data;
test_pr_info("%s: TIMEOUT timer expired", __func__);
td->test_state = TEST_COMPLETED;
wake_up(&td->wait_q);
return;
}
static unsigned int get_timeout_msec(struct test_data *td)
{
if (td->test_info.timeout_msec)
return td->test_info.timeout_msec;
else
return TIMEOUT_TIMER_MS;
}
/**
* test_iosched_start_test() - Prepares and runs the test.
* The members test_duration and test_byte_count of the input
* parameter t_info are modified by this function.
* @t_info: the current test testcase and callbacks
* functions
*
* The function also checks the test result upon test completion
*/
int test_iosched_start_test(struct test_info *t_info)
{
int ret = 0;
unsigned timeout_msec;
int counter = 0;
char *test_name = NULL;
if (!ptd)
return -ENODEV;
if (!t_info) {
ptd->test_result = TEST_FAILED;
return -EINVAL;
}
do {
if (ptd->ignore_round)
/*
* We ignored the last run due to FS write requests.
* Sleep to allow those requests to be issued
*/
msleep(2000);
spin_lock(&ptd->lock);
if (ptd->test_state != TEST_IDLE) {
test_pr_info(
"%s: Another test is running, try again later",
__func__);
spin_unlock(&ptd->lock);
return -EBUSY;
}
if (ptd->start_sector == 0) {
test_pr_err("%s: Invalid start sector", __func__);
ptd->test_result = TEST_FAILED;
spin_unlock(&ptd->lock);
return -EINVAL;
}
memcpy(&ptd->test_info, t_info, sizeof(struct test_info));
ptd->test_result = TEST_NO_RESULT;
ptd->num_of_write_bios = 0;
ptd->unique_next_req_id = UNIQUE_START_REQ_ID;
ptd->wr_rd_next_req_id = WR_RD_START_REQ_ID;
ptd->ignore_round = false;
ptd->fs_wr_reqs_during_test = false;
ptd->test_state = TEST_RUNNING;
spin_unlock(&ptd->lock);
timeout_msec = get_timeout_msec(ptd);
mod_timer(&ptd->timeout_timer, jiffies +
msecs_to_jiffies(timeout_msec));
if (ptd->test_info.get_test_case_str_fn)
test_name = ptd->test_info.get_test_case_str_fn(ptd);
else
test_name = "Unknown testcase";
test_pr_info("%s: Starting test %s", __func__, test_name);
ret = prepare_test(ptd);
if (ret) {
test_pr_err("%s: failed to prepare the test\n",
__func__);
goto error;
}
ptd->test_info.test_duration = jiffies;
ret = run_test(ptd);
if (ret) {
test_pr_err("%s: failed to run the test\n", __func__);
goto error;
}
test_pr_info("%s: Waiting for the test completion", __func__);
wait_event(ptd->wait_q, ptd->test_state == TEST_COMPLETED);
t_info->test_duration = ptd->test_info.test_duration;
t_info->test_byte_count = ptd->test_info.test_byte_count;
del_timer_sync(&ptd->timeout_timer);
ret = check_test_result(ptd);
if (ret) {
test_pr_err("%s: check_test_result failed\n",
__func__);
goto error;
}
ret = post_test(ptd);
if (ret) {
test_pr_err("%s: post_test failed\n", __func__);
goto error;
}
/*
* Wakeup the queue thread to fetch FS requests that might got
* postponded due to the test
*/
__blk_run_queue(ptd->req_q);
if (ptd->ignore_round)
test_pr_info(
"%s: Round canceled (Got wr reqs in the middle)",
__func__);
if (++counter == TEST_MAX_TESTCASE_ROUNDS) {
test_pr_info("%s: Too many rounds, did not succeed...",
__func__);
ptd->test_result = TEST_FAILED;
}
} while ((ptd->ignore_round) && (counter < TEST_MAX_TESTCASE_ROUNDS));
if (ptd->test_result == TEST_PASSED)
return 0;
else
return -EINVAL;
error:
post_test(ptd);
ptd->test_result = TEST_FAILED;
return ret;
}
EXPORT_SYMBOL(test_iosched_start_test);
/**
* test_iosched_register() - register a block device test
* utility.
* @bdt: the block device test type to register
*/
void test_iosched_register(struct blk_dev_test_type *bdt)
{
spin_lock(&blk_dev_test_list_lock);
list_add_tail(&bdt->list, &blk_dev_test_list);
spin_unlock(&blk_dev_test_list_lock);
}
EXPORT_SYMBOL_GPL(test_iosched_register);
/**
* test_iosched_unregister() - unregister a block device test
* utility.
* @bdt: the block device test type to unregister
*/
void test_iosched_unregister(struct blk_dev_test_type *bdt)
{
spin_lock(&blk_dev_test_list_lock);
list_del_init(&bdt->list);
spin_unlock(&blk_dev_test_list_lock);
}
EXPORT_SYMBOL_GPL(test_iosched_unregister);
/**
* test_iosched_set_test_result() - Set the test
* result(PASS/FAIL)
* @test_result: the test result
*/
void test_iosched_set_test_result(int test_result)
{
if (!ptd)
return;
ptd->test_result = test_result;
}
EXPORT_SYMBOL(test_iosched_set_test_result);
/**
* test_iosched_set_ignore_round() - Set the ignore_round flag
* @ignore_round: A flag to indicate if this test round
* should be ignored and re-run
*/
void test_iosched_set_ignore_round(bool ignore_round)
{
if (!ptd)
return;
ptd->ignore_round = ignore_round;
}
EXPORT_SYMBOL(test_iosched_set_ignore_round);
/**
* test_iosched_get_debugfs_tests_root() - returns the root
* debugfs directory for the test_iosched tests
*/
struct dentry *test_iosched_get_debugfs_tests_root(void)
{
if (!ptd)
return NULL;
return ptd->debug.debug_tests_root;
}
EXPORT_SYMBOL(test_iosched_get_debugfs_tests_root);
/**
* test_iosched_get_debugfs_utils_root() - returns the root
* debugfs directory for the test_iosched utils
*/
struct dentry *test_iosched_get_debugfs_utils_root(void)
{
if (!ptd)
return NULL;
return ptd->debug.debug_utils_root;
}
EXPORT_SYMBOL(test_iosched_get_debugfs_utils_root);
static int test_debugfs_init(struct test_data *td)
{
td->debug.debug_root = debugfs_create_dir("test-iosched", NULL);
if (!td->debug.debug_root)
return -ENOENT;
td->debug.debug_tests_root = debugfs_create_dir("tests",
td->debug.debug_root);
if (!td->debug.debug_tests_root)
goto err;
td->debug.debug_utils_root = debugfs_create_dir("utils",
td->debug.debug_root);
if (!td->debug.debug_utils_root)
goto err;
td->debug.debug_test_result = debugfs_create_u32(
"test_result",
S_IRUGO | S_IWUGO,
td->debug.debug_utils_root,
&td->test_result);
if (!td->debug.debug_test_result)
goto err;
td->debug.start_sector = debugfs_create_u32(
"start_sector",
S_IRUGO | S_IWUGO,
td->debug.debug_utils_root,
&td->start_sector);
if (!td->debug.start_sector)
goto err;
return 0;
err:
debugfs_remove_recursive(td->debug.debug_root);
return -ENOENT;
}
static void test_debugfs_cleanup(struct test_data *td)
{
debugfs_remove_recursive(td->debug.debug_root);
}
static void print_req(struct request *req)
{
struct bio *bio;
struct test_request *test_rq;
if (!req)
return;
test_rq = (struct test_request *)req->elv.priv[0];
if (test_rq) {
test_pr_debug("%s: Dispatch request %d: __sector=0x%lx",
__func__, test_rq->req_id, (unsigned long)req->__sector);
test_pr_debug("%s: nr_phys_segments=%d, num_of_sectors=%d",
__func__, req->nr_phys_segments, blk_rq_sectors(req));
bio = req->bio;
test_pr_debug("%s: bio: bi_size=%d, bi_sector=0x%lx",
__func__, bio->bi_size,
(unsigned long)bio->bi_sector);
while ((bio = bio->bi_next) != NULL) {
test_pr_debug("%s: bio: bi_size=%d, bi_sector=0x%lx",
__func__, bio->bi_size,
(unsigned long)bio->bi_sector);
}
}
}
static void test_merged_requests(struct request_queue *q,
struct request *rq, struct request *next)
{
list_del_init(&next->queuelist);
}
/*
* test_dispatch_from(): Dispatch request from @queue to the @dispatched_queue.
* Also update th dispatched_count counter.
*/
static int test_dispatch_from(struct request_queue *q,
struct list_head *queue, unsigned int *count)
{
struct test_request *test_rq;
struct request *rq;
int ret = 0;
if (!ptd)
goto err;
spin_lock_irq(&ptd->lock);
if (!list_empty(queue)) {
test_rq = list_entry(queue->next, struct test_request,
queuelist);
rq = test_rq->rq;
if (!rq) {
pr_err("%s: null request,return", __func__);
spin_unlock_irq(&ptd->lock);
goto err;
}
list_move_tail(&test_rq->queuelist, &ptd->dispatched_queue);
ptd->dispatched_count++;
(*count)--;
spin_unlock_irq(&ptd->lock);
print_req(rq);
elv_dispatch_sort(q, rq);
ptd->test_info.test_byte_count += test_rq->buf_size;
ret = 1;
goto err;
}
spin_unlock_irq(&ptd->lock);
err:
return ret;
}
/*
* Dispatch a test request in case there is a running test Otherwise, dispatch
* a request that was queued by the FS to keep the card functional.
*/
static int test_dispatch_requests(struct request_queue *q, int force)
{
struct test_data *td = q->elevator->elevator_data;
struct request *rq = NULL;
int ret = 0;
switch (td->test_state) {
case TEST_IDLE:
if (!list_empty(&td->queue)) {
rq = list_entry(td->queue.next, struct request,
queuelist);
list_del_init(&rq->queuelist);
elv_dispatch_sort(q, rq);
ret = 1;
goto exit;
}
break;
case TEST_RUNNING:
if (test_dispatch_from(q, &td->urgent_queue,
&td->urgent_count)) {
test_pr_debug("%s: Dispatched from urgent_count=%d",
__func__, ptd->urgent_count);
ret = 1;
goto exit;
}
if (test_dispatch_from(q, &td->reinsert_queue,
&td->reinsert_count)) {
test_pr_debug("%s: Dispatched from reinsert_count=%d",
__func__, ptd->reinsert_count);
ret = 1;
goto exit;
}
if (test_dispatch_from(q, &td->test_queue, &td->test_count)) {
test_pr_debug("%s: Dispatched from test_count=%d",
__func__, ptd->test_count);
ret = 1;
goto exit;
}
break;
case TEST_COMPLETED:
default:
break;
}
exit:
return ret;
}
static void test_add_request(struct request_queue *q, struct request *rq)
{
struct test_data *td = q->elevator->elevator_data;
list_add_tail(&rq->queuelist, &td->queue);
/*
* The write requests can be followed by a FLUSH request that might
* cause unexpected results of the test.
*/
if ((rq_data_dir(rq) == WRITE) && (td->test_state == TEST_RUNNING)) {
test_pr_debug("%s: got WRITE req in the middle of the test",
__func__);
td->fs_wr_reqs_during_test = true;
}
}
static struct request *
test_former_request(struct request_queue *q, struct request *rq)
{
struct test_data *td = q->elevator->elevator_data;
if (rq->queuelist.prev == &td->queue)
return NULL;
return list_entry(rq->queuelist.prev, struct request, queuelist);
}
static struct request *
test_latter_request(struct request_queue *q, struct request *rq)
{
struct test_data *td = q->elevator->elevator_data;
if (rq->queuelist.next == &td->queue)
return NULL;
return list_entry(rq->queuelist.next, struct request, queuelist);
}
static void *test_init_queue(struct request_queue *q)
{
struct blk_dev_test_type *__bdt;
ptd = kmalloc_node(sizeof(struct test_data), GFP_KERNEL,
q->node);
if (!ptd) {
test_pr_err("%s: failed to allocate test data", __func__);
return NULL;
}
memset((void *)ptd, 0, sizeof(struct test_data));
INIT_LIST_HEAD(&ptd->queue);
INIT_LIST_HEAD(&ptd->test_queue);
INIT_LIST_HEAD(&ptd->dispatched_queue);
INIT_LIST_HEAD(&ptd->reinsert_queue);
INIT_LIST_HEAD(&ptd->urgent_queue);
init_waitqueue_head(&ptd->wait_q);
ptd->req_q = q;
setup_timer(&ptd->timeout_timer, test_timeout_handler,
(unsigned long)ptd);
spin_lock_init(&ptd->lock);
if (test_debugfs_init(ptd)) {
test_pr_err("%s: Failed to create debugfs files", __func__);
return NULL;
}
list_for_each_entry(__bdt, &blk_dev_test_list, list)
__bdt->init_fn();
return ptd;
}
static void test_exit_queue(struct elevator_queue *e)
{
struct test_data *td = e->elevator_data;
struct blk_dev_test_type *__bdt;
BUG_ON(!list_empty(&td->queue));
list_for_each_entry(__bdt, &blk_dev_test_list, list)
__bdt->exit_fn();
test_debugfs_cleanup(td);
kfree(td);
}
/**
* test_get_test_data() - Returns a pointer to the test_data
* struct which keeps the current test data.
*
*/
struct test_data *test_get_test_data(void)
{
return ptd;
}
EXPORT_SYMBOL(test_get_test_data);
static bool test_urgent_pending(struct request_queue *q)
{
return !list_empty(&ptd->urgent_queue);
}
/**
* test_iosched_add_urgent_req() - Add an urgent test_request.
* First mark the request as urgent, then add it to the
* urgent_queue test queue.
* @test_rq: pointer to the urgent test_request to be
* added.
*
*/
void test_iosched_add_urgent_req(struct test_request *test_rq)
{
spin_lock_irq(&ptd->lock);
blk_mark_rq_urgent(test_rq->rq);
list_add_tail(&test_rq->queuelist, &ptd->urgent_queue);
ptd->urgent_count++;
spin_unlock_irq(&ptd->lock);
}
EXPORT_SYMBOL(test_iosched_add_urgent_req);
/**
* test_reinsert_req() - Moves the @rq request from
* @dispatched_queue into @reinsert_queue.
* The @rq must be in @dispatched_queue
* @q: request queue
* @rq: request to be inserted
*
*
*/
static int test_reinsert_req(struct request_queue *q,
struct request *rq)
{
struct test_request *test_rq;
int ret = -EINVAL;
if (!ptd)
goto exit;
if (list_empty(&ptd->dispatched_queue)) {
test_pr_err("%s: dispatched_queue is empty", __func__);
goto exit;
}
list_for_each_entry(test_rq, &ptd->dispatched_queue, queuelist) {
if (test_rq->rq == rq) {
list_move(&test_rq->queuelist, &ptd->reinsert_queue);
ptd->dispatched_count--;
ptd->reinsert_count++;
ret = 0;
break;
}
}
exit:
return ret;
}
static struct elevator_type elevator_test_iosched = {
.ops = {
.elevator_merge_req_fn = test_merged_requests,
.elevator_dispatch_fn = test_dispatch_requests,
.elevator_add_req_fn = test_add_request,
.elevator_former_req_fn = test_former_request,
.elevator_latter_req_fn = test_latter_request,
.elevator_init_fn = test_init_queue,
.elevator_exit_fn = test_exit_queue,
.elevator_is_urgent_fn = test_urgent_pending,
.elevator_reinsert_req_fn = test_reinsert_req,
},
.elevator_name = "test-iosched",
.elevator_owner = THIS_MODULE,
};
static int __init test_init(void)
{
elv_register(&elevator_test_iosched);
return 0;
}
static void __exit test_exit(void)
{
elv_unregister(&elevator_test_iosched);
}
module_init(test_init);
module_exit(test_exit);
MODULE_LICENSE("GPL v2");
MODULE_DESCRIPTION("Test IO scheduler");