| /* Copyright (c) 2012-2013, The Linux Foundation. All rights reserved. |
| * |
| * This program is free software; you can redistribute it and/or modify |
| * it under the terms of the GNU General Public License version 2 and |
| * only version 2 as published by the Free Software Foundation. |
| * |
| * This program is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| * GNU General Public License for more details. |
| * |
| */ |
| |
| /* MMC block test */ |
| |
| #include <linux/module.h> |
| #include <linux/blkdev.h> |
| #include <linux/debugfs.h> |
| #include <linux/mmc/card.h> |
| #include <linux/mmc/host.h> |
| #include <linux/delay.h> |
| #include <linux/test-iosched.h> |
| #include <linux/jiffies.h> |
| #include "queue.h" |
| #include <linux/mmc/mmc.h> |
| |
| #define MODULE_NAME "mmc_block_test" |
| #define TEST_MAX_SECTOR_RANGE (600*1024*1024) /* 600 MB */ |
| #define TEST_MAX_BIOS_PER_REQ 128 |
| #define CMD23_PACKED_BIT (1 << 30) |
| #define LARGE_PRIME_1 1103515367 |
| #define LARGE_PRIME_2 35757 |
| #define PACKED_HDR_VER_MASK 0x000000FF |
| #define PACKED_HDR_RW_MASK 0x0000FF00 |
| #define PACKED_HDR_NUM_REQS_MASK 0x00FF0000 |
| #define PACKED_HDR_BITS_16_TO_29_SET 0x3FFF0000 |
| #define SECTOR_SIZE 512 |
| #define NUM_OF_SECTORS_PER_BIO ((BIO_U32_SIZE * 4) / SECTOR_SIZE) |
| #define BIO_TO_SECTOR(x) (x * NUM_OF_SECTORS_PER_BIO) |
| /* the desired long test size to be read */ |
| #define LONG_READ_TEST_MAX_NUM_BYTES (50*1024*1024) /* 50MB */ |
| /* the minimum amount of requests that will be created */ |
| #define LONG_WRITE_TEST_MIN_NUM_REQS 200 /* 100MB */ |
| /* request queue limitation is 128 requests, and we leave 10 spare requests */ |
| #define TEST_MAX_REQUESTS 118 |
| #define LONG_READ_TEST_MAX_NUM_REQS (LONG_READ_TEST_MAX_NUM_BYTES / \ |
| (TEST_MAX_BIOS_PER_REQ * sizeof(int) * BIO_U32_SIZE)) |
| /* this doesn't allow the test requests num to be greater than the maximum */ |
| #define LONG_READ_TEST_ACTUAL_NUM_REQS \ |
| ((TEST_MAX_REQUESTS < LONG_READ_TEST_MAX_NUM_REQS) ? \ |
| TEST_MAX_REQUESTS : LONG_READ_TEST_MAX_NUM_REQS) |
| #define MB_MSEC_RATIO_APPROXIMATION ((1024 * 1024) / 1000) |
| /* actual number of bytes in test */ |
| #define LONG_READ_NUM_BYTES (LONG_READ_TEST_ACTUAL_NUM_REQS * \ |
| (TEST_MAX_BIOS_PER_REQ * sizeof(int) * BIO_U32_SIZE)) |
| /* actual number of MiB in test multiplied by 10, for single digit precision*/ |
| #define BYTE_TO_MB_x_10(x) ((x * 10) / (1024 * 1024)) |
| /* extract integer value */ |
| #define LONG_TEST_SIZE_INTEGER(x) (BYTE_TO_MB_x_10(x) / 10) |
| /* and calculate the MiB value fraction */ |
| #define LONG_TEST_SIZE_FRACTION(x) (BYTE_TO_MB_x_10(x) - \ |
| (LONG_TEST_SIZE_INTEGER(x) * 10)) |
| #define LONG_WRITE_TEST_SLEEP_TIME_MS 5 |
| |
| #define test_pr_debug(fmt, args...) pr_debug("%s: "fmt"\n", MODULE_NAME, args) |
| #define test_pr_info(fmt, args...) pr_info("%s: "fmt"\n", MODULE_NAME, args) |
| #define test_pr_err(fmt, args...) pr_err("%s: "fmt"\n", MODULE_NAME, args) |
| |
| #define SANITIZE_TEST_TIMEOUT 240000 |
| #define NEW_REQ_TEST_SLEEP_TIME 1 |
| #define NEW_REQ_TEST_NUM_BIOS 64 |
| #define TEST_REQUEST_NUM_OF_BIOS 3 |
| |
| #define CHECK_BKOPS_STATS(stats, exp_bkops, exp_hpi, exp_suspend) \ |
| ((stats.bkops != exp_bkops) || \ |
| (stats.hpi != exp_hpi) || \ |
| (stats.suspend != exp_suspend)) |
| #define BKOPS_TEST_TIMEOUT 60000 |
| |
| enum is_random { |
| NON_RANDOM_TEST, |
| RANDOM_TEST, |
| }; |
| |
| enum mmc_block_test_testcases { |
| /* Start of send write packing test group */ |
| SEND_WRITE_PACKING_MIN_TESTCASE, |
| TEST_STOP_DUE_TO_READ = SEND_WRITE_PACKING_MIN_TESTCASE, |
| TEST_STOP_DUE_TO_READ_AFTER_MAX_REQS, |
| TEST_STOP_DUE_TO_FLUSH, |
| TEST_STOP_DUE_TO_FLUSH_AFTER_MAX_REQS, |
| TEST_STOP_DUE_TO_EMPTY_QUEUE, |
| TEST_STOP_DUE_TO_MAX_REQ_NUM, |
| TEST_STOP_DUE_TO_THRESHOLD, |
| SEND_WRITE_PACKING_MAX_TESTCASE = TEST_STOP_DUE_TO_THRESHOLD, |
| |
| /* Start of err check test group */ |
| ERR_CHECK_MIN_TESTCASE, |
| TEST_RET_ABORT = ERR_CHECK_MIN_TESTCASE, |
| TEST_RET_PARTIAL_FOLLOWED_BY_SUCCESS, |
| TEST_RET_PARTIAL_FOLLOWED_BY_ABORT, |
| TEST_RET_PARTIAL_MULTIPLE_UNTIL_SUCCESS, |
| TEST_RET_PARTIAL_MAX_FAIL_IDX, |
| TEST_RET_RETRY, |
| TEST_RET_CMD_ERR, |
| TEST_RET_DATA_ERR, |
| ERR_CHECK_MAX_TESTCASE = TEST_RET_DATA_ERR, |
| |
| /* Start of send invalid test group */ |
| INVALID_CMD_MIN_TESTCASE, |
| TEST_HDR_INVALID_VERSION = INVALID_CMD_MIN_TESTCASE, |
| TEST_HDR_WRONG_WRITE_CODE, |
| TEST_HDR_INVALID_RW_CODE, |
| TEST_HDR_DIFFERENT_ADDRESSES, |
| TEST_HDR_REQ_NUM_SMALLER_THAN_ACTUAL, |
| TEST_HDR_REQ_NUM_LARGER_THAN_ACTUAL, |
| TEST_HDR_CMD23_PACKED_BIT_SET, |
| TEST_CMD23_MAX_PACKED_WRITES, |
| TEST_CMD23_ZERO_PACKED_WRITES, |
| TEST_CMD23_PACKED_BIT_UNSET, |
| TEST_CMD23_REL_WR_BIT_SET, |
| TEST_CMD23_BITS_16TO29_SET, |
| TEST_CMD23_HDR_BLK_NOT_IN_COUNT, |
| INVALID_CMD_MAX_TESTCASE = TEST_CMD23_HDR_BLK_NOT_IN_COUNT, |
| |
| /* |
| * Start of packing control test group. |
| * in these next testcases the abbreviation FB = followed by |
| */ |
| PACKING_CONTROL_MIN_TESTCASE, |
| TEST_PACKING_EXP_ONE_OVER_TRIGGER_FB_READ = |
| PACKING_CONTROL_MIN_TESTCASE, |
| TEST_PACKING_EXP_N_OVER_TRIGGER, |
| TEST_PACKING_EXP_N_OVER_TRIGGER_FB_READ, |
| TEST_PACKING_EXP_N_OVER_TRIGGER_FLUSH_N, |
| TEST_PACKING_EXP_THRESHOLD_OVER_TRIGGER, |
| TEST_PACKING_NOT_EXP_LESS_THAN_TRIGGER_REQUESTS, |
| TEST_PACKING_NOT_EXP_TRIGGER_REQUESTS, |
| TEST_PACKING_NOT_EXP_TRIGGER_READ_TRIGGER, |
| TEST_PACKING_NOT_EXP_TRIGGER_FLUSH_TRIGGER, |
| TEST_PACK_MIX_PACKED_NO_PACKED_PACKED, |
| TEST_PACK_MIX_NO_PACKED_PACKED_NO_PACKED, |
| PACKING_CONTROL_MAX_TESTCASE = TEST_PACK_MIX_NO_PACKED_PACKED_NO_PACKED, |
| |
| TEST_WRITE_DISCARD_SANITIZE_READ, |
| |
| /* Start of bkops test group */ |
| BKOPS_MIN_TESTCASE, |
| BKOPS_DELAYED_WORK_LEVEL_1 = BKOPS_MIN_TESTCASE, |
| BKOPS_DELAYED_WORK_LEVEL_1_HPI, |
| BKOPS_CANCEL_DELAYED_WORK, |
| BKOPS_URGENT_LEVEL_2, |
| BKOPS_URGENT_LEVEL_2_TWO_REQS, |
| BKOPS_URGENT_LEVEL_3, |
| BKOPS_MAX_TESTCASE = BKOPS_URGENT_LEVEL_3, |
| |
| TEST_LONG_SEQUENTIAL_READ, |
| TEST_LONG_SEQUENTIAL_WRITE, |
| |
| TEST_NEW_REQ_NOTIFICATION, |
| }; |
| |
| enum mmc_block_test_group { |
| TEST_NO_GROUP, |
| TEST_GENERAL_GROUP, |
| TEST_SEND_WRITE_PACKING_GROUP, |
| TEST_ERR_CHECK_GROUP, |
| TEST_SEND_INVALID_GROUP, |
| TEST_PACKING_CONTROL_GROUP, |
| TEST_BKOPS_GROUP, |
| TEST_NEW_NOTIFICATION_GROUP, |
| }; |
| |
| enum bkops_test_stages { |
| BKOPS_STAGE_1, |
| BKOPS_STAGE_2, |
| BKOPS_STAGE_3, |
| BKOPS_STAGE_4, |
| }; |
| |
| struct mmc_block_test_debug { |
| struct dentry *send_write_packing_test; |
| struct dentry *err_check_test; |
| struct dentry *send_invalid_packed_test; |
| struct dentry *random_test_seed; |
| struct dentry *packing_control_test; |
| struct dentry *discard_sanitize_test; |
| struct dentry *bkops_test; |
| struct dentry *long_sequential_read_test; |
| struct dentry *long_sequential_write_test; |
| struct dentry *new_req_notification_test; |
| }; |
| |
| struct mmc_block_test_data { |
| /* The number of write requests that the test will issue */ |
| int num_requests; |
| /* The expected write packing statistics for the current test */ |
| struct mmc_wr_pack_stats exp_packed_stats; |
| /* |
| * A user-defined seed for random choices of number of bios written in |
| * a request, and of number of requests issued in a test |
| * This field is randomly updated after each use |
| */ |
| unsigned int random_test_seed; |
| /* A retry counter used in err_check tests */ |
| int err_check_counter; |
| /* Can be one of the values of enum test_group */ |
| enum mmc_block_test_group test_group; |
| /* |
| * Indicates if the current testcase is running with random values of |
| * num_requests and num_bios (in each request) |
| */ |
| int is_random; |
| /* Data structure for debugfs dentrys */ |
| struct mmc_block_test_debug debug; |
| /* |
| * Data structure containing individual test information, including |
| * self-defined specific data |
| */ |
| struct test_info test_info; |
| /* mmc block device test */ |
| struct blk_dev_test_type bdt; |
| /* Current BKOPs test stage */ |
| enum bkops_test_stages bkops_stage; |
| /* A wait queue for BKOPs tests */ |
| wait_queue_head_t bkops_wait_q; |
| /* A counter for the number of test requests completed */ |
| unsigned int completed_req_count; |
| }; |
| |
| static struct mmc_block_test_data *mbtd; |
| |
| void print_mmc_packing_stats(struct mmc_card *card) |
| { |
| int i; |
| int max_num_of_packed_reqs = 0; |
| |
| if ((!card) || (!card->wr_pack_stats.packing_events)) |
| return; |
| |
| max_num_of_packed_reqs = card->ext_csd.max_packed_writes; |
| |
| spin_lock(&card->wr_pack_stats.lock); |
| |
| pr_info("%s: write packing statistics:\n", |
| mmc_hostname(card->host)); |
| |
| for (i = 1 ; i <= max_num_of_packed_reqs ; ++i) { |
| if (card->wr_pack_stats.packing_events[i] != 0) |
| pr_info("%s: Packed %d reqs - %d times\n", |
| mmc_hostname(card->host), i, |
| card->wr_pack_stats.packing_events[i]); |
| } |
| |
| pr_info("%s: stopped packing due to the following reasons:\n", |
| mmc_hostname(card->host)); |
| |
| if (card->wr_pack_stats.pack_stop_reason[EXCEEDS_SEGMENTS]) |
| pr_info("%s: %d times: exceedmax num of segments\n", |
| mmc_hostname(card->host), |
| card->wr_pack_stats.pack_stop_reason[EXCEEDS_SEGMENTS]); |
| if (card->wr_pack_stats.pack_stop_reason[EXCEEDS_SECTORS]) |
| pr_info("%s: %d times: exceeding the max num of sectors\n", |
| mmc_hostname(card->host), |
| card->wr_pack_stats.pack_stop_reason[EXCEEDS_SECTORS]); |
| if (card->wr_pack_stats.pack_stop_reason[WRONG_DATA_DIR]) |
| pr_info("%s: %d times: wrong data direction\n", |
| mmc_hostname(card->host), |
| card->wr_pack_stats.pack_stop_reason[WRONG_DATA_DIR]); |
| if (card->wr_pack_stats.pack_stop_reason[FLUSH_OR_DISCARD]) |
| pr_info("%s: %d times: flush or discard\n", |
| mmc_hostname(card->host), |
| card->wr_pack_stats.pack_stop_reason[FLUSH_OR_DISCARD]); |
| if (card->wr_pack_stats.pack_stop_reason[EMPTY_QUEUE]) |
| pr_info("%s: %d times: empty queue\n", |
| mmc_hostname(card->host), |
| card->wr_pack_stats.pack_stop_reason[EMPTY_QUEUE]); |
| if (card->wr_pack_stats.pack_stop_reason[REL_WRITE]) |
| pr_info("%s: %d times: rel write\n", |
| mmc_hostname(card->host), |
| card->wr_pack_stats.pack_stop_reason[REL_WRITE]); |
| if (card->wr_pack_stats.pack_stop_reason[THRESHOLD]) |
| pr_info("%s: %d times: Threshold\n", |
| mmc_hostname(card->host), |
| card->wr_pack_stats.pack_stop_reason[THRESHOLD]); |
| |
| spin_unlock(&card->wr_pack_stats.lock); |
| } |
| |
| /* |
| * A callback assigned to the packed_test_fn field. |
| * Called from block layer in mmc_blk_packed_hdr_wrq_prep. |
| * Here we alter the packed header or CMD23 in order to send an invalid |
| * packed command to the card. |
| */ |
| static void test_invalid_packed_cmd(struct request_queue *q, |
| struct mmc_queue_req *mqrq) |
| { |
| struct mmc_queue *mq = q->queuedata; |
| u32 *packed_cmd_hdr = mqrq->packed_cmd_hdr; |
| struct request *req = mqrq->req; |
| struct request *second_rq; |
| struct test_request *test_rq; |
| struct mmc_blk_request *brq = &mqrq->brq; |
| int num_requests; |
| int max_packed_reqs; |
| |
| if (!mq) { |
| test_pr_err("%s: NULL mq", __func__); |
| return; |
| } |
| |
| test_rq = (struct test_request *)req->elv.priv[0]; |
| if (!test_rq) { |
| test_pr_err("%s: NULL test_rq", __func__); |
| return; |
| } |
| max_packed_reqs = mq->card->ext_csd.max_packed_writes; |
| |
| switch (mbtd->test_info.testcase) { |
| case TEST_HDR_INVALID_VERSION: |
| test_pr_info("%s: set invalid header version", __func__); |
| /* Put 0 in header version field (1 byte, offset 0 in header) */ |
| packed_cmd_hdr[0] = packed_cmd_hdr[0] & ~PACKED_HDR_VER_MASK; |
| break; |
| case TEST_HDR_WRONG_WRITE_CODE: |
| test_pr_info("%s: wrong write code", __func__); |
| /* Set R/W field with R value (1 byte, offset 1 in header) */ |
| packed_cmd_hdr[0] = packed_cmd_hdr[0] & ~PACKED_HDR_RW_MASK; |
| packed_cmd_hdr[0] = packed_cmd_hdr[0] | 0x00000100; |
| break; |
| case TEST_HDR_INVALID_RW_CODE: |
| test_pr_info("%s: invalid r/w code", __func__); |
| /* Set R/W field with invalid value */ |
| packed_cmd_hdr[0] = packed_cmd_hdr[0] & ~PACKED_HDR_RW_MASK; |
| packed_cmd_hdr[0] = packed_cmd_hdr[0] | 0x00000400; |
| break; |
| case TEST_HDR_DIFFERENT_ADDRESSES: |
| test_pr_info("%s: different addresses", __func__); |
| second_rq = list_entry(req->queuelist.next, struct request, |
| queuelist); |
| test_pr_info("%s: test_rq->sector=%ld, second_rq->sector=%ld", |
| __func__, (long)req->__sector, |
| (long)second_rq->__sector); |
| /* |
| * Put start sector of second write request in the first write |
| * request's cmd25 argument in the packed header |
| */ |
| packed_cmd_hdr[3] = second_rq->__sector; |
| break; |
| case TEST_HDR_REQ_NUM_SMALLER_THAN_ACTUAL: |
| test_pr_info("%s: request num smaller than actual" , __func__); |
| num_requests = (packed_cmd_hdr[0] & PACKED_HDR_NUM_REQS_MASK) |
| >> 16; |
| /* num of entries is decremented by 1 */ |
| num_requests = (num_requests - 1) << 16; |
| /* |
| * Set number of requests field in packed write header to be |
| * smaller than the actual number (1 byte, offset 2 in header) |
| */ |
| packed_cmd_hdr[0] = (packed_cmd_hdr[0] & |
| ~PACKED_HDR_NUM_REQS_MASK) + num_requests; |
| break; |
| case TEST_HDR_REQ_NUM_LARGER_THAN_ACTUAL: |
| test_pr_info("%s: request num larger than actual" , __func__); |
| num_requests = (packed_cmd_hdr[0] & PACKED_HDR_NUM_REQS_MASK) |
| >> 16; |
| /* num of entries is incremented by 1 */ |
| num_requests = (num_requests + 1) << 16; |
| /* |
| * Set number of requests field in packed write header to be |
| * larger than the actual number (1 byte, offset 2 in header). |
| */ |
| packed_cmd_hdr[0] = (packed_cmd_hdr[0] & |
| ~PACKED_HDR_NUM_REQS_MASK) + num_requests; |
| break; |
| case TEST_HDR_CMD23_PACKED_BIT_SET: |
| test_pr_info("%s: header CMD23 packed bit set" , __func__); |
| /* |
| * Set packed bit (bit 30) in cmd23 argument of first and second |
| * write requests in packed write header. |
| * These are located at bytes 2 and 4 in packed write header |
| */ |
| packed_cmd_hdr[2] = packed_cmd_hdr[2] | CMD23_PACKED_BIT; |
| packed_cmd_hdr[4] = packed_cmd_hdr[4] | CMD23_PACKED_BIT; |
| break; |
| case TEST_CMD23_MAX_PACKED_WRITES: |
| test_pr_info("%s: CMD23 request num > max_packed_reqs", |
| __func__); |
| /* |
| * Set the individual packed cmd23 request num to |
| * max_packed_reqs + 1 |
| */ |
| brq->sbc.arg = MMC_CMD23_ARG_PACKED | (max_packed_reqs + 1); |
| break; |
| case TEST_CMD23_ZERO_PACKED_WRITES: |
| test_pr_info("%s: CMD23 request num = 0", __func__); |
| /* Set the individual packed cmd23 request num to zero */ |
| brq->sbc.arg = MMC_CMD23_ARG_PACKED; |
| break; |
| case TEST_CMD23_PACKED_BIT_UNSET: |
| test_pr_info("%s: CMD23 packed bit unset", __func__); |
| /* |
| * Set the individual packed cmd23 packed bit to 0, |
| * although there is a packed write request |
| */ |
| brq->sbc.arg &= ~CMD23_PACKED_BIT; |
| break; |
| case TEST_CMD23_REL_WR_BIT_SET: |
| test_pr_info("%s: CMD23 REL WR bit set", __func__); |
| /* Set the individual packed cmd23 reliable write bit */ |
| brq->sbc.arg = MMC_CMD23_ARG_PACKED | MMC_CMD23_ARG_REL_WR; |
| break; |
| case TEST_CMD23_BITS_16TO29_SET: |
| test_pr_info("%s: CMD23 bits [16-29] set", __func__); |
| brq->sbc.arg = MMC_CMD23_ARG_PACKED | |
| PACKED_HDR_BITS_16_TO_29_SET; |
| break; |
| case TEST_CMD23_HDR_BLK_NOT_IN_COUNT: |
| test_pr_info("%s: CMD23 hdr not in block count", __func__); |
| brq->sbc.arg = MMC_CMD23_ARG_PACKED | |
| ((rq_data_dir(req) == READ) ? 0 : mqrq->packed_blocks); |
| break; |
| default: |
| test_pr_err("%s: unexpected testcase %d", |
| __func__, mbtd->test_info.testcase); |
| break; |
| } |
| } |
| |
| /* |
| * A callback assigned to the err_check_fn field of the mmc_request by the |
| * MMC/card/block layer. |
| * Called upon request completion by the MMC/core layer. |
| * Here we emulate an error return value from the card. |
| */ |
| static int test_err_check(struct mmc_card *card, struct mmc_async_req *areq) |
| { |
| struct mmc_queue_req *mq_rq = container_of(areq, struct mmc_queue_req, |
| mmc_active); |
| struct request_queue *req_q = test_iosched_get_req_queue(); |
| struct mmc_queue *mq; |
| int max_packed_reqs; |
| int ret = 0; |
| struct mmc_blk_request *brq; |
| |
| if (req_q) |
| mq = req_q->queuedata; |
| else { |
| test_pr_err("%s: NULL request_queue", __func__); |
| return 0; |
| } |
| |
| if (!mq) { |
| test_pr_err("%s: %s: NULL mq", __func__, |
| mmc_hostname(card->host)); |
| return 0; |
| } |
| |
| max_packed_reqs = mq->card->ext_csd.max_packed_writes; |
| |
| if (!mq_rq) { |
| test_pr_err("%s: %s: NULL mq_rq", __func__, |
| mmc_hostname(card->host)); |
| return 0; |
| } |
| brq = &mq_rq->brq; |
| |
| switch (mbtd->test_info.testcase) { |
| case TEST_RET_ABORT: |
| test_pr_info("%s: return abort", __func__); |
| ret = MMC_BLK_ABORT; |
| break; |
| case TEST_RET_PARTIAL_FOLLOWED_BY_SUCCESS: |
| test_pr_info("%s: return partial followed by success", |
| __func__); |
| /* |
| * Since in this testcase num_requests is always >= 2, |
| * we can be sure that packed_fail_idx is always >= 1 |
| */ |
| mq_rq->packed_fail_idx = (mbtd->num_requests / 2); |
| test_pr_info("%s: packed_fail_idx = %d" |
| , __func__, mq_rq->packed_fail_idx); |
| mq->err_check_fn = NULL; |
| ret = MMC_BLK_PARTIAL; |
| break; |
| case TEST_RET_PARTIAL_FOLLOWED_BY_ABORT: |
| if (!mbtd->err_check_counter) { |
| test_pr_info("%s: return partial followed by abort", |
| __func__); |
| mbtd->err_check_counter++; |
| /* |
| * Since in this testcase num_requests is always >= 3, |
| * we have that packed_fail_idx is always >= 1 |
| */ |
| mq_rq->packed_fail_idx = (mbtd->num_requests / 2); |
| test_pr_info("%s: packed_fail_idx = %d" |
| , __func__, mq_rq->packed_fail_idx); |
| ret = MMC_BLK_PARTIAL; |
| break; |
| } |
| mbtd->err_check_counter = 0; |
| mq->err_check_fn = NULL; |
| ret = MMC_BLK_ABORT; |
| break; |
| case TEST_RET_PARTIAL_MULTIPLE_UNTIL_SUCCESS: |
| test_pr_info("%s: return partial multiple until success", |
| __func__); |
| if (++mbtd->err_check_counter >= (mbtd->num_requests)) { |
| mq->err_check_fn = NULL; |
| mbtd->err_check_counter = 0; |
| ret = MMC_BLK_PARTIAL; |
| break; |
| } |
| mq_rq->packed_fail_idx = 1; |
| ret = MMC_BLK_PARTIAL; |
| break; |
| case TEST_RET_PARTIAL_MAX_FAIL_IDX: |
| test_pr_info("%s: return partial max fail_idx", __func__); |
| mq_rq->packed_fail_idx = max_packed_reqs - 1; |
| mq->err_check_fn = NULL; |
| ret = MMC_BLK_PARTIAL; |
| break; |
| case TEST_RET_RETRY: |
| test_pr_info("%s: return retry", __func__); |
| ret = MMC_BLK_RETRY; |
| break; |
| case TEST_RET_CMD_ERR: |
| test_pr_info("%s: return cmd err", __func__); |
| ret = MMC_BLK_CMD_ERR; |
| break; |
| case TEST_RET_DATA_ERR: |
| test_pr_info("%s: return data err", __func__); |
| ret = MMC_BLK_DATA_ERR; |
| break; |
| case BKOPS_URGENT_LEVEL_2: |
| case BKOPS_URGENT_LEVEL_3: |
| case BKOPS_URGENT_LEVEL_2_TWO_REQS: |
| if (mbtd->err_check_counter++ == 0) { |
| test_pr_info("%s: simulate an exception from the card", |
| __func__); |
| brq->cmd.resp[0] |= R1_EXCEPTION_EVENT; |
| } |
| mq->err_check_fn = NULL; |
| break; |
| default: |
| test_pr_err("%s: unexpected testcase %d", |
| __func__, mbtd->test_info.testcase); |
| } |
| |
| return ret; |
| } |
| |
| /* |
| * This is a specific implementation for the get_test_case_str_fn function |
| * pointer in the test_info data structure. Given a valid test_data instance, |
| * the function returns a string resembling the test name, based on the testcase |
| */ |
| static char *get_test_case_str(struct test_data *td) |
| { |
| if (!td) { |
| test_pr_err("%s: NULL td", __func__); |
| return NULL; |
| } |
| |
| switch (td->test_info.testcase) { |
| case TEST_STOP_DUE_TO_FLUSH: |
| return "\"stop due to flush\""; |
| case TEST_STOP_DUE_TO_FLUSH_AFTER_MAX_REQS: |
| return "\"stop due to flush after max-1 reqs\""; |
| case TEST_STOP_DUE_TO_READ: |
| return "\"stop due to read\""; |
| case TEST_STOP_DUE_TO_READ_AFTER_MAX_REQS: |
| return "\"stop due to read after max-1 reqs\""; |
| case TEST_STOP_DUE_TO_EMPTY_QUEUE: |
| return "\"stop due to empty queue\""; |
| case TEST_STOP_DUE_TO_MAX_REQ_NUM: |
| return "\"stop due to max req num\""; |
| case TEST_STOP_DUE_TO_THRESHOLD: |
| return "\"stop due to exceeding threshold\""; |
| case TEST_RET_ABORT: |
| return "\"err_check return abort\""; |
| case TEST_RET_PARTIAL_FOLLOWED_BY_SUCCESS: |
| return "\"err_check return partial followed by success\""; |
| case TEST_RET_PARTIAL_FOLLOWED_BY_ABORT: |
| return "\"err_check return partial followed by abort\""; |
| case TEST_RET_PARTIAL_MULTIPLE_UNTIL_SUCCESS: |
| return "\"err_check return partial multiple until success\""; |
| case TEST_RET_PARTIAL_MAX_FAIL_IDX: |
| return "\"err_check return partial max fail index\""; |
| case TEST_RET_RETRY: |
| return "\"err_check return retry\""; |
| case TEST_RET_CMD_ERR: |
| return "\"err_check return cmd error\""; |
| case TEST_RET_DATA_ERR: |
| return "\"err_check return data error\""; |
| case TEST_HDR_INVALID_VERSION: |
| return "\"invalid - wrong header version\""; |
| case TEST_HDR_WRONG_WRITE_CODE: |
| return "\"invalid - wrong write code\""; |
| case TEST_HDR_INVALID_RW_CODE: |
| return "\"invalid - wrong R/W code\""; |
| case TEST_HDR_DIFFERENT_ADDRESSES: |
| return "\"invalid - header different addresses\""; |
| case TEST_HDR_REQ_NUM_SMALLER_THAN_ACTUAL: |
| return "\"invalid - header req num smaller than actual\""; |
| case TEST_HDR_REQ_NUM_LARGER_THAN_ACTUAL: |
| return "\"invalid - header req num larger than actual\""; |
| case TEST_HDR_CMD23_PACKED_BIT_SET: |
| return "\"invalid - header cmd23 packed bit set\""; |
| case TEST_CMD23_MAX_PACKED_WRITES: |
| return "\"invalid - cmd23 max packed writes\""; |
| case TEST_CMD23_ZERO_PACKED_WRITES: |
| return "\"invalid - cmd23 zero packed writes\""; |
| case TEST_CMD23_PACKED_BIT_UNSET: |
| return "\"invalid - cmd23 packed bit unset\""; |
| case TEST_CMD23_REL_WR_BIT_SET: |
| return "\"invalid - cmd23 rel wr bit set\""; |
| case TEST_CMD23_BITS_16TO29_SET: |
| return "\"invalid - cmd23 bits [16-29] set\""; |
| case TEST_CMD23_HDR_BLK_NOT_IN_COUNT: |
| return "\"invalid - cmd23 header block not in count\""; |
| case TEST_PACKING_EXP_N_OVER_TRIGGER: |
| return "\"packing control - pack n\""; |
| case TEST_PACKING_EXP_N_OVER_TRIGGER_FB_READ: |
| return "\"packing control - pack n followed by read\""; |
| case TEST_PACKING_EXP_N_OVER_TRIGGER_FLUSH_N: |
| return "\"packing control - pack n followed by flush\""; |
| case TEST_PACKING_EXP_ONE_OVER_TRIGGER_FB_READ: |
| return "\"packing control - pack one followed by read\""; |
| case TEST_PACKING_EXP_THRESHOLD_OVER_TRIGGER: |
| return "\"packing control - pack threshold\""; |
| case TEST_PACKING_NOT_EXP_LESS_THAN_TRIGGER_REQUESTS: |
| return "\"packing control - no packing\""; |
| case TEST_PACKING_NOT_EXP_TRIGGER_REQUESTS: |
| return "\"packing control - no packing, trigger requests\""; |
| case TEST_PACKING_NOT_EXP_TRIGGER_READ_TRIGGER: |
| return "\"packing control - no pack, trigger-read-trigger\""; |
| case TEST_PACKING_NOT_EXP_TRIGGER_FLUSH_TRIGGER: |
| return "\"packing control- no pack, trigger-flush-trigger\""; |
| case TEST_PACK_MIX_PACKED_NO_PACKED_PACKED: |
| return "\"packing control - mix: pack -> no pack -> pack\""; |
| case TEST_PACK_MIX_NO_PACKED_PACKED_NO_PACKED: |
| return "\"packing control - mix: no pack->pack->no pack\""; |
| case TEST_WRITE_DISCARD_SANITIZE_READ: |
| return "\"write, discard, sanitize\""; |
| case BKOPS_DELAYED_WORK_LEVEL_1: |
| return "\"delayed work BKOPS level 1\""; |
| case BKOPS_DELAYED_WORK_LEVEL_1_HPI: |
| return "\"delayed work BKOPS level 1 with HPI\""; |
| case BKOPS_CANCEL_DELAYED_WORK: |
| return "\"cancel delayed BKOPS work\""; |
| case BKOPS_URGENT_LEVEL_2: |
| return "\"urgent BKOPS level 2\""; |
| case BKOPS_URGENT_LEVEL_2_TWO_REQS: |
| return "\"urgent BKOPS level 2, followed by a request\""; |
| case BKOPS_URGENT_LEVEL_3: |
| return "\"urgent BKOPS level 3\""; |
| case TEST_LONG_SEQUENTIAL_READ: |
| return "\"long sequential read\""; |
| case TEST_LONG_SEQUENTIAL_WRITE: |
| return "\"long sequential write\""; |
| case TEST_NEW_REQ_NOTIFICATION: |
| return "\"new request notification test\""; |
| default: |
| return " Unknown testcase"; |
| } |
| |
| return NULL; |
| } |
| |
| /* |
| * Compare individual testcase's statistics to the expected statistics: |
| * Compare stop reason and number of packing events |
| */ |
| static int check_wr_packing_statistics(struct test_data *td) |
| { |
| struct mmc_wr_pack_stats *mmc_packed_stats; |
| struct mmc_queue *mq = td->req_q->queuedata; |
| int max_packed_reqs = mq->card->ext_csd.max_packed_writes; |
| int i; |
| struct mmc_card *card = mq->card; |
| struct mmc_wr_pack_stats expected_stats; |
| int *stop_reason; |
| int ret = 0; |
| |
| if (!mq) { |
| test_pr_err("%s: NULL mq", __func__); |
| return -EINVAL; |
| } |
| |
| expected_stats = mbtd->exp_packed_stats; |
| |
| mmc_packed_stats = mmc_blk_get_packed_statistics(card); |
| if (!mmc_packed_stats) { |
| test_pr_err("%s: NULL mmc_packed_stats", __func__); |
| return -EINVAL; |
| } |
| |
| if (!mmc_packed_stats->packing_events) { |
| test_pr_err("%s: NULL packing_events", __func__); |
| return -EINVAL; |
| } |
| |
| spin_lock(&mmc_packed_stats->lock); |
| |
| if (!mmc_packed_stats->enabled) { |
| test_pr_err("%s write packing statistics are not enabled", |
| __func__); |
| ret = -EINVAL; |
| goto exit_err; |
| } |
| |
| stop_reason = mmc_packed_stats->pack_stop_reason; |
| |
| for (i = 1; i <= max_packed_reqs; ++i) { |
| if (mmc_packed_stats->packing_events[i] != |
| expected_stats.packing_events[i]) { |
| test_pr_err( |
| "%s: Wrong pack stats in index %d, got %d, expected %d", |
| __func__, i, mmc_packed_stats->packing_events[i], |
| expected_stats.packing_events[i]); |
| if (td->fs_wr_reqs_during_test) |
| goto cancel_round; |
| ret = -EINVAL; |
| goto exit_err; |
| } |
| } |
| |
| if (mmc_packed_stats->pack_stop_reason[EXCEEDS_SEGMENTS] != |
| expected_stats.pack_stop_reason[EXCEEDS_SEGMENTS]) { |
| test_pr_err( |
| "%s: Wrong pack stop reason EXCEEDS_SEGMENTS %d, expected %d", |
| __func__, stop_reason[EXCEEDS_SEGMENTS], |
| expected_stats.pack_stop_reason[EXCEEDS_SEGMENTS]); |
| if (td->fs_wr_reqs_during_test) |
| goto cancel_round; |
| ret = -EINVAL; |
| goto exit_err; |
| } |
| |
| if (mmc_packed_stats->pack_stop_reason[EXCEEDS_SECTORS] != |
| expected_stats.pack_stop_reason[EXCEEDS_SECTORS]) { |
| test_pr_err( |
| "%s: Wrong pack stop reason EXCEEDS_SECTORS %d, expected %d", |
| __func__, stop_reason[EXCEEDS_SECTORS], |
| expected_stats.pack_stop_reason[EXCEEDS_SECTORS]); |
| if (td->fs_wr_reqs_during_test) |
| goto cancel_round; |
| ret = -EINVAL; |
| goto exit_err; |
| } |
| |
| if (mmc_packed_stats->pack_stop_reason[WRONG_DATA_DIR] != |
| expected_stats.pack_stop_reason[WRONG_DATA_DIR]) { |
| test_pr_err( |
| "%s: Wrong pack stop reason WRONG_DATA_DIR %d, expected %d", |
| __func__, stop_reason[WRONG_DATA_DIR], |
| expected_stats.pack_stop_reason[WRONG_DATA_DIR]); |
| if (td->fs_wr_reqs_during_test) |
| goto cancel_round; |
| ret = -EINVAL; |
| goto exit_err; |
| } |
| |
| if (mmc_packed_stats->pack_stop_reason[FLUSH_OR_DISCARD] != |
| expected_stats.pack_stop_reason[FLUSH_OR_DISCARD]) { |
| test_pr_err( |
| "%s: Wrong pack stop reason FLUSH_OR_DISCARD %d, expected %d", |
| __func__, stop_reason[FLUSH_OR_DISCARD], |
| expected_stats.pack_stop_reason[FLUSH_OR_DISCARD]); |
| if (td->fs_wr_reqs_during_test) |
| goto cancel_round; |
| ret = -EINVAL; |
| goto exit_err; |
| } |
| |
| if (mmc_packed_stats->pack_stop_reason[EMPTY_QUEUE] != |
| expected_stats.pack_stop_reason[EMPTY_QUEUE]) { |
| test_pr_err( |
| "%s: Wrong pack stop reason EMPTY_QUEUE %d, expected %d", |
| __func__, stop_reason[EMPTY_QUEUE], |
| expected_stats.pack_stop_reason[EMPTY_QUEUE]); |
| if (td->fs_wr_reqs_during_test) |
| goto cancel_round; |
| ret = -EINVAL; |
| goto exit_err; |
| } |
| |
| if (mmc_packed_stats->pack_stop_reason[REL_WRITE] != |
| expected_stats.pack_stop_reason[REL_WRITE]) { |
| test_pr_err( |
| "%s: Wrong pack stop reason REL_WRITE %d, expected %d", |
| __func__, stop_reason[REL_WRITE], |
| expected_stats.pack_stop_reason[REL_WRITE]); |
| if (td->fs_wr_reqs_during_test) |
| goto cancel_round; |
| ret = -EINVAL; |
| goto exit_err; |
| } |
| |
| exit_err: |
| spin_unlock(&mmc_packed_stats->lock); |
| if (ret && mmc_packed_stats->enabled) |
| print_mmc_packing_stats(card); |
| return ret; |
| cancel_round: |
| spin_unlock(&mmc_packed_stats->lock); |
| test_iosched_set_ignore_round(true); |
| return 0; |
| } |
| |
| /* |
| * Pseudo-randomly choose a seed based on the last seed, and update it in |
| * seed_number. then return seed_number (mod max_val), or min_val. |
| */ |
| static unsigned int pseudo_random_seed(unsigned int *seed_number, |
| unsigned int min_val, |
| unsigned int max_val) |
| { |
| int ret = 0; |
| |
| if (!seed_number) |
| return 0; |
| |
| *seed_number = ((unsigned int)(((unsigned long)*seed_number * |
| (unsigned long)LARGE_PRIME_1) + LARGE_PRIME_2)); |
| ret = (unsigned int)((*seed_number) % max_val); |
| |
| return (ret > min_val ? ret : min_val); |
| } |
| |
| /* |
| * Given a pseudo-random seed, find a pseudo-random num_of_bios. |
| * Make sure that num_of_bios is not larger than TEST_MAX_SECTOR_RANGE |
| */ |
| static void pseudo_rnd_num_of_bios(unsigned int *num_bios_seed, |
| unsigned int *num_of_bios) |
| { |
| do { |
| *num_of_bios = pseudo_random_seed(num_bios_seed, 1, |
| TEST_MAX_BIOS_PER_REQ); |
| if (!(*num_of_bios)) |
| *num_of_bios = 1; |
| } while ((*num_of_bios) * BIO_U32_SIZE * 4 > TEST_MAX_SECTOR_RANGE); |
| } |
| |
| /* Add a single read request to the given td's request queue */ |
| static int prepare_request_add_read(struct test_data *td) |
| { |
| int ret; |
| int start_sec; |
| |
| if (td) |
| start_sec = td->start_sector; |
| else { |
| test_pr_err("%s: NULL td", __func__); |
| return 0; |
| } |
| |
| test_pr_info("%s: Adding a read request, first req_id=%d", __func__, |
| td->wr_rd_next_req_id); |
| |
| ret = test_iosched_add_wr_rd_test_req(0, READ, start_sec, 2, |
| TEST_PATTERN_5A, NULL); |
| if (ret) { |
| test_pr_err("%s: failed to add a read request", __func__); |
| return ret; |
| } |
| |
| return 0; |
| } |
| |
| /* Add a single flush request to the given td's request queue */ |
| static int prepare_request_add_flush(struct test_data *td) |
| { |
| int ret; |
| |
| if (!td) { |
| test_pr_err("%s: NULL td", __func__); |
| return 0; |
| } |
| |
| test_pr_info("%s: Adding a flush request, first req_id=%d", __func__, |
| td->unique_next_req_id); |
| ret = test_iosched_add_unique_test_req(0, REQ_UNIQUE_FLUSH, |
| 0, 0, NULL); |
| if (ret) { |
| test_pr_err("%s: failed to add a flush request", __func__); |
| return ret; |
| } |
| |
| return ret; |
| } |
| |
| /* |
| * Add num_requets amount of write requests to the given td's request queue. |
| * If random test mode is chosen we pseudo-randomly choose the number of bios |
| * for each write request, otherwise add between 1 to 5 bio per request. |
| */ |
| static int prepare_request_add_write_reqs(struct test_data *td, |
| int num_requests, int is_err_expected, |
| int is_random) |
| { |
| int i; |
| unsigned int start_sec; |
| int num_bios; |
| int ret = 0; |
| unsigned int *bio_seed = &mbtd->random_test_seed; |
| |
| if (td) |
| start_sec = td->start_sector; |
| else { |
| test_pr_err("%s: NULL td", __func__); |
| return ret; |
| } |
| |
| test_pr_info("%s: Adding %d write requests, first req_id=%d", __func__, |
| num_requests, td->wr_rd_next_req_id); |
| |
| for (i = 1 ; i <= num_requests ; i++) { |
| start_sec = |
| td->start_sector + sizeof(int) * |
| BIO_U32_SIZE * td->num_of_write_bios; |
| if (is_random) |
| pseudo_rnd_num_of_bios(bio_seed, &num_bios); |
| else |
| /* |
| * For the non-random case, give num_bios a value |
| * between 1 and 5, to keep a small number of BIOs |
| */ |
| num_bios = (i%5)+1; |
| |
| ret = test_iosched_add_wr_rd_test_req(is_err_expected, WRITE, |
| start_sec, num_bios, TEST_PATTERN_5A, NULL); |
| |
| if (ret) { |
| test_pr_err("%s: failed to add a write request", |
| __func__); |
| return ret; |
| } |
| } |
| return 0; |
| } |
| |
| /* |
| * Prepare the write, read and flush requests for a generic packed commands |
| * testcase |
| */ |
| static int prepare_packed_requests(struct test_data *td, int is_err_expected, |
| int num_requests, int is_random) |
| { |
| int ret = 0; |
| struct mmc_queue *mq; |
| int max_packed_reqs; |
| struct request_queue *req_q; |
| |
| if (!td) { |
| pr_err("%s: NULL td", __func__); |
| return -EINVAL; |
| } |
| |
| req_q = td->req_q; |
| |
| if (!req_q) { |
| pr_err("%s: NULL request queue", __func__); |
| return -EINVAL; |
| } |
| |
| mq = req_q->queuedata; |
| if (!mq) { |
| test_pr_err("%s: NULL mq", __func__); |
| return -EINVAL; |
| } |
| |
| max_packed_reqs = mq->card->ext_csd.max_packed_writes; |
| |
| if (mbtd->random_test_seed <= 0) { |
| mbtd->random_test_seed = |
| (unsigned int)(get_jiffies_64() & 0xFFFF); |
| test_pr_info("%s: got seed from jiffies %d", |
| __func__, mbtd->random_test_seed); |
| } |
| |
| ret = prepare_request_add_write_reqs(td, num_requests, is_err_expected, |
| is_random); |
| if (ret) |
| return ret; |
| |
| /* Avoid memory corruption in upcoming stats set */ |
| if (td->test_info.testcase == TEST_STOP_DUE_TO_THRESHOLD) |
| num_requests--; |
| |
| memset((void *)mbtd->exp_packed_stats.pack_stop_reason, 0, |
| sizeof(mbtd->exp_packed_stats.pack_stop_reason)); |
| memset(mbtd->exp_packed_stats.packing_events, 0, |
| (max_packed_reqs + 1) * sizeof(u32)); |
| if (num_requests <= max_packed_reqs) |
| mbtd->exp_packed_stats.packing_events[num_requests] = 1; |
| |
| switch (td->test_info.testcase) { |
| case TEST_STOP_DUE_TO_FLUSH: |
| case TEST_STOP_DUE_TO_FLUSH_AFTER_MAX_REQS: |
| ret = prepare_request_add_flush(td); |
| if (ret) |
| return ret; |
| |
| mbtd->exp_packed_stats.pack_stop_reason[FLUSH_OR_DISCARD] = 1; |
| break; |
| case TEST_STOP_DUE_TO_READ: |
| case TEST_STOP_DUE_TO_READ_AFTER_MAX_REQS: |
| ret = prepare_request_add_read(td); |
| if (ret) |
| return ret; |
| |
| mbtd->exp_packed_stats.pack_stop_reason[WRONG_DATA_DIR] = 1; |
| break; |
| case TEST_STOP_DUE_TO_THRESHOLD: |
| mbtd->exp_packed_stats.packing_events[num_requests] = 1; |
| mbtd->exp_packed_stats.packing_events[1] = 1; |
| mbtd->exp_packed_stats.pack_stop_reason[THRESHOLD] = 1; |
| mbtd->exp_packed_stats.pack_stop_reason[EMPTY_QUEUE] = 1; |
| break; |
| case TEST_STOP_DUE_TO_MAX_REQ_NUM: |
| case TEST_RET_PARTIAL_MAX_FAIL_IDX: |
| mbtd->exp_packed_stats.pack_stop_reason[THRESHOLD] = 1; |
| break; |
| default: |
| mbtd->exp_packed_stats.pack_stop_reason[EMPTY_QUEUE] = 1; |
| } |
| mbtd->num_requests = num_requests; |
| |
| return 0; |
| } |
| |
| /* |
| * Prepare the write, read and flush requests for the packing control |
| * testcases |
| */ |
| static int prepare_packed_control_tests_requests(struct test_data *td, |
| int is_err_expected, int num_requests, int is_random) |
| { |
| int ret = 0; |
| struct mmc_queue *mq; |
| int max_packed_reqs; |
| int temp_num_req = num_requests; |
| struct request_queue *req_q; |
| int test_packed_trigger; |
| int num_packed_reqs; |
| |
| if (!td) { |
| test_pr_err("%s: NULL td\n", __func__); |
| return -EINVAL; |
| } |
| |
| req_q = td->req_q; |
| |
| if (!req_q) { |
| test_pr_err("%s: NULL request queue\n", __func__); |
| return -EINVAL; |
| } |
| |
| mq = req_q->queuedata; |
| if (!mq) { |
| test_pr_err("%s: NULL mq", __func__); |
| return -EINVAL; |
| } |
| |
| max_packed_reqs = mq->card->ext_csd.max_packed_writes; |
| test_packed_trigger = mq->num_wr_reqs_to_start_packing; |
| num_packed_reqs = num_requests - test_packed_trigger; |
| |
| if (mbtd->random_test_seed == 0) { |
| mbtd->random_test_seed = |
| (unsigned int)(get_jiffies_64() & 0xFFFF); |
| test_pr_info("%s: got seed from jiffies %d", |
| __func__, mbtd->random_test_seed); |
| } |
| |
| if (td->test_info.testcase == |
| TEST_PACK_MIX_NO_PACKED_PACKED_NO_PACKED) { |
| temp_num_req = num_requests; |
| num_requests = test_packed_trigger - 1; |
| } |
| |
| /* Verify that the packing is disabled before starting the test */ |
| mq->wr_packing_enabled = false; |
| mq->num_of_potential_packed_wr_reqs = 0; |
| |
| if (td->test_info.testcase == TEST_PACK_MIX_PACKED_NO_PACKED_PACKED) { |
| mq->num_of_potential_packed_wr_reqs = test_packed_trigger + 1; |
| mq->wr_packing_enabled = true; |
| num_requests = test_packed_trigger + 2; |
| } |
| |
| ret = prepare_request_add_write_reqs(td, num_requests, is_err_expected, |
| is_random); |
| if (ret) |
| goto exit; |
| |
| if (td->test_info.testcase == TEST_PACK_MIX_NO_PACKED_PACKED_NO_PACKED) |
| num_requests = temp_num_req; |
| |
| memset((void *)mbtd->exp_packed_stats.pack_stop_reason, 0, |
| sizeof(mbtd->exp_packed_stats.pack_stop_reason)); |
| memset(mbtd->exp_packed_stats.packing_events, 0, |
| (max_packed_reqs + 1) * sizeof(u32)); |
| |
| switch (td->test_info.testcase) { |
| case TEST_PACKING_EXP_N_OVER_TRIGGER_FB_READ: |
| case TEST_PACKING_EXP_ONE_OVER_TRIGGER_FB_READ: |
| ret = prepare_request_add_read(td); |
| if (ret) |
| goto exit; |
| |
| mbtd->exp_packed_stats.pack_stop_reason[WRONG_DATA_DIR] = 1; |
| mbtd->exp_packed_stats.packing_events[num_packed_reqs] = 1; |
| break; |
| case TEST_PACKING_EXP_N_OVER_TRIGGER_FLUSH_N: |
| ret = prepare_request_add_flush(td); |
| if (ret) |
| goto exit; |
| |
| ret = prepare_request_add_write_reqs(td, num_packed_reqs, |
| is_err_expected, is_random); |
| if (ret) |
| goto exit; |
| |
| mbtd->exp_packed_stats.pack_stop_reason[EMPTY_QUEUE] = 1; |
| mbtd->exp_packed_stats.pack_stop_reason[FLUSH_OR_DISCARD] = 1; |
| mbtd->exp_packed_stats.packing_events[num_packed_reqs] = 2; |
| break; |
| case TEST_PACKING_NOT_EXP_TRIGGER_READ_TRIGGER: |
| ret = prepare_request_add_read(td); |
| if (ret) |
| goto exit; |
| |
| ret = prepare_request_add_write_reqs(td, test_packed_trigger, |
| is_err_expected, is_random); |
| if (ret) |
| goto exit; |
| |
| mbtd->exp_packed_stats.packing_events[num_packed_reqs] = 1; |
| break; |
| case TEST_PACKING_NOT_EXP_TRIGGER_FLUSH_TRIGGER: |
| ret = prepare_request_add_flush(td); |
| if (ret) |
| goto exit; |
| |
| ret = prepare_request_add_write_reqs(td, test_packed_trigger, |
| is_err_expected, is_random); |
| if (ret) |
| goto exit; |
| |
| mbtd->exp_packed_stats.packing_events[num_packed_reqs] = 1; |
| break; |
| case TEST_PACK_MIX_PACKED_NO_PACKED_PACKED: |
| ret = prepare_request_add_read(td); |
| if (ret) |
| goto exit; |
| |
| ret = prepare_request_add_write_reqs(td, test_packed_trigger-1, |
| is_err_expected, is_random); |
| if (ret) |
| goto exit; |
| |
| ret = prepare_request_add_write_reqs(td, num_requests, |
| is_err_expected, is_random); |
| if (ret) |
| goto exit; |
| |
| mbtd->exp_packed_stats.packing_events[num_requests] = 1; |
| mbtd->exp_packed_stats.packing_events[num_requests-1] = 1; |
| mbtd->exp_packed_stats.pack_stop_reason[WRONG_DATA_DIR] = 1; |
| mbtd->exp_packed_stats.pack_stop_reason[EMPTY_QUEUE] = 1; |
| break; |
| case TEST_PACK_MIX_NO_PACKED_PACKED_NO_PACKED: |
| ret = prepare_request_add_read(td); |
| if (ret) |
| goto exit; |
| |
| ret = prepare_request_add_write_reqs(td, num_requests, |
| is_err_expected, is_random); |
| if (ret) |
| goto exit; |
| |
| ret = prepare_request_add_read(td); |
| if (ret) |
| goto exit; |
| |
| ret = prepare_request_add_write_reqs(td, test_packed_trigger-1, |
| is_err_expected, is_random); |
| if (ret) |
| goto exit; |
| |
| mbtd->exp_packed_stats.pack_stop_reason[WRONG_DATA_DIR] = 1; |
| mbtd->exp_packed_stats.packing_events[num_packed_reqs] = 1; |
| break; |
| case TEST_PACKING_NOT_EXP_LESS_THAN_TRIGGER_REQUESTS: |
| case TEST_PACKING_NOT_EXP_TRIGGER_REQUESTS: |
| break; |
| default: |
| BUG_ON(num_packed_reqs < 0); |
| mbtd->exp_packed_stats.pack_stop_reason[EMPTY_QUEUE] = 1; |
| mbtd->exp_packed_stats.packing_events[num_packed_reqs] = 1; |
| } |
| mbtd->num_requests = num_requests; |
| |
| exit: |
| return ret; |
| } |
| |
| /* |
| * Prepare requests for the TEST_RET_PARTIAL_FOLLOWED_BY_ABORT testcase. |
| * In this testcase we have mixed error expectations from different |
| * write requests, hence the special prepare function. |
| */ |
| static int prepare_partial_followed_by_abort(struct test_data *td, |
| int num_requests) |
| { |
| int i, start_address; |
| int is_err_expected = 0; |
| int ret = 0; |
| struct mmc_queue *mq = test_iosched_get_req_queue()->queuedata; |
| int max_packed_reqs; |
| |
| if (!mq) { |
| test_pr_err("%s: NULL mq", __func__); |
| return -EINVAL; |
| } |
| |
| max_packed_reqs = mq->card->ext_csd.max_packed_writes; |
| |
| for (i = 1; i <= num_requests; i++) { |
| if (i > (num_requests / 2)) |
| is_err_expected = 1; |
| |
| start_address = td->start_sector + |
| sizeof(int) * BIO_U32_SIZE * td->num_of_write_bios; |
| ret = test_iosched_add_wr_rd_test_req(is_err_expected, WRITE, |
| start_address, (i % 5) + 1, TEST_PATTERN_5A, |
| NULL); |
| if (ret) { |
| test_pr_err("%s: failed to add a write request", |
| __func__); |
| return ret; |
| } |
| } |
| |
| memset((void *)&mbtd->exp_packed_stats.pack_stop_reason, 0, |
| sizeof(mbtd->exp_packed_stats.pack_stop_reason)); |
| memset(mbtd->exp_packed_stats.packing_events, 0, |
| (max_packed_reqs + 1) * sizeof(u32)); |
| mbtd->exp_packed_stats.packing_events[num_requests] = 1; |
| mbtd->exp_packed_stats.pack_stop_reason[EMPTY_QUEUE] = 1; |
| |
| mbtd->num_requests = num_requests; |
| |
| return ret; |
| } |
| |
| /* |
| * Get number of write requests for current testcase. If random test mode was |
| * chosen, pseudo-randomly choose the number of requests, otherwise set to |
| * two less than the packing threshold. |
| */ |
| static int get_num_requests(struct test_data *td) |
| { |
| int *seed = &mbtd->random_test_seed; |
| struct request_queue *req_q; |
| struct mmc_queue *mq; |
| int max_num_requests; |
| int num_requests; |
| int min_num_requests = 2; |
| int is_random = mbtd->is_random; |
| int max_for_double; |
| int test_packed_trigger; |
| |
| req_q = test_iosched_get_req_queue(); |
| if (req_q) |
| mq = req_q->queuedata; |
| else { |
| test_pr_err("%s: NULL request queue", __func__); |
| return 0; |
| } |
| |
| if (!mq) { |
| test_pr_err("%s: NULL mq", __func__); |
| return -EINVAL; |
| } |
| |
| max_num_requests = mq->card->ext_csd.max_packed_writes; |
| num_requests = max_num_requests - 2; |
| test_packed_trigger = mq->num_wr_reqs_to_start_packing; |
| |
| /* |
| * Here max_for_double is intended for packed control testcases |
| * in which we issue many write requests. It's purpose is to prevent |
| * exceeding max number of req_queue requests. |
| */ |
| max_for_double = max_num_requests - 10; |
| |
| if (td->test_info.testcase == |
| TEST_PACKING_NOT_EXP_LESS_THAN_TRIGGER_REQUESTS) |
| /* Don't expect packing, so issue up to trigger-1 reqs */ |
| num_requests = test_packed_trigger - 1; |
| |
| if (is_random) { |
| if (td->test_info.testcase == |
| TEST_RET_PARTIAL_FOLLOWED_BY_ABORT) |
| /* |
| * Here we don't want num_requests to be less than 1 |
| * as a consequence of division by 2. |
| */ |
| min_num_requests = 3; |
| |
| if (td->test_info.testcase == |
| TEST_PACKING_NOT_EXP_LESS_THAN_TRIGGER_REQUESTS) |
| /* Don't expect packing, so issue up to trigger reqs */ |
| max_num_requests = test_packed_trigger; |
| |
| num_requests = pseudo_random_seed(seed, min_num_requests, |
| max_num_requests - 1); |
| } |
| |
| if (td->test_info.testcase == |
| TEST_PACKING_NOT_EXP_LESS_THAN_TRIGGER_REQUESTS) |
| num_requests -= test_packed_trigger; |
| |
| if (td->test_info.testcase == TEST_PACKING_EXP_N_OVER_TRIGGER_FLUSH_N) |
| num_requests = |
| num_requests > max_for_double ? max_for_double : num_requests; |
| |
| if (mbtd->test_group == TEST_PACKING_CONTROL_GROUP) |
| num_requests += test_packed_trigger; |
| |
| if (td->test_info.testcase == TEST_PACKING_NOT_EXP_TRIGGER_REQUESTS) |
| num_requests = test_packed_trigger; |
| |
| return num_requests; |
| } |
| |
| static int prepare_long_read_test_requests(struct test_data *td) |
| { |
| |
| int ret; |
| int start_sec; |
| int j; |
| |
| if (td) |
| start_sec = td->start_sector; |
| else { |
| test_pr_err("%s: NULL td\n", __func__); |
| return -EINVAL; |
| } |
| |
| test_pr_info("%s: Adding %d read requests, first req_id=%d", __func__, |
| LONG_READ_TEST_ACTUAL_NUM_REQS, td->wr_rd_next_req_id); |
| |
| for (j = 0; j < LONG_READ_TEST_ACTUAL_NUM_REQS; j++) { |
| |
| ret = test_iosched_add_wr_rd_test_req(0, READ, |
| start_sec, |
| TEST_MAX_BIOS_PER_REQ, |
| TEST_NO_PATTERN, NULL); |
| if (ret) { |
| test_pr_err("%s: failed to add a read request, err = %d" |
| , __func__, ret); |
| return ret; |
| } |
| |
| start_sec += |
| (TEST_MAX_BIOS_PER_REQ * sizeof(int) * BIO_U32_SIZE); |
| } |
| |
| return 0; |
| } |
| |
| /* |
| * An implementation for the prepare_test_fn pointer in the test_info |
| * data structure. According to the testcase we add the right number of requests |
| * and decide if an error is expected or not. |
| */ |
| static int prepare_test(struct test_data *td) |
| { |
| struct mmc_queue *mq = test_iosched_get_req_queue()->queuedata; |
| int max_num_requests; |
| int num_requests = 0; |
| int ret = 0; |
| int is_random = mbtd->is_random; |
| int test_packed_trigger = mq->num_wr_reqs_to_start_packing; |
| |
| if (!mq) { |
| test_pr_err("%s: NULL mq", __func__); |
| return -EINVAL; |
| } |
| |
| max_num_requests = mq->card->ext_csd.max_packed_writes; |
| |
| if (is_random && mbtd->random_test_seed == 0) { |
| mbtd->random_test_seed = |
| (unsigned int)(get_jiffies_64() & 0xFFFF); |
| test_pr_info("%s: got seed from jiffies %d", |
| __func__, mbtd->random_test_seed); |
| } |
| |
| num_requests = get_num_requests(td); |
| |
| if (mbtd->test_group == TEST_SEND_INVALID_GROUP) |
| mq->packed_test_fn = |
| test_invalid_packed_cmd; |
| |
| if (mbtd->test_group == TEST_ERR_CHECK_GROUP) |
| mq->err_check_fn = test_err_check; |
| |
| switch (td->test_info.testcase) { |
| case TEST_STOP_DUE_TO_FLUSH: |
| case TEST_STOP_DUE_TO_READ: |
| case TEST_RET_PARTIAL_FOLLOWED_BY_SUCCESS: |
| case TEST_RET_PARTIAL_MULTIPLE_UNTIL_SUCCESS: |
| case TEST_STOP_DUE_TO_EMPTY_QUEUE: |
| case TEST_CMD23_PACKED_BIT_UNSET: |
| ret = prepare_packed_requests(td, 0, num_requests, is_random); |
| break; |
| case TEST_STOP_DUE_TO_FLUSH_AFTER_MAX_REQS: |
| case TEST_STOP_DUE_TO_READ_AFTER_MAX_REQS: |
| ret = prepare_packed_requests(td, 0, max_num_requests - 1, |
| is_random); |
| break; |
| case TEST_RET_PARTIAL_FOLLOWED_BY_ABORT: |
| ret = prepare_partial_followed_by_abort(td, num_requests); |
| break; |
| case TEST_STOP_DUE_TO_MAX_REQ_NUM: |
| case TEST_RET_PARTIAL_MAX_FAIL_IDX: |
| ret = prepare_packed_requests(td, 0, max_num_requests, |
| is_random); |
| break; |
| case TEST_STOP_DUE_TO_THRESHOLD: |
| ret = prepare_packed_requests(td, 0, max_num_requests + 1, |
| is_random); |
| break; |
| case TEST_RET_ABORT: |
| case TEST_RET_RETRY: |
| case TEST_RET_CMD_ERR: |
| case TEST_RET_DATA_ERR: |
| case TEST_HDR_INVALID_VERSION: |
| case TEST_HDR_WRONG_WRITE_CODE: |
| case TEST_HDR_INVALID_RW_CODE: |
| case TEST_HDR_DIFFERENT_ADDRESSES: |
| case TEST_HDR_REQ_NUM_SMALLER_THAN_ACTUAL: |
| case TEST_HDR_REQ_NUM_LARGER_THAN_ACTUAL: |
| case TEST_CMD23_MAX_PACKED_WRITES: |
| case TEST_CMD23_ZERO_PACKED_WRITES: |
| case TEST_CMD23_REL_WR_BIT_SET: |
| case TEST_CMD23_BITS_16TO29_SET: |
| case TEST_CMD23_HDR_BLK_NOT_IN_COUNT: |
| case TEST_HDR_CMD23_PACKED_BIT_SET: |
| ret = prepare_packed_requests(td, 1, num_requests, is_random); |
| break; |
| case TEST_PACKING_EXP_N_OVER_TRIGGER: |
| case TEST_PACKING_EXP_N_OVER_TRIGGER_FB_READ: |
| case TEST_PACKING_NOT_EXP_TRIGGER_REQUESTS: |
| case TEST_PACKING_NOT_EXP_LESS_THAN_TRIGGER_REQUESTS: |
| case TEST_PACK_MIX_PACKED_NO_PACKED_PACKED: |
| case TEST_PACK_MIX_NO_PACKED_PACKED_NO_PACKED: |
| ret = prepare_packed_control_tests_requests(td, 0, num_requests, |
| is_random); |
| break; |
| case TEST_PACKING_EXP_THRESHOLD_OVER_TRIGGER: |
| ret = prepare_packed_control_tests_requests(td, 0, |
| max_num_requests, is_random); |
| break; |
| case TEST_PACKING_EXP_ONE_OVER_TRIGGER_FB_READ: |
| ret = prepare_packed_control_tests_requests(td, 0, |
| test_packed_trigger + 1, |
| is_random); |
| break; |
| case TEST_PACKING_EXP_N_OVER_TRIGGER_FLUSH_N: |
| ret = prepare_packed_control_tests_requests(td, 0, num_requests, |
| is_random); |
| break; |
| case TEST_PACKING_NOT_EXP_TRIGGER_READ_TRIGGER: |
| case TEST_PACKING_NOT_EXP_TRIGGER_FLUSH_TRIGGER: |
| ret = prepare_packed_control_tests_requests(td, 0, |
| test_packed_trigger, is_random); |
| break; |
| case TEST_LONG_SEQUENTIAL_WRITE: |
| case TEST_LONG_SEQUENTIAL_READ: |
| ret = prepare_long_read_test_requests(td); |
| break; |
| default: |
| test_pr_info("%s: Invalid test case...", __func__); |
| ret = -EINVAL; |
| } |
| |
| return ret; |
| } |
| |
| static int run_packed_test(struct test_data *td) |
| { |
| struct mmc_queue *mq; |
| struct request_queue *req_q; |
| |
| if (!td) { |
| pr_err("%s: NULL td", __func__); |
| return -EINVAL; |
| } |
| |
| req_q = td->req_q; |
| |
| if (!req_q) { |
| pr_err("%s: NULL request queue", __func__); |
| return -EINVAL; |
| } |
| |
| mq = req_q->queuedata; |
| if (!mq) { |
| test_pr_err("%s: NULL mq", __func__); |
| return -EINVAL; |
| } |
| mmc_blk_init_packed_statistics(mq->card); |
| |
| if (td->test_info.testcase != TEST_PACK_MIX_PACKED_NO_PACKED_PACKED) { |
| /* |
| * Verify that the packing is disabled before starting the |
| * test |
| */ |
| mq->wr_packing_enabled = false; |
| mq->num_of_potential_packed_wr_reqs = 0; |
| } |
| |
| __blk_run_queue(td->req_q); |
| |
| return 0; |
| } |
| |
| /* |
| * An implementation for the post_test_fn in the test_info data structure. |
| * In our case we just reset the function pointers in the mmc_queue in order for |
| * the FS to be able to dispatch it's requests correctly after the test is |
| * finished. |
| */ |
| static int post_test(struct test_data *td) |
| { |
| struct mmc_queue *mq; |
| |
| if (!td) |
| return -EINVAL; |
| |
| mq = td->req_q->queuedata; |
| |
| if (!mq) { |
| test_pr_err("%s: NULL mq", __func__); |
| return -EINVAL; |
| } |
| |
| mq->packed_test_fn = NULL; |
| mq->err_check_fn = NULL; |
| |
| return 0; |
| } |
| |
| /* |
| * This function checks, based on the current test's test_group, that the |
| * packed commands capability and control are set right. In addition, we check |
| * if the card supports the packed command feature. |
| */ |
| static int validate_packed_commands_settings(void) |
| { |
| struct request_queue *req_q; |
| struct mmc_queue *mq; |
| int max_num_requests; |
| struct mmc_host *host; |
| |
| req_q = test_iosched_get_req_queue(); |
| if (!req_q) { |
| test_pr_err("%s: test_iosched_get_req_queue failed", __func__); |
| test_iosched_set_test_result(TEST_FAILED); |
| return -EINVAL; |
| } |
| |
| mq = req_q->queuedata; |
| if (!mq) { |
| test_pr_err("%s: NULL mq", __func__); |
| return -EINVAL; |
| } |
| |
| max_num_requests = mq->card->ext_csd.max_packed_writes; |
| host = mq->card->host; |
| |
| if (!(host->caps2 && MMC_CAP2_PACKED_WR)) { |
| test_pr_err("%s: Packed Write capability disabled, exit test", |
| __func__); |
| test_iosched_set_test_result(TEST_NOT_SUPPORTED); |
| return -EINVAL; |
| } |
| |
| if (max_num_requests == 0) { |
| test_pr_err( |
| "%s: no write packing support, ext_csd.max_packed_writes=%d", |
| __func__, mq->card->ext_csd.max_packed_writes); |
| test_iosched_set_test_result(TEST_NOT_SUPPORTED); |
| return -EINVAL; |
| } |
| |
| test_pr_info("%s: max number of packed requests supported is %d ", |
| __func__, max_num_requests); |
| |
| switch (mbtd->test_group) { |
| case TEST_SEND_WRITE_PACKING_GROUP: |
| case TEST_ERR_CHECK_GROUP: |
| case TEST_SEND_INVALID_GROUP: |
| /* disable the packing control */ |
| host->caps2 &= ~MMC_CAP2_PACKED_WR_CONTROL; |
| break; |
| case TEST_PACKING_CONTROL_GROUP: |
| host->caps2 |= MMC_CAP2_PACKED_WR_CONTROL; |
| break; |
| default: |
| break; |
| } |
| |
| return 0; |
| } |
| |
| static void pseudo_rnd_sector_and_size(unsigned int *seed, |
| unsigned int min_start_sector, |
| unsigned int *start_sector, |
| unsigned int *num_of_bios) |
| { |
| unsigned int max_sec = min_start_sector + TEST_MAX_SECTOR_RANGE; |
| do { |
| *start_sector = pseudo_random_seed(seed, |
| 1, max_sec); |
| *num_of_bios = pseudo_random_seed(seed, |
| 1, TEST_MAX_BIOS_PER_REQ); |
| if (!(*num_of_bios)) |
| *num_of_bios = 1; |
| } while ((*start_sector < min_start_sector) || |
| (*start_sector + (*num_of_bios * BIO_U32_SIZE * 4)) > max_sec); |
| } |
| |
| /* sanitize test functions */ |
| static int prepare_write_discard_sanitize_read(struct test_data *td) |
| { |
| unsigned int start_sector; |
| unsigned int num_of_bios = 0; |
| static unsigned int total_bios; |
| unsigned int *num_bios_seed; |
| int i = 0; |
| |
| if (mbtd->random_test_seed == 0) { |
| mbtd->random_test_seed = |
| (unsigned int)(get_jiffies_64() & 0xFFFF); |
| test_pr_info("%s: got seed from jiffies %d", |
| __func__, mbtd->random_test_seed); |
| } |
| num_bios_seed = &mbtd->random_test_seed; |
| |
| do { |
| pseudo_rnd_sector_and_size(num_bios_seed, td->start_sector, |
| &start_sector, &num_of_bios); |
| |
| /* DISCARD */ |
| total_bios += num_of_bios; |
| test_pr_info("%s: discard req: id=%d, startSec=%d, NumBios=%d", |
| __func__, td->unique_next_req_id, start_sector, |
| num_of_bios); |
| test_iosched_add_unique_test_req(0, REQ_UNIQUE_DISCARD, |
| start_sector, BIO_TO_SECTOR(num_of_bios), |
| NULL); |
| |
| } while (++i < (BLKDEV_MAX_RQ-10)); |
| |
| test_pr_info("%s: total discard bios = %d", __func__, total_bios); |
| |
| test_pr_info("%s: add sanitize req", __func__); |
| test_iosched_add_unique_test_req(0, REQ_UNIQUE_SANITIZE, 0, 0, NULL); |
| |
| return 0; |
| } |
| |
| /* |
| * Post test operations for BKOPs test |
| * Disable the BKOPs statistics and clear the feature flags |
| */ |
| static int bkops_post_test(struct test_data *td) |
| { |
| struct request_queue *q = td->req_q; |
| struct mmc_queue *mq = (struct mmc_queue *)q->queuedata; |
| struct mmc_card *card = mq->card; |
| |
| mmc_card_clr_doing_bkops(mq->card); |
| card->ext_csd.raw_bkops_status = 0; |
| |
| spin_lock(&card->bkops_info.bkops_stats.lock); |
| card->bkops_info.bkops_stats.enabled = false; |
| spin_unlock(&card->bkops_info.bkops_stats.lock); |
| |
| return 0; |
| } |
| |
| /* |
| * Verify the BKOPs statsistics |
| */ |
| static int check_bkops_result(struct test_data *td) |
| { |
| struct request_queue *q = td->req_q; |
| struct mmc_queue *mq = (struct mmc_queue *)q->queuedata; |
| struct mmc_card *card = mq->card; |
| struct mmc_bkops_stats *bkops_stat; |
| |
| if (!card) |
| goto fail; |
| |
| bkops_stat = &card->bkops_info.bkops_stats; |
| |
| test_pr_info("%s: Test results: bkops:(%d,%d,%d) hpi:%d, suspend:%d", |
| __func__, |
| bkops_stat->bkops_level[BKOPS_SEVERITY_1_INDEX], |
| bkops_stat->bkops_level[BKOPS_SEVERITY_2_INDEX], |
| bkops_stat->bkops_level[BKOPS_SEVERITY_3_INDEX], |
| bkops_stat->hpi, |
| bkops_stat->suspend); |
| |
| switch (mbtd->test_info.testcase) { |
| case BKOPS_DELAYED_WORK_LEVEL_1: |
| if ((bkops_stat->bkops_level[BKOPS_SEVERITY_1_INDEX] == 1) && |
| (bkops_stat->suspend == 1) && |
| (bkops_stat->hpi == 0)) |
| goto exit; |
| else |
| goto fail; |
| break; |
| case BKOPS_DELAYED_WORK_LEVEL_1_HPI: |
| if ((bkops_stat->bkops_level[BKOPS_SEVERITY_1_INDEX] == 1) && |
| (bkops_stat->suspend == 0) && |
| (bkops_stat->hpi == 1)) |
| goto exit; |
| /* this might happen due to timing issues */ |
| else if |
| ((bkops_stat->bkops_level[BKOPS_SEVERITY_1_INDEX] == 0) && |
| (bkops_stat->suspend == 0) && |
| (bkops_stat->hpi == 0)) |
| goto ignore; |
| else |
| goto fail; |
| break; |
| case BKOPS_CANCEL_DELAYED_WORK: |
| if ((bkops_stat->bkops_level[BKOPS_SEVERITY_1_INDEX] == 0) && |
| (bkops_stat->bkops_level[BKOPS_SEVERITY_2_INDEX] == 0) && |
| (bkops_stat->bkops_level[BKOPS_SEVERITY_3_INDEX] == 0) && |
| (bkops_stat->suspend == 0) && |
| (bkops_stat->hpi == 0)) |
| goto exit; |
| else |
| goto fail; |
| case BKOPS_URGENT_LEVEL_2: |
| case BKOPS_URGENT_LEVEL_2_TWO_REQS: |
| if ((bkops_stat->bkops_level[BKOPS_SEVERITY_2_INDEX] == 1) && |
| (bkops_stat->suspend == 0) && |
| (bkops_stat->hpi == 0)) |
| goto exit; |
| else |
| goto fail; |
| case BKOPS_URGENT_LEVEL_3: |
| if ((bkops_stat->bkops_level[BKOPS_SEVERITY_3_INDEX] == 1) && |
| (bkops_stat->suspend == 0) && |
| (bkops_stat->hpi == 0)) |
| goto exit; |
| else |
| goto fail; |
| default: |
| return -EINVAL; |
| } |
| |
| exit: |
| return 0; |
| ignore: |
| test_iosched_set_ignore_round(true); |
| return 0; |
| fail: |
| if (td->fs_wr_reqs_during_test) { |
| test_pr_info("%s: wr reqs during test, cancel the round", |
| __func__); |
| test_iosched_set_ignore_round(true); |
| return 0; |
| } |
| |
| test_pr_info("%s: BKOPs statistics are not as expected, test failed", |
| __func__); |
| return -EINVAL; |
| } |
| |
| static void bkops_end_io_final_fn(struct request *rq, int err) |
| { |
| struct test_request *test_rq = |
| (struct test_request *)rq->elv.priv[0]; |
| BUG_ON(!test_rq); |
| |
| test_rq->req_completed = 1; |
| test_rq->req_result = err; |
| |
| test_pr_info("%s: request %d completed, err=%d", |
| __func__, test_rq->req_id, err); |
| |
| mbtd->bkops_stage = BKOPS_STAGE_4; |
| wake_up(&mbtd->bkops_wait_q); |
| } |
| |
| static void bkops_end_io_fn(struct request *rq, int err) |
| { |
| struct test_request *test_rq = |
| (struct test_request *)rq->elv.priv[0]; |
| BUG_ON(!test_rq); |
| |
| test_rq->req_completed = 1; |
| test_rq->req_result = err; |
| |
| test_pr_info("%s: request %d completed, err=%d", |
| __func__, test_rq->req_id, err); |
| mbtd->bkops_stage = BKOPS_STAGE_2; |
| wake_up(&mbtd->bkops_wait_q); |
| |
| } |
| |
| static int prepare_bkops(struct test_data *td) |
| { |
| int ret = 0; |
| struct request_queue *q = td->req_q; |
| struct mmc_queue *mq = (struct mmc_queue *)q->queuedata; |
| struct mmc_card *card = mq->card; |
| struct mmc_bkops_stats *bkops_stat; |
| |
| if (!card) |
| return -EINVAL; |
| |
| bkops_stat = &card->bkops_info.bkops_stats; |
| |
| if (!card->ext_csd.bkops_en) { |
| test_pr_err("%s: BKOPS is not enabled by card or host)", |
| __func__); |
| return -ENOTSUPP; |
| } |
| if (mmc_card_doing_bkops(card)) { |
| test_pr_err("%s: BKOPS in progress, try later", __func__); |
| return -EAGAIN; |
| } |
| |
| mmc_blk_init_bkops_statistics(card); |
| |
| if ((mbtd->test_info.testcase == BKOPS_URGENT_LEVEL_2) || |
| (mbtd->test_info.testcase == BKOPS_URGENT_LEVEL_2_TWO_REQS) || |
| (mbtd->test_info.testcase == BKOPS_URGENT_LEVEL_3)) |
| mq->err_check_fn = test_err_check; |
| mbtd->err_check_counter = 0; |
| |
| return ret; |
| } |
| |
| static int run_bkops(struct test_data *td) |
| { |
| int ret = 0; |
| struct request_queue *q = td->req_q; |
| struct mmc_queue *mq = (struct mmc_queue *)q->queuedata; |
| struct mmc_card *card = mq->card; |
| struct mmc_bkops_stats *bkops_stat; |
| |
| if (!card) |
| return -EINVAL; |
| |
| bkops_stat = &card->bkops_info.bkops_stats; |
| |
| switch (mbtd->test_info.testcase) { |
| case BKOPS_DELAYED_WORK_LEVEL_1: |
| bkops_stat->ignore_card_bkops_status = true; |
| card->ext_csd.raw_bkops_status = 1; |
| card->bkops_info.sectors_changed = |
| card->bkops_info.min_sectors_to_queue_delayed_work + 1; |
| mbtd->bkops_stage = BKOPS_STAGE_1; |
| |
| __blk_run_queue(q); |
| /* this long sleep makes sure the host starts bkops and |
| also, gets into suspend */ |
| msleep(10000); |
| |
| bkops_stat->ignore_card_bkops_status = false; |
| card->ext_csd.raw_bkops_status = 0; |
| |
| test_iosched_mark_test_completion(); |
| break; |
| |
| case BKOPS_DELAYED_WORK_LEVEL_1_HPI: |
| bkops_stat->ignore_card_bkops_status = true; |
| card->ext_csd.raw_bkops_status = 1; |
| card->bkops_info.sectors_changed = |
| card->bkops_info.min_sectors_to_queue_delayed_work + 1; |
| mbtd->bkops_stage = BKOPS_STAGE_1; |
| |
| __blk_run_queue(q); |
| msleep(card->bkops_info.delay_ms); |
| |
| ret = test_iosched_add_wr_rd_test_req(0, WRITE, |
| td->start_sector, |
| TEST_REQUEST_NUM_OF_BIOS, |
| TEST_PATTERN_5A, |
| bkops_end_io_final_fn); |
| if (ret) { |
| test_pr_err("%s: failed to add a write request", |
| __func__); |
| ret = -EINVAL; |
| break; |
| } |
| |
| __blk_run_queue(q); |
| wait_event(mbtd->bkops_wait_q, |
| mbtd->bkops_stage == BKOPS_STAGE_4); |
| bkops_stat->ignore_card_bkops_status = false; |
| |
| test_iosched_mark_test_completion(); |
| break; |
| |
| case BKOPS_CANCEL_DELAYED_WORK: |
| bkops_stat->ignore_card_bkops_status = true; |
| card->ext_csd.raw_bkops_status = 1; |
| card->bkops_info.sectors_changed = |
| card->bkops_info.min_sectors_to_queue_delayed_work + 1; |
| mbtd->bkops_stage = BKOPS_STAGE_1; |
| |
| __blk_run_queue(q); |
| |
| ret = test_iosched_add_wr_rd_test_req(0, WRITE, |
| td->start_sector, |
| TEST_REQUEST_NUM_OF_BIOS, |
| TEST_PATTERN_5A, |
| bkops_end_io_final_fn); |
| if (ret) { |
| test_pr_err("%s: failed to add a write request", |
| __func__); |
| ret = -EINVAL; |
| break; |
| } |
| |
| __blk_run_queue(q); |
| wait_event(mbtd->bkops_wait_q, |
| mbtd->bkops_stage == BKOPS_STAGE_4); |
| bkops_stat->ignore_card_bkops_status = false; |
| |
| test_iosched_mark_test_completion(); |
| break; |
| |
| case BKOPS_URGENT_LEVEL_2: |
| case BKOPS_URGENT_LEVEL_3: |
| bkops_stat->ignore_card_bkops_status = true; |
| if (mbtd->test_info.testcase == BKOPS_URGENT_LEVEL_2) |
| card->ext_csd.raw_bkops_status = 2; |
| else |
| card->ext_csd.raw_bkops_status = 3; |
| mbtd->bkops_stage = BKOPS_STAGE_1; |
| |
| ret = test_iosched_add_wr_rd_test_req(0, WRITE, |
| td->start_sector, |
| TEST_REQUEST_NUM_OF_BIOS, |
| TEST_PATTERN_5A, |
| bkops_end_io_fn); |
| if (ret) { |
| test_pr_err("%s: failed to add a write request", |
| __func__); |
| ret = -EINVAL; |
| break; |
| } |
| |
| __blk_run_queue(q); |
| wait_event(mbtd->bkops_wait_q, |
| mbtd->bkops_stage == BKOPS_STAGE_2); |
| card->ext_csd.raw_bkops_status = 0; |
| |
| ret = test_iosched_add_wr_rd_test_req(0, WRITE, |
| td->start_sector, |
| TEST_REQUEST_NUM_OF_BIOS, |
| TEST_PATTERN_5A, |
| bkops_end_io_final_fn); |
| if (ret) { |
| test_pr_err("%s: failed to add a write request", |
| __func__); |
| ret = -EINVAL; |
| break; |
| } |
| |
| __blk_run_queue(q); |
| |
| wait_event(mbtd->bkops_wait_q, |
| mbtd->bkops_stage == BKOPS_STAGE_4); |
| |
| bkops_stat->ignore_card_bkops_status = false; |
| test_iosched_mark_test_completion(); |
| break; |
| |
| case BKOPS_URGENT_LEVEL_2_TWO_REQS: |
| mq->wr_packing_enabled = false; |
| bkops_stat->ignore_card_bkops_status = true; |
| card->ext_csd.raw_bkops_status = 2; |
| mbtd->bkops_stage = BKOPS_STAGE_1; |
| |
| ret = test_iosched_add_wr_rd_test_req(0, WRITE, |
| td->start_sector, |
| TEST_REQUEST_NUM_OF_BIOS, |
| TEST_PATTERN_5A, |
| NULL); |
| if (ret) { |
| test_pr_err("%s: failed to add a write request", |
| __func__); |
| ret = -EINVAL; |
| break; |
| } |
| |
| ret = test_iosched_add_wr_rd_test_req(0, WRITE, |
| td->start_sector, |
| TEST_REQUEST_NUM_OF_BIOS, |
| TEST_PATTERN_5A, |
| bkops_end_io_fn); |
| if (ret) { |
| test_pr_err("%s: failed to add a write request", |
| __func__); |
| ret = -EINVAL; |
| break; |
| } |
| |
| __blk_run_queue(q); |
| wait_event(mbtd->bkops_wait_q, |
| mbtd->bkops_stage == BKOPS_STAGE_2); |
| card->ext_csd.raw_bkops_status = 0; |
| |
| ret = test_iosched_add_wr_rd_test_req(0, WRITE, |
| td->start_sector, |
| TEST_REQUEST_NUM_OF_BIOS, |
| TEST_PATTERN_5A, |
| bkops_end_io_final_fn); |
| if (ret) { |
| test_pr_err("%s: failed to add a write request", |
| __func__); |
| ret = -EINVAL; |
| break; |
| } |
| |
| __blk_run_queue(q); |
| |
| wait_event(mbtd->bkops_wait_q, |
| mbtd->bkops_stage == BKOPS_STAGE_4); |
| |
| bkops_stat->ignore_card_bkops_status = false; |
| test_iosched_mark_test_completion(); |
| |
| break; |
| default: |
| test_pr_err("%s: wrong testcase: %d", __func__, |
| mbtd->test_info.testcase); |
| ret = -EINVAL; |
| } |
| return ret; |
| } |
| |
| /* |
| * new_req_post_test() - Do post test operations for |
| * new_req_notification test: disable the statistics and clear |
| * the feature flags. |
| * @td The test_data for the new_req test that has |
| * ended. |
| */ |
| static int new_req_post_test(struct test_data *td) |
| { |
| struct mmc_queue *mq; |
| |
| if (!td || !td->req_q) |
| goto exit; |
| |
| mq = (struct mmc_queue *)td->req_q->queuedata; |
| |
| if (!mq || !mq->card) |
| goto exit; |
| |
| test_pr_info("Completed %d requests", |
| mbtd->completed_req_count); |
| |
| exit: |
| return 0; |
| } |
| |
| /* |
| * check_new_req_result() - Print out the number of completed |
| * requests. Assigned to the check_test_result_fn pointer, |
| * therefore the name. |
| * @td The test_data for the new_req test that has |
| * ended. |
| */ |
| static int check_new_req_result(struct test_data *td) |
| { |
| test_pr_info("%s: Test results: Completed %d requests", |
| __func__, mbtd->completed_req_count); |
| return 0; |
| } |
| |
| /* |
| * new_req_free_end_io_fn() - Remove request from queuelist and |
| * free request's allocated memory. Used as a call-back |
| * assigned to end_io member in request struct. |
| * @rq The request to be freed |
| * @err Unused |
| */ |
| static void new_req_free_end_io_fn(struct request *rq, int err) |
| { |
| struct test_request *test_rq = |
| (struct test_request *)rq->elv.priv[0]; |
| struct test_data *ptd = test_get_test_data(); |
| |
| BUG_ON(!test_rq); |
| |
| spin_lock_irq(&ptd->lock); |
| list_del_init(&test_rq->queuelist); |
| ptd->dispatched_count--; |
| spin_unlock_irq(&ptd->lock); |
| |
| __blk_put_request(ptd->req_q, test_rq->rq); |
| kfree(test_rq->bios_buffer); |
| kfree(test_rq); |
| mbtd->completed_req_count++; |
| } |
| |
| static int prepare_new_req(struct test_data *td) |
| { |
| struct request_queue *q = td->req_q; |
| struct mmc_queue *mq = (struct mmc_queue *)q->queuedata; |
| |
| mmc_blk_init_packed_statistics(mq->card); |
| mbtd->completed_req_count = 0; |
| |
| return 0; |
| } |
| |
| static int run_new_req(struct test_data *ptd) |
| { |
| int ret = 0; |
| int i; |
| unsigned int requests_count = 2; |
| unsigned int bio_num; |
| struct test_request *test_rq = NULL; |
| |
| while (1) { |
| for (i = 0; i < requests_count; i++) { |
| bio_num = TEST_MAX_BIOS_PER_REQ; |
| test_rq = test_iosched_create_test_req(0, READ, |
| ptd->start_sector, |
| bio_num, TEST_PATTERN_5A, |
| new_req_free_end_io_fn); |
| if (test_rq) { |
| spin_lock_irq(ptd->req_q->queue_lock); |
| list_add_tail(&test_rq->queuelist, |
| &ptd->test_queue); |
| ptd->test_count++; |
| spin_unlock_irq(ptd->req_q->queue_lock); |
| } else { |
| test_pr_err("%s: failed to create read request", |
| __func__); |
| ret = -ENODEV; |
| break; |
| } |
| } |
| |
| __blk_run_queue(ptd->req_q); |
| /* wait while a mmc layer will send all requests in test_queue*/ |
| while (!list_empty(&ptd->test_queue)) |
| msleep(NEW_REQ_TEST_SLEEP_TIME); |
| |
| /* test finish criteria */ |
| if (mbtd->completed_req_count > 1000) { |
| if (ptd->dispatched_count) |
| continue; |
| else |
| break; |
| } |
| |
| for (i = 0; i < requests_count; i++) { |
| bio_num = NEW_REQ_TEST_NUM_BIOS; |
| test_rq = test_iosched_create_test_req(0, READ, |
| ptd->start_sector, |
| bio_num, TEST_PATTERN_5A, |
| new_req_free_end_io_fn); |
| if (test_rq) { |
| spin_lock_irq(ptd->req_q->queue_lock); |
| list_add_tail(&test_rq->queuelist, |
| &ptd->test_queue); |
| ptd->test_count++; |
| spin_unlock_irq(ptd->req_q->queue_lock); |
| } else { |
| test_pr_err("%s: failed to create read request", |
| __func__); |
| ret = -ENODEV; |
| break; |
| } |
| } |
| __blk_run_queue(ptd->req_q); |
| } |
| |
| test_iosched_mark_test_completion(); |
| test_pr_info("%s: EXIT: %d code", __func__, ret); |
| |
| return ret; |
| } |
| |
| static bool message_repeat; |
| static int test_open(struct inode *inode, struct file *file) |
| { |
| file->private_data = inode->i_private; |
| message_repeat = 1; |
| return 0; |
| } |
| |
| /* send_packing TEST */ |
| static ssize_t send_write_packing_test_write(struct file *file, |
| const char __user *buf, |
| size_t count, |
| loff_t *ppos) |
| { |
| int ret = 0; |
| int i = 0; |
| int number = -1; |
| int j = 0; |
| |
| test_pr_info("%s: -- send_write_packing TEST --", __func__); |
| |
| sscanf(buf, "%d", &number); |
| |
| if (number <= 0) |
| number = 1; |
| |
| |
| mbtd->test_group = TEST_SEND_WRITE_PACKING_GROUP; |
| |
| if (validate_packed_commands_settings()) |
| return count; |
| |
| if (mbtd->random_test_seed > 0) |
| test_pr_info("%s: Test seed: %d", __func__, |
| mbtd->random_test_seed); |
| |
| memset(&mbtd->test_info, 0, sizeof(struct test_info)); |
| |
| mbtd->test_info.data = mbtd; |
| mbtd->test_info.prepare_test_fn = prepare_test; |
| mbtd->test_info.run_test_fn = run_packed_test; |
| mbtd->test_info.check_test_result_fn = check_wr_packing_statistics; |
| mbtd->test_info.get_test_case_str_fn = get_test_case_str; |
| mbtd->test_info.post_test_fn = post_test; |
| |
| for (i = 0; i < number; ++i) { |
| test_pr_info("%s: Cycle # %d / %d", __func__, i+1, number); |
| test_pr_info("%s: ====================", __func__); |
| |
| for (j = SEND_WRITE_PACKING_MIN_TESTCASE; |
| j <= SEND_WRITE_PACKING_MAX_TESTCASE; j++) { |
| |
| mbtd->test_info.testcase = j; |
| mbtd->is_random = RANDOM_TEST; |
| ret = test_iosched_start_test(&mbtd->test_info); |
| if (ret) |
| break; |
| /* Allow FS requests to be dispatched */ |
| msleep(1000); |
| mbtd->test_info.testcase = j; |
| mbtd->is_random = NON_RANDOM_TEST; |
| ret = test_iosched_start_test(&mbtd->test_info); |
| if (ret) |
| break; |
| /* Allow FS requests to be dispatched */ |
| msleep(1000); |
| } |
| } |
| |
| test_pr_info("%s: Completed all the test cases.", __func__); |
| |
| return count; |
| } |
| |
| static ssize_t send_write_packing_test_read(struct file *file, |
| char __user *buffer, |
| size_t count, |
| loff_t *offset) |
| { |
| memset((void *)buffer, 0, count); |
| |
| snprintf(buffer, count, |
| "\nsend_write_packing_test\n" |
| "=========\n" |
| "Description:\n" |
| "This test checks the following scenarios\n" |
| "- Pack due to FLUSH message\n" |
| "- Pack due to FLUSH after threshold writes\n" |
| "- Pack due to READ message\n" |
| "- Pack due to READ after threshold writes\n" |
| "- Pack due to empty queue\n" |
| "- Pack due to threshold writes\n" |
| "- Pack due to one over threshold writes\n"); |
| |
| if (message_repeat == 1) { |
| message_repeat = 0; |
| return strnlen(buffer, count); |
| } else { |
| return 0; |
| } |
| } |
| |
| const struct file_operations send_write_packing_test_ops = { |
| .open = test_open, |
| .write = send_write_packing_test_write, |
| .read = send_write_packing_test_read, |
| }; |
| |
| /* err_check TEST */ |
| static ssize_t err_check_test_write(struct file *file, |
| const char __user *buf, |
| size_t count, |
| loff_t *ppos) |
| { |
| int ret = 0; |
| int i = 0; |
| int number = -1; |
| int j = 0; |
| |
| test_pr_info("%s: -- err_check TEST --", __func__); |
| |
| sscanf(buf, "%d", &number); |
| |
| if (number <= 0) |
| number = 1; |
| |
| mbtd->test_group = TEST_ERR_CHECK_GROUP; |
| |
| if (validate_packed_commands_settings()) |
| return count; |
| |
| if (mbtd->random_test_seed > 0) |
| test_pr_info("%s: Test seed: %d", __func__, |
| mbtd->random_test_seed); |
| |
| memset(&mbtd->test_info, 0, sizeof(struct test_info)); |
| |
| mbtd->test_info.data = mbtd; |
| mbtd->test_info.prepare_test_fn = prepare_test; |
| mbtd->test_info.run_test_fn = run_packed_test; |
| mbtd->test_info.check_test_result_fn = check_wr_packing_statistics; |
| mbtd->test_info.get_test_case_str_fn = get_test_case_str; |
| mbtd->test_info.post_test_fn = post_test; |
| |
| for (i = 0; i < number; ++i) { |
| test_pr_info("%s: Cycle # %d / %d", __func__, i+1, number); |
| test_pr_info("%s: ====================", __func__); |
| |
| for (j = ERR_CHECK_MIN_TESTCASE; |
| j <= ERR_CHECK_MAX_TESTCASE ; j++) { |
| mbtd->test_info.testcase = j; |
| mbtd->is_random = RANDOM_TEST; |
| ret = test_iosched_start_test(&mbtd->test_info); |
| if (ret) |
| break; |
| /* Allow FS requests to be dispatched */ |
| msleep(1000); |
| mbtd->test_info.testcase = j; |
| mbtd->is_random = NON_RANDOM_TEST; |
| ret = test_iosched_start_test(&mbtd->test_info); |
| if (ret) |
| break; |
| /* Allow FS requests to be dispatched */ |
| msleep(1000); |
| } |
| } |
| |
| test_pr_info("%s: Completed all the test cases.", __func__); |
| |
| return count; |
| } |
| |
| static ssize_t err_check_test_read(struct file *file, |
| char __user *buffer, |
| size_t count, |
| loff_t *offset) |
| { |
| memset((void *)buffer, 0, count); |
| |
| snprintf(buffer, count, |
| "\nerr_check_TEST\n" |
| "=========\n" |
| "Description:\n" |
| "This test checks the following scenarios\n" |
| "- Return ABORT\n" |
| "- Return PARTIAL followed by success\n" |
| "- Return PARTIAL followed by abort\n" |
| "- Return PARTIAL multiple times until success\n" |
| "- Return PARTIAL with fail index = threshold\n" |
| "- Return RETRY\n" |
| "- Return CMD_ERR\n" |
| "- Return DATA_ERR\n"); |
| |
| if (message_repeat == 1) { |
| message_repeat = 0; |
| return strnlen(buffer, count); |
| } else { |
| return 0; |
| } |
| } |
| |
| const struct file_operations err_check_test_ops = { |
| .open = test_open, |
| .write = err_check_test_write, |
| .read = err_check_test_read, |
| }; |
| |
| /* send_invalid_packed TEST */ |
| static ssize_t send_invalid_packed_test_write(struct file *file, |
| const char __user *buf, |
| size_t count, |
| loff_t *ppos) |
| { |
| int ret = 0; |
| int i = 0; |
| int number = -1; |
| int j = 0; |
| int num_of_failures = 0; |
| |
| test_pr_info("%s: -- send_invalid_packed TEST --", __func__); |
| |
| sscanf(buf, "%d", &number); |
| |
| if (number <= 0) |
| number = 1; |
| |
| mbtd->test_group = TEST_SEND_INVALID_GROUP; |
| |
| if (validate_packed_commands_settings()) |
| return count; |
| |
| if (mbtd->random_test_seed > 0) |
| test_pr_info("%s: Test seed: %d", __func__, |
| mbtd->random_test_seed); |
| |
| memset(&mbtd->test_info, 0, sizeof(struct test_info)); |
| |
| mbtd->test_info.data = mbtd; |
| mbtd->test_info.prepare_test_fn = prepare_test; |
| mbtd->test_info.run_test_fn = run_packed_test; |
| mbtd->test_info.check_test_result_fn = check_wr_packing_statistics; |
| mbtd->test_info.get_test_case_str_fn = get_test_case_str; |
| mbtd->test_info.post_test_fn = post_test; |
| |
| for (i = 0; i < number; ++i) { |
| test_pr_info("%s: Cycle # %d / %d", __func__, i+1, number); |
| test_pr_info("%s: ====================", __func__); |
| |
| for (j = INVALID_CMD_MIN_TESTCASE; |
| j <= INVALID_CMD_MAX_TESTCASE ; j++) { |
| |
| mbtd->test_info.testcase = j; |
| mbtd->is_random = RANDOM_TEST; |
| ret = test_iosched_start_test(&mbtd->test_info); |
| if (ret) |
| num_of_failures++; |
| /* Allow FS requests to be dispatched */ |
| msleep(1000); |
| |
| mbtd->test_info.testcase = j; |
| mbtd->is_random = NON_RANDOM_TEST; |
| ret = test_iosched_start_test(&mbtd->test_info); |
| if (ret) |
| num_of_failures++; |
| /* Allow FS requests to be dispatched */ |
| msleep(1000); |
| } |
| } |
| |
| test_pr_info("%s: Completed all the test cases.", __func__); |
| |
| if (num_of_failures > 0) { |
| test_iosched_set_test_result(TEST_FAILED); |
| test_pr_err( |
| "There were %d failures during the test, TEST FAILED", |
| num_of_failures); |
| } |
| return count; |
| } |
| |
| static ssize_t send_invalid_packed_test_read(struct file *file, |
| char __user *buffer, |
| size_t count, |
| loff_t *offset) |
| { |
| memset((void *)buffer, 0, count); |
| |
| snprintf(buffer, count, |
| "\nsend_invalid_packed_TEST\n" |
| "=========\n" |
| "Description:\n" |
| "This test checks the following scenarios\n" |
| "- Send an invalid header version\n" |
| "- Send the wrong write code\n" |
| "- Send an invalid R/W code\n" |
| "- Send wrong start address in header\n" |
| "- Send header with block_count smaller than actual\n" |
| "- Send header with block_count larger than actual\n" |
| "- Send header CMD23 packed bit set\n" |
| "- Send CMD23 with block count over threshold\n" |
| "- Send CMD23 with block_count equals zero\n" |
| "- Send CMD23 packed bit unset\n" |
| "- Send CMD23 reliable write bit set\n" |
| "- Send CMD23 bits [16-29] set\n" |
| "- Send CMD23 header block not in block_count\n"); |
| |
| if (message_repeat == 1) { |
| message_repeat = 0; |
| return strnlen(buffer, count); |
| } else { |
| return 0; |
| } |
| } |
| |
| const struct file_operations send_invalid_packed_test_ops = { |
| .open = test_open, |
| .write = send_invalid_packed_test_write, |
| .read = send_invalid_packed_test_read, |
| }; |
| |
| /* packing_control TEST */ |
| static ssize_t write_packing_control_test_write(struct file *file, |
| const char __user *buf, |
| size_t count, |
| loff_t *ppos) |
| { |
| int ret = 0; |
| int i = 0; |
| int number = -1; |
| int j = 0; |
| struct mmc_queue *mq = test_iosched_get_req_queue()->queuedata; |
| int max_num_requests = mq->card->ext_csd.max_packed_writes; |
| int test_successful = 1; |
| |
| test_pr_info("%s: -- write_packing_control TEST --", __func__); |
| |
| sscanf(buf, "%d", &number); |
| |
| if (number <= 0) |
| number = 1; |
| |
| test_pr_info("%s: max_num_requests = %d ", __func__, |
| max_num_requests); |
| |
| memset(&mbtd->test_info, 0, sizeof(struct test_info)); |
| mbtd->test_group = TEST_PACKING_CONTROL_GROUP; |
| |
| if (validate_packed_commands_settings()) |
| return count; |
| |
| mbtd->test_info.data = mbtd; |
| mbtd->test_info.prepare_test_fn = prepare_test; |
| mbtd->test_info.run_test_fn = run_packed_test; |
| mbtd->test_info.check_test_result_fn = check_wr_packing_statistics; |
| mbtd->test_info.get_test_case_str_fn = get_test_case_str; |
| |
| for (i = 0; i < number; ++i) { |
| test_pr_info("%s: Cycle # %d / %d", __func__, i+1, number); |
| test_pr_info("%s: ====================", __func__); |
| |
| for (j = PACKING_CONTROL_MIN_TESTCASE; |
| j <= PACKING_CONTROL_MAX_TESTCASE; j++) { |
| |
| test_successful = 1; |
| mbtd->test_info.testcase = j; |
| mbtd->is_random = RANDOM_TEST; |
| ret = test_iosched_start_test(&mbtd->test_info); |
| if (ret) { |
| test_successful = 0; |
| break; |
| } |
| /* Allow FS requests to be dispatched */ |
| msleep(1000); |
| |
| mbtd->test_info.testcase = j; |
| mbtd->is_random = NON_RANDOM_TEST; |
| ret = test_iosched_start_test(&mbtd->test_info); |
| if (ret) { |
| test_successful = 0; |
| break; |
| } |
| /* Allow FS requests to be dispatched */ |
| msleep(1000); |
| } |
| |
| if (!test_successful) |
| break; |
| } |
| |
| test_pr_info("%s: Completed all the test cases.", __func__); |
| |
| return count; |
| } |
| |
| static ssize_t write_packing_control_test_read(struct file *file, |
| char __user *buffer, |
| size_t count, |
| loff_t *offset) |
| { |
| memset((void *)buffer, 0, count); |
| |
| snprintf(buffer, count, |
| "\nwrite_packing_control_test\n" |
| "=========\n" |
| "Description:\n" |
| "This test checks the following scenarios\n" |
| "- Packing expected - one over trigger\n" |
| "- Packing expected - N over trigger\n" |
| "- Packing expected - N over trigger followed by read\n" |
| "- Packing expected - N over trigger followed by flush\n" |
| "- Packing expected - threshold over trigger FB by flush\n" |
| "- Packing not expected - less than trigger\n" |
| "- Packing not expected - trigger requests\n" |
| "- Packing not expected - trigger, read, trigger\n" |
| "- Mixed state - packing -> no packing -> packing\n" |
| "- Mixed state - no packing -> packing -> no packing\n"); |
| |
| if (message_repeat == 1) { |
| message_repeat = 0; |
| return strnlen(buffer, count); |
| } else { |
| return 0; |
| } |
| } |
| |
| const struct file_operations write_packing_control_test_ops = { |
| .open = test_open, |
| .write = write_packing_control_test_write, |
| .read = write_packing_control_test_read, |
| }; |
| |
| static ssize_t write_discard_sanitize_test_write(struct file *file, |
| const char __user *buf, |
| size_t count, |
| loff_t *ppos) |
| { |
| int ret = 0; |
| int i = 0; |
| int number = -1; |
| |
| sscanf(buf, "%d", &number); |
| if (number <= 0) |
| number = 1; |
| |
| test_pr_info("%s: -- write_discard_sanitize TEST --\n", __func__); |
| |
| memset(&mbtd->test_info, 0, sizeof(struct test_info)); |
| |
| mbtd->test_group = TEST_GENERAL_GROUP; |
| |
| mbtd->test_info.data = mbtd; |
| mbtd->test_info.prepare_test_fn = prepare_write_discard_sanitize_read; |
| mbtd->test_info.get_test_case_str_fn = get_test_case_str; |
| mbtd->test_info.timeout_msec = SANITIZE_TEST_TIMEOUT; |
| |
| for (i = 0 ; i < number ; ++i) { |
| test_pr_info("%s: Cycle # %d / %d\n", __func__, i+1, number); |
| test_pr_info("%s: ===================", __func__); |
| |
| mbtd->test_info.testcase = TEST_WRITE_DISCARD_SANITIZE_READ; |
| ret = test_iosched_start_test(&mbtd->test_info); |
| |
| if (ret) |
| break; |
| } |
| |
| return count; |
| } |
| |
| const struct file_operations write_discard_sanitize_test_ops = { |
| .open = test_open, |
| .write = write_discard_sanitize_test_write, |
| }; |
| |
| static ssize_t bkops_test_write(struct file *file, |
| const char __user *buf, |
| size_t count, |
| loff_t *ppos) |
| { |
| int ret = 0; |
| int i = 0, j; |
| int number = -1; |
| |
| test_pr_info("%s: -- bkops_test TEST --", __func__); |
| |
| sscanf(buf, "%d", &number); |
| |
| if (number <= 0) |
| number = 1; |
| |
| mbtd->test_group = TEST_BKOPS_GROUP; |
| |
| memset(&mbtd->test_info, 0, sizeof(struct test_info)); |
| |
| mbtd->test_info.data = mbtd; |
| mbtd->test_info.prepare_test_fn = prepare_bkops; |
| mbtd->test_info.check_test_result_fn = check_bkops_result; |
| mbtd->test_info.get_test_case_str_fn = get_test_case_str; |
| mbtd->test_info.run_test_fn = run_bkops; |
| mbtd->test_info.timeout_msec = BKOPS_TEST_TIMEOUT; |
| mbtd->test_info.post_test_fn = bkops_post_test; |
| |
| for (i = 0 ; i < number ; ++i) { |
| test_pr_info("%s: Cycle # %d / %d", __func__, i+1, number); |
| test_pr_info("%s: ===================", __func__); |
| for (j = BKOPS_MIN_TESTCASE ; |
| j <= BKOPS_MAX_TESTCASE ; j++) { |
| mbtd->test_info.testcase = j; |
| ret = test_iosched_start_test(&mbtd->test_info); |
| if (ret) |
| break; |
| } |
| } |
| |
| test_pr_info("%s: Completed all the test cases.", __func__); |
| |
| return count; |
| } |
| |
| static ssize_t bkops_test_read(struct file *file, |
| char __user *buffer, |
| size_t count, |
| loff_t *offset) |
| { |
| memset((void *)buffer, 0, count); |
| |
| snprintf(buffer, count, |
| "\nbkops_test\n========================\n" |
| "Description:\n" |
| "This test simulates BKOPS status from card\n" |
| "and verifies that:\n" |
| " - Starting BKOPS delayed work, level 1\n" |
| " - Starting BKOPS delayed work, level 1, with HPI\n" |
| " - Cancel starting BKOPS delayed work, " |
| " when a request is received\n" |
| " - Starting BKOPS urgent, level 2,3\n" |
| " - Starting BKOPS urgent with 2 requests\n"); |
| return strnlen(buffer, count); |
| } |
| |
| const struct file_operations bkops_test_ops = { |
| .open = test_open, |
| .write = bkops_test_write, |
| .read = bkops_test_read, |
| }; |
| |
| static ssize_t long_sequential_read_test_write(struct file *file, |
| const char __user *buf, |
| size_t count, |
| loff_t *ppos) |
| { |
| int ret = 0; |
| int i = 0; |
| int number = -1; |
| unsigned long mtime, integer, fraction; |
| |
| test_pr_info("%s: -- Long Sequential Read TEST --", __func__); |
| |
| sscanf(buf, "%d", &number); |
| |
| if (number <= 0) |
| number = 1; |
| |
| memset(&mbtd->test_info, 0, sizeof(struct test_info)); |
| mbtd->test_group = TEST_GENERAL_GROUP; |
| |
| mbtd->test_info.data = mbtd; |
| mbtd->test_info.prepare_test_fn = prepare_test; |
| mbtd->test_info.get_test_case_str_fn = get_test_case_str; |
| |
| for (i = 0 ; i < number ; ++i) { |
| test_pr_info("%s: Cycle # %d / %d", __func__, i+1, number); |
| test_pr_info("%s: ====================", __func__); |
| |
| mbtd->test_info.testcase = TEST_LONG_SEQUENTIAL_READ; |
| mbtd->is_random = NON_RANDOM_TEST; |
| ret = test_iosched_start_test(&mbtd->test_info); |
| if (ret) |
| break; |
| |
| mtime = jiffies_to_msecs(mbtd->test_info.test_duration); |
| |
| test_pr_info("%s: time is %lu msec, size is %u.%u MiB", |
| __func__, mtime, |
| LONG_TEST_SIZE_INTEGER(LONG_READ_NUM_BYTES), |
| LONG_TEST_SIZE_FRACTION(LONG_READ_NUM_BYTES)); |
| |
| /* we first multiply in order not to lose precision */ |
| mtime *= MB_MSEC_RATIO_APPROXIMATION; |
| /* divide values to get a MiB/sec integer value with one |
| digit of precision. Multiply by 10 for one digit precision |
| */ |
| fraction = integer = (LONG_READ_NUM_BYTES * 10) / mtime; |
| integer /= 10; |
| /* and calculate the MiB value fraction */ |
| fraction -= integer * 10; |
| |
| test_pr_info("%s: Throughput: %lu.%lu MiB/sec\n" |
| , __func__, integer, fraction); |
| |
| /* Allow FS requests to be dispatched */ |
| msleep(1000); |
| } |
| |
| return count; |
| } |
| |
| static ssize_t long_sequential_read_test_read(struct file *file, |
| char __user *buffer, |
| size_t count, |
| loff_t *offset) |
| { |
| memset((void *)buffer, 0, count); |
| |
| snprintf(buffer, count, |
| "\nlong_sequential_read_test\n" |
| "=========\n" |
| "Description:\n" |
| "This test runs the following scenarios\n" |
| "- Long Sequential Read Test: this test measures read " |
| "throughput at the driver level by sequentially reading many " |
| "large requests.\n"); |
| |
| if (message_repeat == 1) { |
| message_repeat = 0; |
| return strnlen(buffer, count); |
| } else |
| return 0; |
| } |
| |
| const struct file_operations long_sequential_read_test_ops = { |
| .open = test_open, |
| .write = long_sequential_read_test_write, |
| .read = long_sequential_read_test_read, |
| }; |
| |
| static void long_seq_write_free_end_io_fn(struct request *rq, int err) |
| { |
| struct test_request *test_rq = |
| (struct test_request *)rq->elv.priv[0]; |
| struct test_data *ptd = test_get_test_data(); |
| |
| BUG_ON(!test_rq); |
| |
| spin_lock_irq(&ptd->lock); |
| list_del_init(&test_rq->queuelist); |
| ptd->dispatched_count--; |
| __blk_put_request(ptd->req_q, test_rq->rq); |
| spin_unlock_irq(&ptd->lock); |
| |
| kfree(test_rq->bios_buffer); |
| kfree(test_rq); |
| mbtd->completed_req_count++; |
| |
| check_test_completion(); |
| } |
| |
| static int run_long_seq_write(struct test_data *td) |
| { |
| int ret = 0; |
| int i; |
| |
| td->test_count = 0; |
| mbtd->completed_req_count = 0; |
| |
| test_pr_info("%s: Adding at least %d write requests, first req_id=%d", |
| __func__, LONG_WRITE_TEST_MIN_NUM_REQS, |
| td->wr_rd_next_req_id); |
| |
| do { |
| for (i = 0; i < TEST_MAX_REQUESTS; i++) { |
| /* |
| * since our requests come from a pool containing 128 |
| * requests, we don't want to exhaust this quantity, |
| * therefore we add up to TEST_MAX_REQUESTS (which |
| * includes a safety margin) and then call the mmc layer |
| * to fetch them |
| */ |
| if (td->test_count > TEST_MAX_REQUESTS) |
| break; |
| |
| ret = test_iosched_add_wr_rd_test_req(0, WRITE, |
| td->start_sector, TEST_MAX_BIOS_PER_REQ, |
| TEST_PATTERN_5A, |
| long_seq_write_free_end_io_fn); |
| if (ret) { |
| test_pr_err("%s: failed to create write request" |
| , __func__); |
| break; |
| } |
| } |
| |
| __blk_run_queue(td->req_q); |
| |
| } while (mbtd->completed_req_count < LONG_WRITE_TEST_MIN_NUM_REQS); |
| |
| test_pr_info("%s: completed %d requests", __func__, |
| mbtd->completed_req_count); |
| |
| return ret; |
| } |
| |
| static ssize_t long_sequential_write_test_write(struct file *file, |
| const char __user *buf, |
| size_t count, |
| loff_t *ppos) |
| { |
| int ret = 0; |
| int i = 0; |
| int number = -1; |
| unsigned long mtime, integer, fraction, byte_count; |
| |
| test_pr_info("%s: -- Long Sequential Write TEST --", __func__); |
| |
| sscanf(buf, "%d", &number); |
| |
| if (number <= 0) |
| number = 1; |
| |
| memset(&mbtd->test_info, 0, sizeof(struct test_info)); |
| mbtd->test_group = TEST_GENERAL_GROUP; |
| |
| mbtd->test_info.data = mbtd; |
| mbtd->test_info.get_test_case_str_fn = get_test_case_str; |
| mbtd->test_info.run_test_fn = run_long_seq_write; |
| |
| for (i = 0 ; i < number ; ++i) { |
| test_pr_info("%s: Cycle # %d / %d", __func__, i+1, number); |
| test_pr_info("%s: ====================", __func__); |
| |
| integer = 0; |
| fraction = 0; |
| mbtd->test_info.test_byte_count = 0; |
| mbtd->test_info.testcase = TEST_LONG_SEQUENTIAL_WRITE; |
| mbtd->is_random = NON_RANDOM_TEST; |
| ret = test_iosched_start_test(&mbtd->test_info); |
| if (ret) |
| break; |
| |
| mtime = jiffies_to_msecs(mbtd->test_info.test_duration); |
| byte_count = mbtd->test_info.test_byte_count; |
| |
| test_pr_info("%s: time is %lu msec, size is %lu.%lu MiB", |
| __func__, mtime, LONG_TEST_SIZE_INTEGER(byte_count), |
| LONG_TEST_SIZE_FRACTION(byte_count)); |
| |
| /* we first multiply in order not to lose precision */ |
| mtime *= MB_MSEC_RATIO_APPROXIMATION; |
| /* divide values to get a MiB/sec integer value with one |
| digit of precision |
| */ |
| fraction = integer = (byte_count * 10) / mtime; |
| integer /= 10; |
| /* and calculate the MiB value fraction */ |
| fraction -= integer * 10; |
| |
| test_pr_info("%s: Throughput: %lu.%lu MiB/sec\n", |
| __func__, integer, fraction); |
| |
| /* Allow FS requests to be dispatched */ |
| msleep(1000); |
| } |
| |
| return count; |
| } |
| |
| static ssize_t long_sequential_write_test_read(struct file *file, |
| char __user *buffer, |
| size_t count, |
| loff_t *offset) |
| { |
| memset((void *)buffer, 0, count); |
| |
| snprintf(buffer, count, |
| "\nlong_sequential_write_test\n" |
| "=========\n" |
| "Description:\n" |
| "This test runs the following scenarios\n" |
| "- Long Sequential Write Test: this test measures write " |
| "throughput at the driver level by sequentially writing many " |
| "large requests\n"); |
| |
| if (message_repeat == 1) { |
| message_repeat = 0; |
| return strnlen(buffer, count); |
| } else |
| return 0; |
| } |
| |
| const struct file_operations long_sequential_write_test_ops = { |
| .open = test_open, |
| .write = long_sequential_write_test_write, |
| .read = long_sequential_write_test_read, |
| }; |
| |
| static ssize_t new_req_notification_test_write(struct file *file, |
| const char __user *buf, |
| size_t count, |
| loff_t *ppos) |
| { |
| int ret = 0; |
| int i = 0; |
| int number = -1; |
| |
| test_pr_info("%s: -- new_req_notification TEST --", __func__); |
| |
| sscanf(buf, "%d", &number); |
| |
| if (number <= 0) |
| number = 1; |
| |
| mbtd->test_group = TEST_NEW_NOTIFICATION_GROUP; |
| |
| memset(&mbtd->test_info, 0, sizeof(struct test_info)); |
| |
| mbtd->test_info.data = mbtd; |
| mbtd->test_info.prepare_test_fn = prepare_new_req; |
| mbtd->test_info.check_test_result_fn = check_new_req_result; |
| mbtd->test_info.get_test_case_str_fn = get_test_case_str; |
| mbtd->test_info.run_test_fn = run_new_req; |
| mbtd->test_info.timeout_msec = 10 * 60 * 1000; /* 1 min */ |
| mbtd->test_info.post_test_fn = new_req_post_test; |
| |
| for (i = 0 ; i < number ; ++i) { |
| test_pr_info("%s: Cycle # %d / %d", __func__, i+1, number); |
| test_pr_info("%s: ===================", __func__); |
| test_pr_info("%s: start test case TEST_NEW_REQ_NOTIFICATION", |
| __func__); |
| mbtd->test_info.testcase = TEST_NEW_REQ_NOTIFICATION; |
| ret = test_iosched_start_test(&mbtd->test_info); |
| if (ret) { |
| test_pr_info("%s: break from new_req tests loop", |
| __func__); |
| break; |
| } |
| } |
| return count; |
| } |
| |
| static ssize_t new_req_notification_test_read(struct file *file, |
| char __user *buffer, |
| size_t count, |
| loff_t *offset) |
| { |
| memset((void *)buffer, 0, count); |
| |
| snprintf(buffer, count, |
| "\nnew_req_notification_test\n========================\n" |
| "Description:\n" |
| "This test checks following scenarious\n" |
| "- new request arrives after a NULL request was sent to the " |
| "mmc_queue,\n" |
| "which is waiting for completion of a former request\n"); |
| |
| return strnlen(buffer, count); |
| } |
| |
| const struct file_operations new_req_notification_test_ops = { |
| .open = test_open, |
| .write = new_req_notification_test_write, |
| .read = new_req_notification_test_read, |
| }; |
| |
| static void mmc_block_test_debugfs_cleanup(void) |
| { |
| debugfs_remove(mbtd->debug.random_test_seed); |
| debugfs_remove(mbtd->debug.send_write_packing_test); |
| debugfs_remove(mbtd->debug.err_check_test); |
| debugfs_remove(mbtd->debug.send_invalid_packed_test); |
| debugfs_remove(mbtd->debug.packing_control_test); |
| debugfs_remove(mbtd->debug.discard_sanitize_test); |
| debugfs_remove(mbtd->debug.bkops_test); |
| debugfs_remove(mbtd->debug.long_sequential_read_test); |
| debugfs_remove(mbtd->debug.long_sequential_write_test); |
| debugfs_remove(mbtd->debug.new_req_notification_test); |
| } |
| |
| static int mmc_block_test_debugfs_init(void) |
| { |
| struct dentry *utils_root, *tests_root; |
| |
| utils_root = test_iosched_get_debugfs_utils_root(); |
| tests_root = test_iosched_get_debugfs_tests_root(); |
| |
| if (!utils_root || !tests_root) |
| return -EINVAL; |
| |
| mbtd->debug.random_test_seed = debugfs_create_u32( |
| "random_test_seed", |
| S_IRUGO | S_IWUGO, |
| utils_root, |
| &mbtd->random_test_seed); |
| |
| if (!mbtd->debug.random_test_seed) |
| goto err_nomem; |
| |
| mbtd->debug.send_write_packing_test = |
| debugfs_create_file("send_write_packing_test", |
| S_IRUGO | S_IWUGO, |
| tests_root, |
| NULL, |
| &send_write_packing_test_ops); |
| |
| if (!mbtd->debug.send_write_packing_test) |
| goto err_nomem; |
| |
| mbtd->debug.err_check_test = |
| debugfs_create_file("err_check_test", |
| S_IRUGO | S_IWUGO, |
| tests_root, |
| NULL, |
| &err_check_test_ops); |
| |
| if (!mbtd->debug.err_check_test) |
| goto err_nomem; |
| |
| mbtd->debug.send_invalid_packed_test = |
| debugfs_create_file("send_invalid_packed_test", |
| S_IRUGO | S_IWUGO, |
| tests_root, |
| NULL, |
| &send_invalid_packed_test_ops); |
| |
| if (!mbtd->debug.send_invalid_packed_test) |
| goto err_nomem; |
| |
| mbtd->debug.packing_control_test = debugfs_create_file( |
| "packing_control_test", |
| S_IRUGO | S_IWUGO, |
| tests_root, |
| NULL, |
| &write_packing_control_test_ops); |
| |
| if (!mbtd->debug.packing_control_test) |
| goto err_nomem; |
| |
| mbtd->debug.discard_sanitize_test = |
| debugfs_create_file("write_discard_sanitize_test", |
| S_IRUGO | S_IWUGO, |
| tests_root, |
| NULL, |
| &write_discard_sanitize_test_ops); |
| if (!mbtd->debug.discard_sanitize_test) { |
| mmc_block_test_debugfs_cleanup(); |
| return -ENOMEM; |
| } |
| |
| mbtd->debug.bkops_test = |
| debugfs_create_file("bkops_test", |
| S_IRUGO | S_IWUGO, |
| tests_root, |
| NULL, |
| &bkops_test_ops); |
| |
| mbtd->debug.new_req_notification_test = |
| debugfs_create_file("new_req_notification_test", |
| S_IRUGO | S_IWUGO, |
| tests_root, |
| NULL, |
| &new_req_notification_test_ops); |
| |
| if (!mbtd->debug.new_req_notification_test) |
| goto err_nomem; |
| |
| if (!mbtd->debug.bkops_test) |
| goto err_nomem; |
| |
| mbtd->debug.long_sequential_read_test = debugfs_create_file( |
| "long_sequential_read_test", |
| S_IRUGO | S_IWUGO, |
| tests_root, |
| NULL, |
| &long_sequential_read_test_ops); |
| |
| if (!mbtd->debug.long_sequential_read_test) |
| goto err_nomem; |
| |
| mbtd->debug.long_sequential_write_test = debugfs_create_file( |
| "long_sequential_write_test", |
| S_IRUGO | S_IWUGO, |
| tests_root, |
| NULL, |
| &long_sequential_write_test_ops); |
| |
| if (!mbtd->debug.long_sequential_write_test) |
| goto err_nomem; |
| |
| return 0; |
| |
| err_nomem: |
| mmc_block_test_debugfs_cleanup(); |
| return -ENOMEM; |
| } |
| |
| static void mmc_block_test_probe(void) |
| { |
| struct request_queue *q = test_iosched_get_req_queue(); |
| struct mmc_queue *mq; |
| int max_packed_reqs; |
| |
| if (!q) { |
| test_pr_err("%s: NULL request queue", __func__); |
| return; |
| } |
| |
| mq = q->queuedata; |
| if (!mq) { |
| test_pr_err("%s: NULL mq", __func__); |
| return; |
| } |
| |
| max_packed_reqs = mq->card->ext_csd.max_packed_writes; |
| mbtd->exp_packed_stats.packing_events = |
| kzalloc((max_packed_reqs + 1) * |
| sizeof(*mbtd->exp_packed_stats.packing_events), |
| GFP_KERNEL); |
| |
| mmc_block_test_debugfs_init(); |
| } |
| |
| static void mmc_block_test_remove(void) |
| { |
| mmc_block_test_debugfs_cleanup(); |
| } |
| |
| static int __init mmc_block_test_init(void) |
| { |
| mbtd = kzalloc(sizeof(struct mmc_block_test_data), GFP_KERNEL); |
| if (!mbtd) { |
| test_pr_err("%s: failed to allocate mmc_block_test_data", |
| __func__); |
| return -ENODEV; |
| } |
| |
| init_waitqueue_head(&mbtd->bkops_wait_q); |
| mbtd->bdt.init_fn = mmc_block_test_probe; |
| mbtd->bdt.exit_fn = mmc_block_test_remove; |
| INIT_LIST_HEAD(&mbtd->bdt.list); |
| test_iosched_register(&mbtd->bdt); |
| |
| return 0; |
| } |
| |
| static void __exit mmc_block_test_exit(void) |
| { |
| test_iosched_unregister(&mbtd->bdt); |
| kfree(mbtd); |
| } |
| |
| module_init(mmc_block_test_init); |
| module_exit(mmc_block_test_exit); |
| |
| MODULE_LICENSE("GPL v2"); |
| MODULE_DESCRIPTION("MMC block test"); |