rdma: cleanup and fixes

Mostly style cleanups, but also a fix for checking and using
OS rand if set. rdma should just always use the internal rand
generator.

Signed-off-by: Jens Axboe <axboe@kernel.dk>
diff --git a/engines/rdma.c b/engines/rdma.c
index 4d81c7d..9b18301 100644
--- a/engines/rdma.c
+++ b/engines/rdma.c
@@ -18,13 +18,13 @@
  * Exchanging steps of RDMA ioengine control messages:
  *	1. client side sends test mode (RDMA_WRITE/RDMA_READ/SEND)
  *	   to server side.
- * 	2. server side parses test mode, and sends back confirmation
+ *	2. server side parses test mode, and sends back confirmation
  *	   to client side. In RDMA WRITE/READ test, this confirmation
- * 	   includes memory information, such as rkey, address.
+ *	   includes memory information, such as rkey, address.
  *	3. client side initiates test loop.
- * 	4. In RDMA WRITE/READ test, client side sends a completion
+ *	4. In RDMA WRITE/READ test, client side sends a completion
  *	   notification to server side. Server side updates its
- * 	   td->done as true.
+ *	   td->done as true.
  *
  */
 #include <stdio.h>
@@ -325,6 +325,7 @@
 		rd->pd = ibv_alloc_pd(rd->child_cm_id->verbs);
 	else
 		rd->pd = ibv_alloc_pd(rd->cm_id->verbs);
+
 	if (rd->pd == NULL) {
 		log_err("fio: ibv_alloc_pd fail\n");
 		return 1;
@@ -416,7 +417,7 @@
 	/* setup work request */
 	/* recv wq */
 	rd->recv_sgl.addr = (uint64_t) (unsigned long)&rd->recv_buf;
-	rd->recv_sgl.length = sizeof rd->recv_buf;
+	rd->recv_sgl.length = sizeof(rd->recv_buf);
 	rd->recv_sgl.lkey = rd->recv_mr->lkey;
 	rd->rq_wr.sg_list = &rd->recv_sgl;
 	rd->rq_wr.num_sge = 1;
@@ -424,7 +425,7 @@
 
 	/* send wq */
 	rd->send_sgl.addr = (uint64_t) (unsigned long)&rd->send_buf;
-	rd->send_sgl.length = sizeof rd->send_buf;
+	rd->send_sgl.length = sizeof(rd->send_buf);
 	rd->send_sgl.lkey = rd->send_mr->lkey;
 
 	rd->sq_wr.opcode = IBV_WR_SEND;
@@ -441,13 +442,12 @@
 				  enum rdma_cm_event_type wait_event)
 {
 	struct rdmaio_data *rd = td->io_ops->data;
-
-	int ret;
 	struct rdma_cm_event *event;
+	int ret;
 
 	ret = rdma_get_cm_event(channel, &event);
 	if (ret) {
-		log_err("fio: rdma_get_cm_event");
+		log_err("fio: rdma_get_cm_event: %d\n", ret);
 		return 1;
 	}
 
@@ -521,9 +521,9 @@
 	int i;
 
 	io_u = rd->io_us_completed[0];
-	for (i = 0; i < rd->io_u_completed_nr - 1; i++) {
+	for (i = 0; i < rd->io_u_completed_nr - 1; i++)
 		rd->io_us_completed[i] = rd->io_us_completed[i + 1];
-	}
+
 	rd->io_u_completed_nr--;
 
 	dprint_io_u(io_u, "fio_rdmaio_event");
@@ -535,14 +535,11 @@
 				unsigned int max, struct timespec *t)
 {
 	struct rdmaio_data *rd = td->io_ops->data;
-	int r;
 	enum ibv_wc_opcode comp_opcode;
 	comp_opcode = IBV_WC_RDMA_WRITE;
 	struct ibv_cq *ev_cq;
 	void *ev_ctx;
-	int ret;
-
-	r = 0;
+	int ret, r = 0;
 
 	switch (rd->rdma_protocol) {
 	case FIO_RDMA_MEM_WRITE:
@@ -617,10 +614,7 @@
 		case FIO_RDMA_MEM_WRITE:
 			/* compose work request */
 			r_io_u_d = io_us[i]->engine_data;
-			if (td->o.use_os_rand)
-				index = os_random_long(&td->random_state) % rd->rmt_nr;
-			else
-				index = __rand(&rd->rand_state) % rd->rmt_nr;
+			index = __rand(&rd->rand_state) % rd->rmt_nr;
 			r_io_u_d->sq_wr.opcode = IBV_WR_RDMA_WRITE;
 			r_io_u_d->sq_wr.wr.rdma.rkey = rd->rmt_us[index].rkey;
 			r_io_u_d->sq_wr.wr.rdma.remote_addr = \
@@ -630,10 +624,7 @@
 		case FIO_RDMA_MEM_READ:
 			/* compose work request */
 			r_io_u_d = io_us[i]->engine_data;
-			if (td->o.use_os_rand)
-				index = os_random_long(&td->random_state) % rd->rmt_nr;
-			else
-				index = __rand(&rd->rand_state) % rd->rmt_nr;
+			index = __rand(&rd->rand_state) % rd->rmt_nr;
 			r_io_u_d->sq_wr.opcode = IBV_WR_RDMA_READ;
 			r_io_u_d->sq_wr.wr.rdma.rkey = rd->rmt_us[index].rkey;
 			r_io_u_d->sq_wr.wr.rdma.remote_addr = \
@@ -755,11 +746,11 @@
 	io_us = rd->io_us_queued;
 	do {
 		/* RDMA_WRITE or RDMA_READ */
-		if (rd->is_client) {
+		if (rd->is_client)
 			ret = fio_rdmaio_send(td, io_us, rd->io_u_queued_nr);
-		} else if (!rd->is_client) {
+		else if (!rd->is_client)
 			ret = fio_rdmaio_recv(td, io_us, rd->io_u_queued_nr);
-		} else
+		else
 			ret = 0;	/* must be a SYNC */
 
 		if (ret > 0) {
@@ -781,7 +772,7 @@
 	struct rdma_conn_param conn_param;
 	struct ibv_send_wr *bad_wr;
 
-	memset(&conn_param, 0, sizeof conn_param);
+	memset(&conn_param, 0, sizeof(conn_param));
 	conn_param.responder_resources = 1;
 	conn_param.initiator_depth = 1;
 	conn_param.retry_count = 10;
@@ -814,7 +805,7 @@
 	/* In SEND/RECV test, it's a good practice to setup the iodepth of
 	 * of the RECV side deeper than that of the SEND side to
 	 * avoid RNR (receiver not ready) error. The
-	 * SEND side may send so many unsolicited message before 
+	 * SEND side may send so many unsolicited message before
 	 * RECV side commits sufficient recv buffers into recv queue.
 	 * This may lead to RNR error. Here, SEND side pauses for a while
 	 * during which RECV side commits sufficient recv buffers.
@@ -831,7 +822,7 @@
 	struct ibv_send_wr *bad_wr;
 
 	/* rdma_accept() - then wait for accept success */
-	memset(&conn_param, 0, sizeof conn_param);
+	memset(&conn_param, 0, sizeof(conn_param));
 	conn_param.responder_resources = 1;
 	conn_param.initiator_depth = 1;
 
@@ -894,14 +885,17 @@
 		rdma_disconnect(rd->cm_id);
 	else {
 		rdma_disconnect(rd->child_cm_id);
-/*        rdma_disconnect(rd->cm_id); */
+#if 0
+		rdma_disconnect(rd->cm_id);
+#endif
 	}
 
-/*    if (get_next_channel_event(td, rd->cm_channel, RDMA_CM_EVENT_DISCONNECTED) != 0)
-    {
-        log_err("fio: wait for RDMA_CM_EVENT_DISCONNECTED\n");
-        return 1;
-    }*/
+#if 0
+	if (get_next_channel_event(td, rd->cm_channel, RDMA_CM_EVENT_DISCONNECTED) != 0) {
+		log_err("fio: wait for RDMA_CM_EVENT_DISCONNECTED\n");
+		return 1;
+	}
+#endif
 
 	ibv_destroy_cq(rd->cq);
 	ibv_destroy_qp(rd->qp);
@@ -924,6 +918,7 @@
 {
 	struct rdmaio_data *rd = td->io_ops->data;
 	struct ibv_recv_wr *bad_wr;
+	int err;
 
 	rd->addr.sin_family = AF_INET;
 	rd->addr.sin_port = htons(port);
@@ -941,28 +936,28 @@
 	}
 
 	/* resolve route */
-	if (rdma_resolve_addr(rd->cm_id, NULL,
-			      (struct sockaddr *)&rd->addr, 2000) != 0) {
-		log_err("fio: rdma_resolve_addr");
+	err = rdma_resolve_addr(rd->cm_id, NULL, (struct sockaddr *)&rd->addr, 2000);
+	if (err != 0) {
+		log_err("fio: rdma_resolve_addr: %d\n", err);
 		return 1;
 	}
 
-	if (get_next_channel_event
-	    (td, rd->cm_channel, RDMA_CM_EVENT_ADDR_RESOLVED)
-	    != 0) {
-		log_err("fio: get_next_channel_event");
+	err = get_next_channel_event(td, rd->cm_channel, RDMA_CM_EVENT_ADDR_RESOLVED);
+	if (err != 0) {
+		log_err("fio: get_next_channel_event: %d\n", err);
 		return 1;
 	}
 
 	/* resolve route */
-	if (rdma_resolve_route(rd->cm_id, 2000) != 0) {
-		log_err("fio: rdma_resolve_route");
+	err = rdma_resolve_route(rd->cm_id, 2000);
+	if (err != 0) {
+		log_err("fio: rdma_resolve_route: %d\n", err);
 		return 1;
 	}
 
-	if (get_next_channel_event
-	    (td, rd->cm_channel, RDMA_CM_EVENT_ROUTE_RESOLVED) != 0) {
-		log_err("fio: get_next_channel_event");
+	err = get_next_channel_event(td, rd->cm_channel, RDMA_CM_EVENT_ROUTE_RESOLVED);
+	if (err != 0) {
+		log_err("fio: get_next_channel_event: %d\n", err);
 		return 1;
 	}
 
@@ -974,8 +969,9 @@
 		return 1;
 
 	/* post recv buf */
-	if (ibv_post_recv(rd->qp, &rd->rq_wr, &bad_wr) != 0) {
-		log_err("fio: ibv_post_recv fail\n");
+	err = ibv_post_recv(rd->qp, &rd->rq_wr, &bad_wr);
+	if (err != 0) {
+		log_err("fio: ibv_post_recv fail: %d\n", err);
 		return 1;
 	}
 
@@ -1027,10 +1023,12 @@
 static int fio_rdmaio_init(struct thread_data *td)
 {
 	struct rdmaio_data *rd = td->io_ops->data;
+	struct flist_head *entry;
+	unsigned int max_bs;
 	unsigned int port;
 	char host[64], buf[128];
 	char *sep, *portp, *modep;
-	int ret;
+	int ret, i = 0;
 	struct rlimit rl;
 
 	if (td_rw(td)) {
@@ -1150,11 +1148,8 @@
 		ret = fio_rdmaio_setup_connect(td, host, port);
 	}
 
-	struct flist_head *entry;
-	unsigned int max_bs;
 	max_bs = max(td->o.max_bs[DDIR_READ], td->o.max_bs[DDIR_WRITE]);
 	/* register each io_u in the free list */
-	int i = 0;
 	flist_for_each(entry, &td->io_u_freelist) {
 		struct io_u *io_u = flist_entry(entry, struct io_u, list);
 
@@ -1176,8 +1171,9 @@
 		rd->send_buf.rmt_us[i].rkey = htonl(io_u->mr->rkey);
 		rd->send_buf.rmt_us[i].size = htonl(max_bs);
 
-/*    log_info("fio: Send rkey %x addr %" PRIx64 " len %d to client\n",
-          io_u->mr->rkey, io_u->buf, max_bs); */
+#if 0
+		log_info("fio: Send rkey %x addr %" PRIx64 " len %d to client\n", io_u->mr->rkey, io_u->buf, max_bs); */
+#endif
 		i++;
 	}
 
@@ -1193,16 +1189,8 @@
 {
 	struct rdmaio_data *rd = td->io_ops->data;
 
-	if (rd) {
-/*        if (nd->listenfd != -1)
-            close(nd->listenfd);
-        if (nd->pipes[0] != -1)
-            close(nd->pipes[0]);
-        if (nd->pipes[1] != -1)
-            close(nd->pipes[1]);
-*/
+	if (rd)
 		free(rd);
-	}
 }
 
 static int fio_rdmaio_setup(struct thread_data *td)
@@ -1210,7 +1198,7 @@
 	struct rdmaio_data *rd;
 
 	if (!td->io_ops->data) {
-		rd = malloc(sizeof(*rd));;
+		rd = malloc(sizeof(*rd));
 
 		memset(rd, 0, sizeof(*rd));
 		init_rand_seed(&rd->rand_state, (unsigned int) GOLDEN_RATIO_PRIME);