Tom Tucker | f94b533 | 2006-09-22 15:22:48 -0700 | [diff] [blame] | 1 | /* |
| 2 | * Copyright (c) 2005 Ammasso, Inc. All rights reserved. |
| 3 | * Copyright (c) 2005 Open Grid Computing, Inc. All rights reserved. |
| 4 | * |
| 5 | * This software is available to you under a choice of one of two |
| 6 | * licenses. You may choose to be licensed under the terms of the GNU |
| 7 | * General Public License (GPL) Version 2, available from the file |
| 8 | * COPYING in the main directory of this source tree, or the |
| 9 | * OpenIB.org BSD license below: |
| 10 | * |
| 11 | * Redistribution and use in source and binary forms, with or |
| 12 | * without modification, are permitted provided that the following |
| 13 | * conditions are met: |
| 14 | * |
| 15 | * - Redistributions of source code must retain the above |
| 16 | * copyright notice, this list of conditions and the following |
| 17 | * disclaimer. |
| 18 | * |
| 19 | * - Redistributions in binary form must reproduce the above |
| 20 | * copyright notice, this list of conditions and the following |
| 21 | * disclaimer in the documentation and/or other materials |
| 22 | * provided with the distribution. |
| 23 | * |
| 24 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, |
| 25 | * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
| 26 | * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND |
| 27 | * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS |
| 28 | * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN |
| 29 | * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN |
| 30 | * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
| 31 | * SOFTWARE. |
| 32 | */ |
Tejun Heo | 5a0e3ad | 2010-03-24 17:04:11 +0900 | [diff] [blame] | 33 | #include <linux/slab.h> |
| 34 | |
Tom Tucker | f94b533 | 2006-09-22 15:22:48 -0700 | [diff] [blame] | 35 | #include "c2.h" |
| 36 | #include "c2_vq.h" |
| 37 | |
| 38 | #define PBL_VIRT 1 |
| 39 | #define PBL_PHYS 2 |
| 40 | |
| 41 | /* |
| 42 | * Send all the PBL messages to convey the remainder of the PBL |
| 43 | * Wait for the adapter's reply on the last one. |
| 44 | * This is indicated by setting the MEM_PBL_COMPLETE in the flags. |
| 45 | * |
| 46 | * NOTE: vq_req is _not_ freed by this function. The VQ Host |
| 47 | * Reply buffer _is_ freed by this function. |
| 48 | */ |
| 49 | static int |
Roland Dreier | dc544bc | 2008-04-16 21:01:08 -0700 | [diff] [blame] | 50 | send_pbl_messages(struct c2_dev *c2dev, __be32 stag_index, |
Tom Tucker | f94b533 | 2006-09-22 15:22:48 -0700 | [diff] [blame] | 51 | unsigned long va, u32 pbl_depth, |
| 52 | struct c2_vq_req *vq_req, int pbl_type) |
| 53 | { |
| 54 | u32 pbe_count; /* amt that fits in a PBL msg */ |
| 55 | u32 count; /* amt in this PBL MSG. */ |
| 56 | struct c2wr_nsmr_pbl_req *wr; /* PBL WR ptr */ |
| 57 | struct c2wr_nsmr_pbl_rep *reply; /* reply ptr */ |
| 58 | int err, pbl_virt, pbl_index, i; |
| 59 | |
| 60 | switch (pbl_type) { |
| 61 | case PBL_VIRT: |
| 62 | pbl_virt = 1; |
| 63 | break; |
| 64 | case PBL_PHYS: |
| 65 | pbl_virt = 0; |
| 66 | break; |
| 67 | default: |
| 68 | return -EINVAL; |
| 69 | break; |
| 70 | } |
| 71 | |
| 72 | pbe_count = (c2dev->req_vq.msg_size - |
| 73 | sizeof(struct c2wr_nsmr_pbl_req)) / sizeof(u64); |
| 74 | wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL); |
| 75 | if (!wr) { |
| 76 | return -ENOMEM; |
| 77 | } |
| 78 | c2_wr_set_id(wr, CCWR_NSMR_PBL); |
| 79 | |
| 80 | /* |
| 81 | * Only the last PBL message will generate a reply from the verbs, |
| 82 | * so we set the context to 0 indicating there is no kernel verbs |
| 83 | * handler blocked awaiting this reply. |
| 84 | */ |
| 85 | wr->hdr.context = 0; |
| 86 | wr->rnic_handle = c2dev->adapter_handle; |
| 87 | wr->stag_index = stag_index; /* already swapped */ |
| 88 | wr->flags = 0; |
| 89 | pbl_index = 0; |
| 90 | while (pbl_depth) { |
| 91 | count = min(pbe_count, pbl_depth); |
| 92 | wr->addrs_length = cpu_to_be32(count); |
| 93 | |
| 94 | /* |
| 95 | * If this is the last message, then reference the |
| 96 | * vq request struct cuz we're gonna wait for a reply. |
| 97 | * also make this PBL msg as the last one. |
| 98 | */ |
| 99 | if (count == pbl_depth) { |
| 100 | /* |
| 101 | * reference the request struct. dereferenced in the |
| 102 | * int handler. |
| 103 | */ |
| 104 | vq_req_get(c2dev, vq_req); |
| 105 | wr->flags = cpu_to_be32(MEM_PBL_COMPLETE); |
| 106 | |
| 107 | /* |
| 108 | * This is the last PBL message. |
| 109 | * Set the context to our VQ Request Object so we can |
| 110 | * wait for the reply. |
| 111 | */ |
| 112 | wr->hdr.context = (unsigned long) vq_req; |
| 113 | } |
| 114 | |
| 115 | /* |
| 116 | * If pbl_virt is set then va is a virtual address |
| 117 | * that describes a virtually contiguous memory |
| 118 | * allocation. The wr needs the start of each virtual page |
| 119 | * to be converted to the corresponding physical address |
| 120 | * of the page. If pbl_virt is not set then va is an array |
| 121 | * of physical addresses and there is no conversion to do. |
| 122 | * Just fill in the wr with what is in the array. |
| 123 | */ |
| 124 | for (i = 0; i < count; i++) { |
| 125 | if (pbl_virt) { |
| 126 | va += PAGE_SIZE; |
| 127 | } else { |
| 128 | wr->paddrs[i] = |
| 129 | cpu_to_be64(((u64 *)va)[pbl_index + i]); |
| 130 | } |
| 131 | } |
| 132 | |
| 133 | /* |
| 134 | * Send WR to adapter |
| 135 | */ |
| 136 | err = vq_send_wr(c2dev, (union c2wr *) wr); |
| 137 | if (err) { |
| 138 | if (count <= pbe_count) { |
| 139 | vq_req_put(c2dev, vq_req); |
| 140 | } |
| 141 | goto bail0; |
| 142 | } |
| 143 | pbl_depth -= count; |
| 144 | pbl_index += count; |
| 145 | } |
| 146 | |
| 147 | /* |
| 148 | * Now wait for the reply... |
| 149 | */ |
| 150 | err = vq_wait_for_reply(c2dev, vq_req); |
| 151 | if (err) { |
| 152 | goto bail0; |
| 153 | } |
| 154 | |
| 155 | /* |
| 156 | * Process reply |
| 157 | */ |
| 158 | reply = (struct c2wr_nsmr_pbl_rep *) (unsigned long) vq_req->reply_msg; |
| 159 | if (!reply) { |
| 160 | err = -ENOMEM; |
| 161 | goto bail0; |
| 162 | } |
| 163 | |
| 164 | err = c2_errno(reply); |
| 165 | |
| 166 | vq_repbuf_free(c2dev, reply); |
| 167 | bail0: |
| 168 | kfree(wr); |
| 169 | return err; |
| 170 | } |
| 171 | |
| 172 | #define C2_PBL_MAX_DEPTH 131072 |
| 173 | int |
| 174 | c2_nsmr_register_phys_kern(struct c2_dev *c2dev, u64 *addr_list, |
| 175 | int page_size, int pbl_depth, u32 length, |
| 176 | u32 offset, u64 *va, enum c2_acf acf, |
| 177 | struct c2_mr *mr) |
| 178 | { |
| 179 | struct c2_vq_req *vq_req; |
| 180 | struct c2wr_nsmr_register_req *wr; |
| 181 | struct c2wr_nsmr_register_rep *reply; |
| 182 | u16 flags; |
| 183 | int i, pbe_count, count; |
| 184 | int err; |
| 185 | |
| 186 | if (!va || !length || !addr_list || !pbl_depth) |
| 187 | return -EINTR; |
| 188 | |
| 189 | /* |
| 190 | * Verify PBL depth is within rnic max |
| 191 | */ |
| 192 | if (pbl_depth > C2_PBL_MAX_DEPTH) { |
| 193 | return -EINTR; |
| 194 | } |
| 195 | |
| 196 | /* |
| 197 | * allocate verbs request object |
| 198 | */ |
| 199 | vq_req = vq_req_alloc(c2dev); |
| 200 | if (!vq_req) |
| 201 | return -ENOMEM; |
| 202 | |
| 203 | wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL); |
| 204 | if (!wr) { |
| 205 | err = -ENOMEM; |
| 206 | goto bail0; |
| 207 | } |
| 208 | |
| 209 | /* |
| 210 | * build the WR |
| 211 | */ |
| 212 | c2_wr_set_id(wr, CCWR_NSMR_REGISTER); |
| 213 | wr->hdr.context = (unsigned long) vq_req; |
| 214 | wr->rnic_handle = c2dev->adapter_handle; |
| 215 | |
| 216 | flags = (acf | MEM_VA_BASED | MEM_REMOTE); |
| 217 | |
| 218 | /* |
| 219 | * compute how many pbes can fit in the message |
| 220 | */ |
| 221 | pbe_count = (c2dev->req_vq.msg_size - |
| 222 | sizeof(struct c2wr_nsmr_register_req)) / sizeof(u64); |
| 223 | |
| 224 | if (pbl_depth <= pbe_count) { |
| 225 | flags |= MEM_PBL_COMPLETE; |
| 226 | } |
| 227 | wr->flags = cpu_to_be16(flags); |
| 228 | wr->stag_key = 0; //stag_key; |
| 229 | wr->va = cpu_to_be64(*va); |
| 230 | wr->pd_id = mr->pd->pd_id; |
| 231 | wr->pbe_size = cpu_to_be32(page_size); |
| 232 | wr->length = cpu_to_be32(length); |
| 233 | wr->pbl_depth = cpu_to_be32(pbl_depth); |
| 234 | wr->fbo = cpu_to_be32(offset); |
| 235 | count = min(pbl_depth, pbe_count); |
| 236 | wr->addrs_length = cpu_to_be32(count); |
| 237 | |
| 238 | /* |
| 239 | * fill out the PBL for this message |
| 240 | */ |
| 241 | for (i = 0; i < count; i++) { |
| 242 | wr->paddrs[i] = cpu_to_be64(addr_list[i]); |
| 243 | } |
| 244 | |
| 245 | /* |
| 246 | * regerence the request struct |
| 247 | */ |
| 248 | vq_req_get(c2dev, vq_req); |
| 249 | |
| 250 | /* |
| 251 | * send the WR to the adapter |
| 252 | */ |
| 253 | err = vq_send_wr(c2dev, (union c2wr *) wr); |
| 254 | if (err) { |
| 255 | vq_req_put(c2dev, vq_req); |
| 256 | goto bail1; |
| 257 | } |
| 258 | |
| 259 | /* |
| 260 | * wait for reply from adapter |
| 261 | */ |
| 262 | err = vq_wait_for_reply(c2dev, vq_req); |
| 263 | if (err) { |
| 264 | goto bail1; |
| 265 | } |
| 266 | |
| 267 | /* |
| 268 | * process reply |
| 269 | */ |
| 270 | reply = |
| 271 | (struct c2wr_nsmr_register_rep *) (unsigned long) (vq_req->reply_msg); |
| 272 | if (!reply) { |
| 273 | err = -ENOMEM; |
| 274 | goto bail1; |
| 275 | } |
| 276 | if ((err = c2_errno(reply))) { |
| 277 | goto bail2; |
| 278 | } |
| 279 | //*p_pb_entries = be32_to_cpu(reply->pbl_depth); |
| 280 | mr->ibmr.lkey = mr->ibmr.rkey = be32_to_cpu(reply->stag_index); |
| 281 | vq_repbuf_free(c2dev, reply); |
| 282 | |
| 283 | /* |
| 284 | * if there are still more PBEs we need to send them to |
| 285 | * the adapter and wait for a reply on the final one. |
| 286 | * reuse vq_req for this purpose. |
| 287 | */ |
| 288 | pbl_depth -= count; |
| 289 | if (pbl_depth) { |
| 290 | |
| 291 | vq_req->reply_msg = (unsigned long) NULL; |
| 292 | atomic_set(&vq_req->reply_ready, 0); |
| 293 | err = send_pbl_messages(c2dev, |
| 294 | cpu_to_be32(mr->ibmr.lkey), |
| 295 | (unsigned long) &addr_list[i], |
| 296 | pbl_depth, vq_req, PBL_PHYS); |
| 297 | if (err) { |
| 298 | goto bail1; |
| 299 | } |
| 300 | } |
| 301 | |
| 302 | vq_req_free(c2dev, vq_req); |
| 303 | kfree(wr); |
| 304 | |
| 305 | return err; |
| 306 | |
| 307 | bail2: |
| 308 | vq_repbuf_free(c2dev, reply); |
| 309 | bail1: |
| 310 | kfree(wr); |
| 311 | bail0: |
| 312 | vq_req_free(c2dev, vq_req); |
| 313 | return err; |
| 314 | } |
| 315 | |
| 316 | int c2_stag_dealloc(struct c2_dev *c2dev, u32 stag_index) |
| 317 | { |
| 318 | struct c2_vq_req *vq_req; /* verbs request object */ |
| 319 | struct c2wr_stag_dealloc_req wr; /* work request */ |
| 320 | struct c2wr_stag_dealloc_rep *reply; /* WR reply */ |
| 321 | int err; |
| 322 | |
| 323 | |
| 324 | /* |
| 325 | * allocate verbs request object |
| 326 | */ |
| 327 | vq_req = vq_req_alloc(c2dev); |
| 328 | if (!vq_req) { |
| 329 | return -ENOMEM; |
| 330 | } |
| 331 | |
| 332 | /* |
| 333 | * Build the WR |
| 334 | */ |
| 335 | c2_wr_set_id(&wr, CCWR_STAG_DEALLOC); |
| 336 | wr.hdr.context = (u64) (unsigned long) vq_req; |
| 337 | wr.rnic_handle = c2dev->adapter_handle; |
| 338 | wr.stag_index = cpu_to_be32(stag_index); |
| 339 | |
| 340 | /* |
| 341 | * reference the request struct. dereferenced in the int handler. |
| 342 | */ |
| 343 | vq_req_get(c2dev, vq_req); |
| 344 | |
| 345 | /* |
| 346 | * Send WR to adapter |
| 347 | */ |
| 348 | err = vq_send_wr(c2dev, (union c2wr *) & wr); |
| 349 | if (err) { |
| 350 | vq_req_put(c2dev, vq_req); |
| 351 | goto bail0; |
| 352 | } |
| 353 | |
| 354 | /* |
| 355 | * Wait for reply from adapter |
| 356 | */ |
| 357 | err = vq_wait_for_reply(c2dev, vq_req); |
| 358 | if (err) { |
| 359 | goto bail0; |
| 360 | } |
| 361 | |
| 362 | /* |
| 363 | * Process reply |
| 364 | */ |
| 365 | reply = (struct c2wr_stag_dealloc_rep *) (unsigned long) vq_req->reply_msg; |
| 366 | if (!reply) { |
| 367 | err = -ENOMEM; |
| 368 | goto bail0; |
| 369 | } |
| 370 | |
| 371 | err = c2_errno(reply); |
| 372 | |
| 373 | vq_repbuf_free(c2dev, reply); |
| 374 | bail0: |
| 375 | vq_req_free(c2dev, vq_req); |
| 376 | return err; |
| 377 | } |