43#include <linux/list.h> 44#include <linux/in.h> 45#include <linux/device.h> 46#include <linux/pci.h> 47#include <linux/sched.h> 48 49#include <asm/atomic.h> 50 51#include <rdma/ib_verbs.h> 52#include <rdma/rdma_cm.h> 53 54#include "krping.h" 55#include "getopt.h" 56 57extern int krping_debug; 58#define DEBUG_LOG(cb, x...) if (krping_debug) krping_printf((cb)->cookie, x) 59#define PRINTF(cb, x...) krping_printf((cb)->cookie, x) 60 61MODULE_AUTHOR("Steve Wise"); 62MODULE_DESCRIPTION("RDMA ping client/server"); 63MODULE_LICENSE("Dual BSD/GPL"); 64 65static __inline uint64_t 66get_cycles(void) 67{ 68 uint32_t low, high; 69 __asm __volatile("rdtsc" : "=a" (low), "=d" (high)); 70 return (low | ((u_int64_t)high << 32)); 71} 72 73typedef uint64_t cycles_t; 74 75enum mem_type { 76 DMA = 1, 77 FASTREG = 2, 78 MW = 3, 79 MR = 4 80}; 81 82static const struct krping_option krping_opts[] = { 83 {"count", OPT_INT, 'C'}, 84 {"size", OPT_INT, 'S'}, 85 {"addr", OPT_STRING, 'a'}, 86 {"port", OPT_INT, 'p'}, 87 {"verbose", OPT_NOPARAM, 'v'}, 88 {"validate", OPT_NOPARAM, 'V'}, 89 {"server", OPT_NOPARAM, 's'}, 90 {"client", OPT_NOPARAM, 'c'}, 91 {"mem_mode", OPT_STRING, 'm'}, 92 {"server_inv", OPT_NOPARAM, 'I'}, 93 {"wlat", OPT_NOPARAM, 'l'}, 94 {"rlat", OPT_NOPARAM, 'L'}, 95 {"bw", OPT_NOPARAM, 'B'}, 96 {"duplex", OPT_NOPARAM, 'd'}, 97 {"txdepth", OPT_INT, 'T'}, 98 {"poll", OPT_NOPARAM, 'P'}, 99 {"local_dma_lkey", OPT_NOPARAM, 'Z'}, 100 {"read_inv", OPT_NOPARAM, 'R'}, 101 {"fr", OPT_NOPARAM, 'f'}, 102 {NULL, 0, 0} 103}; 104 105#define htonll(x) cpu_to_be64((x)) 106#define ntohll(x) cpu_to_be64((x)) 107 108static struct mutex krping_mutex; 109 110/* 111 * List of running krping threads. 112 */ 113static LIST_HEAD(krping_cbs); 114 115/* 116 * krping "ping/pong" loop: 117 * client sends source rkey/addr/len 118 * server receives source rkey/add/len 119 * server rdma reads "ping" data from source 120 * server sends "go ahead" on rdma read completion 121 * client sends sink rkey/addr/len 122 * server receives sink rkey/addr/len 123 * server rdma writes "pong" data to sink 124 * server sends "go ahead" on rdma write completion 125 * <repeat loop> 126 */ 127 128/* 129 * These states are used to signal events between the completion handler 130 * and the main client or server thread. 131 * 132 * Once CONNECTED, they cycle through RDMA_READ_ADV, RDMA_WRITE_ADV, 133 * and RDMA_WRITE_COMPLETE for each ping. 134 */ 135enum test_state { 136 IDLE = 1, 137 CONNECT_REQUEST, 138 ADDR_RESOLVED, 139 ROUTE_RESOLVED, 140 CONNECTED, 141 RDMA_READ_ADV, 142 RDMA_READ_COMPLETE, 143 RDMA_WRITE_ADV, 144 RDMA_WRITE_COMPLETE, 145 ERROR 146}; 147 148struct krping_rdma_info { 149 uint64_t buf; 150 uint32_t rkey; 151 uint32_t size; 152}; 153 154/* 155 * Default max buffer size for IO... 156 */ 157#define RPING_BUFSIZE 128*1024 158#define RPING_SQ_DEPTH 64 159 160/* 161 * Control block struct. 162 */ 163struct krping_cb { 164 void *cookie; 165 int server; /* 0 iff client */ 166 struct ib_cq *cq; 167 struct ib_pd *pd; 168 struct ib_qp *qp; 169 170 enum mem_type mem; 171 struct ib_mr *dma_mr; 172 173 struct ib_fast_reg_page_list *page_list; 174 int page_list_len; 175 struct ib_send_wr fastreg_wr; 176 struct ib_send_wr invalidate_wr; 177 struct ib_mr *fastreg_mr; 178 int server_invalidate; 179 int read_inv; 180 u8 key; 181 182 struct ib_mw *mw; 183 struct ib_mw_bind bind_attr; 184 185 struct ib_recv_wr rq_wr; /* recv work request record */ 186 struct ib_sge recv_sgl; /* recv single SGE */ 187 struct krping_rdma_info recv_buf;/* malloc'd buffer */ 188 u64 recv_dma_addr; 189 DECLARE_PCI_UNMAP_ADDR(recv_mapping) 190 struct ib_mr *recv_mr; 191 192 struct ib_send_wr sq_wr; /* send work requrest record */ 193 struct ib_sge send_sgl; 194 struct krping_rdma_info send_buf;/* single send buf */ 195 u64 send_dma_addr; 196 DECLARE_PCI_UNMAP_ADDR(send_mapping) 197 struct ib_mr *send_mr; 198 199 struct ib_send_wr rdma_sq_wr; /* rdma work request record */ 200 struct ib_sge rdma_sgl; /* rdma single SGE */ 201 char *rdma_buf; /* used as rdma sink */ 202 u64 rdma_dma_addr; 203 DECLARE_PCI_UNMAP_ADDR(rdma_mapping) 204 struct ib_mr *rdma_mr; 205 206 uint32_t remote_rkey; /* remote guys RKEY */ 207 uint64_t remote_addr; /* remote guys TO */ 208 uint32_t remote_len; /* remote guys LEN */ 209 210 char *start_buf; /* rdma read src */ 211 u64 start_dma_addr; 212 DECLARE_PCI_UNMAP_ADDR(start_mapping) 213 struct ib_mr *start_mr; 214 215 enum test_state state; /* used for cond/signalling */ 216 wait_queue_head_t sem; 217 struct krping_stats stats; 218 219 uint16_t port; /* dst port in NBO */ 220 struct in_addr addr; /* dst addr in NBO */ 221 char *addr_str; /* dst addr string */ 222 int verbose; /* verbose logging */ 223 int count; /* ping count */ 224 int size; /* ping data size */ 225 int validate; /* validate ping data */ 226 int wlat; /* run wlat test */ 227 int rlat; /* run rlat test */ 228 int bw; /* run bw test */ 229 int duplex; /* run bw full duplex test */ 230 int poll; /* poll or block for rlat test */ 231 int txdepth; /* SQ depth */ 232 int local_dma_lkey; /* use 0 for lkey */ 233 int frtest; /* fastreg test */ 234 235 /* CM stuff */ 236 struct rdma_cm_id *cm_id; /* connection on client side,*/ 237 /* listener on server side. */ 238 struct rdma_cm_id *child_cm_id; /* connection on server side */ 239 struct list_head list; 240}; 241 242static int krping_cma_event_handler(struct rdma_cm_id *cma_id, 243 struct rdma_cm_event *event) 244{ 245 int ret; 246 struct krping_cb *cb = cma_id->context; 247 248 DEBUG_LOG(cb, "cma_event type %d cma_id %p (%s)\n", event->event, 249 cma_id, (cma_id == cb->cm_id) ? "parent" : "child"); 250 251 switch (event->event) { 252 case RDMA_CM_EVENT_ADDR_RESOLVED: 253 cb->state = ADDR_RESOLVED; 254 ret = rdma_resolve_route(cma_id, 2000); 255 if (ret) { 256 PRINTF(cb, "rdma_resolve_route error %d\n", ret); 257 wake_up_interruptible(&cb->sem); 258 } 259 break; 260 261 case RDMA_CM_EVENT_ROUTE_RESOLVED: 262 cb->state = ROUTE_RESOLVED; 263 wake_up_interruptible(&cb->sem); 264 break; 265 266 case RDMA_CM_EVENT_CONNECT_REQUEST: 267 cb->state = CONNECT_REQUEST; 268 cb->child_cm_id = cma_id; 269 DEBUG_LOG(cb, "child cma %p\n", cb->child_cm_id); 270 wake_up_interruptible(&cb->sem); 271 break; 272 273 case RDMA_CM_EVENT_ESTABLISHED: 274 DEBUG_LOG(cb, "ESTABLISHED\n"); 275 if (!cb->server) { 276 cb->state = CONNECTED; 277 } 278 wake_up_interruptible(&cb->sem); 279 break; 280 281 case RDMA_CM_EVENT_ADDR_ERROR: 282 case RDMA_CM_EVENT_ROUTE_ERROR: 283 case RDMA_CM_EVENT_CONNECT_ERROR: 284 case RDMA_CM_EVENT_UNREACHABLE: 285 case RDMA_CM_EVENT_REJECTED: 286 PRINTF(cb, "cma event %d, error %d\n", event->event, 287 event->status); 288 cb->state = ERROR; 289 wake_up_interruptible(&cb->sem); 290 break; 291 292 case RDMA_CM_EVENT_DISCONNECTED: 293 PRINTF(cb, "DISCONNECT EVENT...\n"); 294 cb->state = ERROR; 295 wake_up_interruptible(&cb->sem); 296 break; 297 298 case RDMA_CM_EVENT_DEVICE_REMOVAL: 299 PRINTF(cb, "cma detected device removal!!!!\n"); 300 break; 301 302 default: 303 PRINTF(cb, "oof bad type!\n"); 304 wake_up_interruptible(&cb->sem); 305 break; 306 } 307 return 0; 308} 309 310static int server_recv(struct krping_cb *cb, struct ib_wc *wc) 311{ 312 if (wc->byte_len != sizeof(cb->recv_buf)) { 313 PRINTF(cb, "Received bogus data, size %d\n", 314 wc->byte_len); 315 return -1; 316 } 317 318 cb->remote_rkey = ntohl(cb->recv_buf.rkey); 319 cb->remote_addr = ntohll(cb->recv_buf.buf); 320 cb->remote_len = ntohl(cb->recv_buf.size); 321 DEBUG_LOG(cb, "Received rkey %x addr %llx len %d from peer\n", 322 cb->remote_rkey, (unsigned long long)cb->remote_addr, 323 cb->remote_len); 324 325 if (cb->state <= CONNECTED || cb->state == RDMA_WRITE_COMPLETE) 326 cb->state = RDMA_READ_ADV; 327 else 328 cb->state = RDMA_WRITE_ADV; 329 330 return 0; 331} 332 333static int client_recv(struct krping_cb *cb, struct ib_wc *wc) 334{ 335 if (wc->byte_len != sizeof(cb->recv_buf)) { 336 PRINTF(cb, "Received bogus data, size %d\n", 337 wc->byte_len); 338 return -1; 339 } 340 341 if (cb->state == RDMA_READ_ADV) 342 cb->state = RDMA_WRITE_ADV; 343 else 344 cb->state = RDMA_WRITE_COMPLETE; 345 346 return 0; 347} 348 349static void krping_cq_event_handler(struct ib_cq *cq, void *ctx) 350{ 351 struct krping_cb *cb = ctx; 352 struct ib_wc wc; 353 struct ib_recv_wr *bad_wr; 354 int ret; 355 356 BUG_ON(cb->cq != cq); 357 if (cb->state == ERROR) { 358 PRINTF(cb, "cq completion in ERROR state\n"); 359 return; 360 } 361 if (cb->frtest) { 362 PRINTF(cb, "cq completion event in frtest!\n"); 363 return; 364 } 365 if (!cb->wlat && !cb->rlat && !cb->bw) 366 ib_req_notify_cq(cb->cq, IB_CQ_NEXT_COMP); 367 while ((ret = ib_poll_cq(cb->cq, 1, &wc)) == 1) { 368 if (wc.status) { 369 if (wc.status == IB_WC_WR_FLUSH_ERR) { 370 DEBUG_LOG(cb, "cq flushed\n"); 371 continue; 372 } else { 373 PRINTF(cb, "cq completion failed with " 374 "wr_id %Lx status %d opcode %d vender_err %x\n", 375 wc.wr_id, wc.status, wc.opcode, wc.vendor_err); 376 goto error; 377 } 378 } 379 380 switch (wc.opcode) { 381 case IB_WC_SEND: 382 DEBUG_LOG(cb, "send completion\n"); 383 cb->stats.send_bytes += cb->send_sgl.length; 384 cb->stats.send_msgs++; 385 break; 386 387 case IB_WC_RDMA_WRITE: 388 DEBUG_LOG(cb, "rdma write completion\n"); 389 cb->stats.write_bytes += cb->rdma_sq_wr.sg_list->length; 390 cb->stats.write_msgs++; 391 cb->state = RDMA_WRITE_COMPLETE; 392 wake_up_interruptible(&cb->sem); 393 break; 394 395 case IB_WC_RDMA_READ: 396 DEBUG_LOG(cb, "rdma read completion\n"); 397 cb->stats.read_bytes += cb->rdma_sq_wr.sg_list->length; 398 cb->stats.read_msgs++; 399 cb->state = RDMA_READ_COMPLETE; 400 wake_up_interruptible(&cb->sem); 401 break; 402 403 case IB_WC_RECV: 404 DEBUG_LOG(cb, "recv completion\n"); 405 cb->stats.recv_bytes += sizeof(cb->recv_buf); 406 cb->stats.recv_msgs++; 407 if (cb->wlat || cb->rlat || cb->bw) 408 ret = server_recv(cb, &wc); 409 else 410 ret = cb->server ? server_recv(cb, &wc) : 411 client_recv(cb, &wc); 412 if (ret) { 413 PRINTF(cb, "recv wc error: %d\n", ret); 414 goto error; 415 } 416 417 ret = ib_post_recv(cb->qp, &cb->rq_wr, &bad_wr); 418 if (ret) { 419 PRINTF(cb, "post recv error: %d\n", 420 ret); 421 goto error; 422 } 423 wake_up_interruptible(&cb->sem); 424 break; 425 426 default: 427 PRINTF(cb, 428 "%s:%d Unexpected opcode %d, Shutting down\n", 429 __func__, __LINE__, wc.opcode); 430 goto error; 431 } 432 } 433 if (ret) { 434 PRINTF(cb, "poll error %d\n", ret); 435 goto error; 436 } 437 return; 438error: 439 cb->state = ERROR; 440 wake_up_interruptible(&cb->sem); 441} 442 443static int krping_accept(struct krping_cb *cb) 444{ 445 struct rdma_conn_param conn_param; 446 int ret; 447 448 DEBUG_LOG(cb, "accepting client connection request\n"); 449 450 memset(&conn_param, 0, sizeof conn_param); 451 conn_param.responder_resources = 1; 452 conn_param.initiator_depth = 1; 453 454 ret = rdma_accept(cb->child_cm_id, &conn_param); 455 if (ret) { 456 PRINTF(cb, "rdma_accept error: %d\n", ret); 457 return ret; 458 } 459 460 if (!cb->wlat && !cb->rlat && !cb->bw) { 461 wait_event_interruptible(cb->sem, cb->state >= CONNECTED); 462 if (cb->state == ERROR) { 463 PRINTF(cb, "wait for CONNECTED state %d\n", 464 cb->state); 465 return -1; 466 } 467 } 468 return 0; 469} 470 471static void krping_setup_wr(struct krping_cb *cb) 472{ 473 cb->recv_sgl.addr = cb->recv_dma_addr; 474 cb->recv_sgl.length = sizeof cb->recv_buf; 475 if (cb->local_dma_lkey) 476 cb->recv_sgl.lkey = cb->qp->device->local_dma_lkey; 477 else if (cb->mem == DMA) 478 cb->recv_sgl.lkey = cb->dma_mr->lkey; 479 else 480 cb->recv_sgl.lkey = cb->recv_mr->lkey; 481 cb->rq_wr.sg_list = &cb->recv_sgl; 482 cb->rq_wr.num_sge = 1; 483 484 cb->send_sgl.addr = cb->send_dma_addr; 485 cb->send_sgl.length = sizeof cb->send_buf; 486 if (cb->local_dma_lkey) 487 cb->send_sgl.lkey = cb->qp->device->local_dma_lkey; 488 else if (cb->mem == DMA) 489 cb->send_sgl.lkey = cb->dma_mr->lkey; 490 else 491 cb->send_sgl.lkey = cb->send_mr->lkey; 492 493 cb->sq_wr.opcode = IB_WR_SEND; 494 cb->sq_wr.send_flags = IB_SEND_SIGNALED; 495 cb->sq_wr.sg_list = &cb->send_sgl; 496 cb->sq_wr.num_sge = 1; 497 498 if (cb->server || cb->wlat || cb->rlat || cb->bw) { 499 cb->rdma_sgl.addr = cb->rdma_dma_addr; 500 if (cb->mem == MR) 501 cb->rdma_sgl.lkey = cb->rdma_mr->lkey; 502 cb->rdma_sq_wr.send_flags = IB_SEND_SIGNALED; 503 cb->rdma_sq_wr.sg_list = &cb->rdma_sgl; 504 cb->rdma_sq_wr.num_sge = 1; 505 } 506 507 switch(cb->mem) { 508 case FASTREG: 509 510 /* 511 * A chain of 2 WRs, INVALDATE_MR + FAST_REG_MR. 512 * both unsignaled. The client uses them to reregister 513 * the rdma buffers with a new key each iteration. 514 */ 515 cb->fastreg_wr.opcode = IB_WR_FAST_REG_MR; 516 cb->fastreg_wr.wr.fast_reg.page_shift = PAGE_SHIFT; 517 cb->fastreg_wr.wr.fast_reg.length = cb->size; 518 cb->fastreg_wr.wr.fast_reg.page_list = cb->page_list; 519 cb->fastreg_wr.wr.fast_reg.page_list_len = cb->page_list_len; 520 521 cb->invalidate_wr.next = &cb->fastreg_wr; 522 cb->invalidate_wr.opcode = IB_WR_LOCAL_INV; 523 break; 524 case MW: 525 cb->bind_attr.wr_id = 0xabbaabba; 526 cb->bind_attr.send_flags = 0; /* unsignaled */ 527 cb->bind_attr.length = cb->size; 528 break; 529 default: 530 break; 531 } 532} 533 534static int krping_setup_buffers(struct krping_cb *cb) 535{ 536 int ret; 537 struct ib_phys_buf buf; 538 u64 iovbase; 539 540 DEBUG_LOG(cb, "krping_setup_buffers called on cb %p\n", cb); 541 542 cb->recv_dma_addr = dma_map_single(cb->pd->device->dma_device, 543 &cb->recv_buf, 544 sizeof(cb->recv_buf), DMA_BIDIRECTIONAL); 545 pci_unmap_addr_set(cb, recv_mapping, cb->recv_dma_addr); 546 cb->send_dma_addr = dma_map_single(cb->pd->device->dma_device, 547 &cb->send_buf, sizeof(cb->send_buf), 548 DMA_BIDIRECTIONAL); 549 pci_unmap_addr_set(cb, send_mapping, cb->send_dma_addr); 550 551 if (cb->mem == DMA) { 552 cb->dma_mr = ib_get_dma_mr(cb->pd, IB_ACCESS_LOCAL_WRITE| 553 IB_ACCESS_REMOTE_READ| 554 IB_ACCESS_REMOTE_WRITE); 555 if (IS_ERR(cb->dma_mr)) { 556 DEBUG_LOG(cb, "reg_dmamr failed\n"); 557 ret = PTR_ERR(cb->dma_mr); 558 goto bail; 559 } 560 } else { 561 if (!cb->local_dma_lkey) { 562 buf.addr = cb->recv_dma_addr; 563 buf.size = sizeof cb->recv_buf; 564 DEBUG_LOG(cb, "recv buf dma_addr %llx size %d\n", buf.addr, 565 (int)buf.size); 566 iovbase = cb->recv_dma_addr; 567 cb->recv_mr = ib_reg_phys_mr(cb->pd, &buf, 1, 568 IB_ACCESS_LOCAL_WRITE, 569 &iovbase); 570 571 if (IS_ERR(cb->recv_mr)) { 572 DEBUG_LOG(cb, "recv_buf reg_mr failed\n"); 573 ret = PTR_ERR(cb->recv_mr); 574 goto bail; 575 } 576 577 buf.addr = cb->send_dma_addr; 578 buf.size = sizeof cb->send_buf; 579 DEBUG_LOG(cb, "send buf dma_addr %llx size %d\n", buf.addr, 580 (int)buf.size); 581 iovbase = cb->send_dma_addr; 582 cb->send_mr = ib_reg_phys_mr(cb->pd, &buf, 1, 583 0, &iovbase); 584 585 if (IS_ERR(cb->send_mr)) { 586 DEBUG_LOG(cb, "send_buf reg_mr failed\n"); 587 ret = PTR_ERR(cb->send_mr); 588 goto bail; 589 } 590 } 591 } 592 593 cb->rdma_buf = kmalloc(cb->size, GFP_KERNEL); 594 if (!cb->rdma_buf) { 595 DEBUG_LOG(cb, "rdma_buf malloc failed\n"); 596 ret = -ENOMEM; 597 goto bail; 598 } 599 600 cb->rdma_dma_addr = dma_map_single(cb->pd->device->dma_device, 601 cb->rdma_buf, cb->size, 602 DMA_BIDIRECTIONAL); 603 pci_unmap_addr_set(cb, rdma_mapping, cb->rdma_dma_addr); 604 if (cb->mem != DMA) { 605 switch (cb->mem) { 606 case FASTREG: 607 cb->page_list_len = (((cb->size - 1) & PAGE_MASK) + 608 PAGE_SIZE) >> PAGE_SHIFT; 609 cb->page_list = ib_alloc_fast_reg_page_list( 610 cb->pd->device, 611 cb->page_list_len); 612 if (IS_ERR(cb->page_list)) { 613 DEBUG_LOG(cb, "recv_buf reg_mr failed\n"); 614 ret = PTR_ERR(cb->page_list); 615 goto bail; 616 } 617 cb->fastreg_mr = ib_alloc_fast_reg_mr(cb->pd, 618 cb->page_list->max_page_list_len); 619 if (IS_ERR(cb->fastreg_mr)) { 620 DEBUG_LOG(cb, "recv_buf reg_mr failed\n"); 621 ret = PTR_ERR(cb->fastreg_mr); 622 goto bail; 623 } 624 DEBUG_LOG(cb, "fastreg rkey 0x%x page_list %p" 625 " page_list_len %u\n", cb->fastreg_mr->rkey, 626 cb->page_list, cb->page_list_len); 627 break; 628 case MW: 629 cb->mw = ib_alloc_mw(cb->pd); 630 if (IS_ERR(cb->mw)) { 631 DEBUG_LOG(cb, "recv_buf alloc_mw failed\n"); 632 ret = PTR_ERR(cb->mw); 633 goto bail; 634 } 635 DEBUG_LOG(cb, "mw rkey 0x%x\n", cb->mw->rkey); 636 /*FALLTHROUGH*/ 637 case MR: 638 buf.addr = cb->rdma_dma_addr; 639 buf.size = cb->size; 640 iovbase = cb->rdma_dma_addr; 641 cb->rdma_mr = ib_reg_phys_mr(cb->pd, &buf, 1, 642 IB_ACCESS_REMOTE_READ| 643 IB_ACCESS_REMOTE_WRITE, 644 &iovbase); 645 if (IS_ERR(cb->rdma_mr)) { 646 DEBUG_LOG(cb, "rdma_buf reg_mr failed\n"); 647 ret = PTR_ERR(cb->rdma_mr); 648 goto bail; 649 } 650 DEBUG_LOG(cb, "rdma buf dma_addr %llx size %d mr rkey 0x%x\n", 651 buf.addr, (int)buf.size, cb->rdma_mr->rkey); 652 break; 653 default: 654 ret = -EINVAL; 655 goto bail; 656 break; 657 } 658 } 659 660 if (!cb->server || cb->wlat || cb->rlat || cb->bw) { 661 662 cb->start_buf = kmalloc(cb->size, GFP_KERNEL); 663 if (!cb->start_buf) { 664 DEBUG_LOG(cb, "start_buf malloc failed\n"); 665 ret = -ENOMEM; 666 goto bail; 667 } 668 669 cb->start_dma_addr = dma_map_single(cb->pd->device->dma_device, 670 cb->start_buf, cb->size, 671 DMA_BIDIRECTIONAL); 672 pci_unmap_addr_set(cb, start_mapping, cb->start_dma_addr); 673 674 if (cb->mem == MR || cb->mem == MW) { 675 unsigned flags = IB_ACCESS_REMOTE_READ; 676 677 if (cb->wlat || cb->rlat || cb->bw) 678 flags |= IB_ACCESS_REMOTE_WRITE; 679 680 buf.addr = cb->start_dma_addr; 681 buf.size = cb->size; 682 DEBUG_LOG(cb, "start buf dma_addr %llx size %d\n", 683 buf.addr, (int)buf.size); 684 iovbase = cb->start_dma_addr; 685 cb->start_mr = ib_reg_phys_mr(cb->pd, &buf, 1, 686 flags, 687 &iovbase); 688 689 if (IS_ERR(cb->start_mr)) { 690 DEBUG_LOG(cb, "start_buf reg_mr failed\n"); 691 ret = PTR_ERR(cb->start_mr); 692 goto bail; 693 } 694 } 695 } 696 697 krping_setup_wr(cb); 698 DEBUG_LOG(cb, "allocated & registered buffers...\n"); 699 return 0; 700bail: 701 if (cb->fastreg_mr && !IS_ERR(cb->fastreg_mr)) 702 ib_dereg_mr(cb->fastreg_mr); 703 if (cb->mw && !IS_ERR(cb->mw)) 704 ib_dealloc_mw(cb->mw); 705 if (cb->rdma_mr && !IS_ERR(cb->rdma_mr)) 706 ib_dereg_mr(cb->rdma_mr); 707 if (cb->page_list && !IS_ERR(cb->page_list)) 708 ib_free_fast_reg_page_list(cb->page_list); 709 if (cb->dma_mr && !IS_ERR(cb->dma_mr)) 710 ib_dereg_mr(cb->dma_mr); 711 if (cb->recv_mr && !IS_ERR(cb->recv_mr)) 712 ib_dereg_mr(cb->recv_mr); 713 if (cb->send_mr && !IS_ERR(cb->send_mr)) 714 ib_dereg_mr(cb->send_mr); 715 if (cb->rdma_buf) 716 kfree(cb->rdma_buf); 717 if (cb->start_buf) 718 kfree(cb->start_buf); 719 return ret; 720} 721 722static void krping_free_buffers(struct krping_cb *cb) 723{ 724 DEBUG_LOG(cb, "krping_free_buffers called on cb %p\n", cb); 725 726 if (cb->dma_mr) 727 ib_dereg_mr(cb->dma_mr); 728 if (cb->send_mr) 729 ib_dereg_mr(cb->send_mr); 730 if (cb->recv_mr) 731 ib_dereg_mr(cb->recv_mr); 732 if (cb->rdma_mr) 733 ib_dereg_mr(cb->rdma_mr); 734 if (cb->start_mr) 735 ib_dereg_mr(cb->start_mr); 736 if (cb->fastreg_mr) 737 ib_dereg_mr(cb->fastreg_mr); 738 if (cb->mw) 739 ib_dealloc_mw(cb->mw); 740 741 dma_unmap_single(cb->pd->device->dma_device, 742 pci_unmap_addr(cb, recv_mapping), 743 sizeof(cb->recv_buf), DMA_BIDIRECTIONAL); 744 dma_unmap_single(cb->pd->device->dma_device, 745 pci_unmap_addr(cb, send_mapping), 746 sizeof(cb->send_buf), DMA_BIDIRECTIONAL); 747 dma_unmap_single(cb->pd->device->dma_device, 748 pci_unmap_addr(cb, rdma_mapping), 749 cb->size, DMA_BIDIRECTIONAL); 750 kfree(cb->rdma_buf); 751 if (cb->start_buf) { 752 dma_unmap_single(cb->pd->device->dma_device, 753 pci_unmap_addr(cb, start_mapping), 754 cb->size, DMA_BIDIRECTIONAL); 755 kfree(cb->start_buf); 756 } 757} 758 759static int krping_create_qp(struct krping_cb *cb) 760{ 761 struct ib_qp_init_attr init_attr; 762 int ret; 763 764 memset(&init_attr, 0, sizeof(init_attr)); 765 init_attr.cap.max_send_wr = cb->txdepth; 766 init_attr.cap.max_recv_wr = 2; 767 init_attr.cap.max_recv_sge = 1; 768 init_attr.cap.max_send_sge = 1; 769 init_attr.qp_type = IB_QPT_RC; 770 init_attr.send_cq = cb->cq; 771 init_attr.recv_cq = cb->cq; 772 init_attr.sq_sig_type = IB_SIGNAL_REQ_WR; 773 774 if (cb->server) { 775 ret = rdma_create_qp(cb->child_cm_id, cb->pd, &init_attr); 776 if (!ret) 777 cb->qp = cb->child_cm_id->qp; 778 } else { 779 ret = rdma_create_qp(cb->cm_id, cb->pd, &init_attr); 780 if (!ret) 781 cb->qp = cb->cm_id->qp; 782 } 783 784 return ret; 785} 786 787static void krping_free_qp(struct krping_cb *cb) 788{ 789 ib_destroy_qp(cb->qp); 790 ib_destroy_cq(cb->cq); 791 ib_dealloc_pd(cb->pd); 792} 793 794static int krping_setup_qp(struct krping_cb *cb, struct rdma_cm_id *cm_id) 795{ 796 int ret; 797 cb->pd = ib_alloc_pd(cm_id->device); 798 if (IS_ERR(cb->pd)) { 799 PRINTF(cb, "ib_alloc_pd failed\n"); 800 return PTR_ERR(cb->pd); 801 } 802 DEBUG_LOG(cb, "created pd %p\n", cb->pd); 803 804 strlcpy(cb->stats.name, cb->pd->device->name, sizeof(cb->stats.name)); 805 806 cb->cq = ib_create_cq(cm_id->device, krping_cq_event_handler, NULL, 807 cb, cb->txdepth * 2, 0); 808 if (IS_ERR(cb->cq)) { 809 PRINTF(cb, "ib_create_cq failed\n"); 810 ret = PTR_ERR(cb->cq); 811 goto err1; 812 } 813 DEBUG_LOG(cb, "created cq %p\n", cb->cq); 814 815 if (!cb->wlat && !cb->rlat && !cb->bw && !cb->frtest) { 816 ret = ib_req_notify_cq(cb->cq, IB_CQ_NEXT_COMP); 817 if (ret) { 818 PRINTF(cb, "ib_create_cq failed\n"); 819 goto err2; 820 } 821 } 822 823 ret = krping_create_qp(cb); 824 if (ret) { 825 PRINTF(cb, "krping_create_qp failed: %d\n", ret); 826 goto err2; 827 } 828 DEBUG_LOG(cb, "created qp %p\n", cb->qp); 829 return 0; 830err2: 831 ib_destroy_cq(cb->cq); 832err1: 833 ib_dealloc_pd(cb->pd); 834 return ret; 835} 836 837/* 838 * return the (possibly rebound) rkey for the rdma buffer. 839 * FASTREG mode: invalidate and rebind via fastreg wr. 840 * MW mode: rebind the MW. 841 * other modes: just return the mr rkey. 842 */ 843static u32 krping_rdma_rkey(struct krping_cb *cb, u64 buf, int post_inv) 844{ 845 u32 rkey = 0xffffffff; 846 u64 p; 847 struct ib_send_wr *bad_wr; 848 int i; 849 int ret; 850 851 switch (cb->mem) { 852 case FASTREG: 853 cb->invalidate_wr.ex.invalidate_rkey = cb->fastreg_mr->rkey; 854 855 /* 856 * Update the fastreg key. 857 */ 858 ib_update_fast_reg_key(cb->fastreg_mr, ++cb->key); 859 cb->fastreg_wr.wr.fast_reg.rkey = cb->fastreg_mr->rkey; 860 861 /* 862 * Update the fastreg WR with new buf info. 863 */ 864 if (buf == (u64)cb->start_dma_addr) 865 cb->fastreg_wr.wr.fast_reg.access_flags = IB_ACCESS_REMOTE_READ; 866 else 867 cb->fastreg_wr.wr.fast_reg.access_flags = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE; 868 cb->fastreg_wr.wr.fast_reg.iova_start = buf; 869 p = (u64)(buf & PAGE_MASK); 870 for (i=0; i < cb->fastreg_wr.wr.fast_reg.page_list_len; 871 i++, p += PAGE_SIZE) { 872 cb->page_list->page_list[i] = p; 873 DEBUG_LOG(cb, "page_list[%d] 0x%llx\n", i, p); 874 } 875 876 DEBUG_LOG(cb, "post_inv = %d, fastreg new rkey 0x%x shift %u len %u" 877 " iova_start %llx page_list_len %u\n", 878 post_inv, 879 cb->fastreg_wr.wr.fast_reg.rkey, 880 cb->fastreg_wr.wr.fast_reg.page_shift, 881 cb->fastreg_wr.wr.fast_reg.length, 882 cb->fastreg_wr.wr.fast_reg.iova_start, 883 cb->fastreg_wr.wr.fast_reg.page_list_len); 884 885 if (post_inv) 886 ret = ib_post_send(cb->qp, &cb->invalidate_wr, &bad_wr); 887 else 888 ret = ib_post_send(cb->qp, &cb->fastreg_wr, &bad_wr); 889 if (ret) { 890 PRINTF(cb, "post send error %d\n", ret); 891 cb->state = ERROR; 892 } 893 rkey = cb->fastreg_mr->rkey; 894 break; 895 case MW: 896 /* 897 * Update the MW with new buf info. 898 */ 899 if (buf == (u64)cb->start_dma_addr) { 900 cb->bind_attr.mw_access_flags = IB_ACCESS_REMOTE_READ; 901 cb->bind_attr.mr = cb->start_mr; 902 } else { 903 cb->bind_attr.mw_access_flags = IB_ACCESS_REMOTE_WRITE; 904 cb->bind_attr.mr = cb->rdma_mr; 905 } 906 cb->bind_attr.addr = buf; 907 DEBUG_LOG(cb, "binding mw rkey 0x%x to buf %llx mr rkey 0x%x\n", 908 cb->mw->rkey, buf, cb->bind_attr.mr->rkey); 909 ret = ib_bind_mw(cb->qp, cb->mw, &cb->bind_attr); 910 if (ret) { 911 PRINTF(cb, "bind mw error %d\n", ret); 912 cb->state = ERROR; 913 } else 914 rkey = cb->mw->rkey; 915 break; 916 case MR: 917 if (buf == (u64)cb->start_dma_addr) 918 rkey = cb->start_mr->rkey; 919 else 920 rkey = cb->rdma_mr->rkey; 921 break; 922 case DMA: 923 rkey = cb->dma_mr->rkey; 924 break; 925 default: 926 PRINTF(cb, "%s:%d case ERROR\n", __func__, __LINE__); 927 cb->state = ERROR; 928 break; 929 } 930 return rkey; 931} 932 933static void krping_format_send(struct krping_cb *cb, u64 buf) 934{ 935 struct krping_rdma_info *info = &cb->send_buf; 936 u32 rkey; 937 938 /* 939 * Client side will do fastreg or mw bind before 940 * advertising the rdma buffer. Server side 941 * sends have no data. 942 */ 943 if (!cb->server || cb->wlat || cb->rlat || cb->bw) { 944 rkey = krping_rdma_rkey(cb, buf, !cb->server_invalidate); 945 info->buf = htonll(buf); 946 info->rkey = htonl(rkey); 947 info->size = htonl(cb->size); 948 DEBUG_LOG(cb, "RDMA addr %llx rkey %x len %d\n", 949 (unsigned long long)buf, rkey, cb->size); 950 } 951} 952 953static void krping_test_server(struct krping_cb *cb) 954{ 955 struct ib_send_wr *bad_wr, inv; 956 int ret; 957 958 while (1) { 959 /* Wait for client's Start STAG/TO/Len */ 960 wait_event_interruptible(cb->sem, cb->state >= RDMA_READ_ADV); 961 if (cb->state != RDMA_READ_ADV) { 962 PRINTF(cb, "wait for RDMA_READ_ADV state %d\n", 963 cb->state); 964 break; 965 } 966 967 DEBUG_LOG(cb, "server received sink adv\n"); 968 969 cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey; 970 cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr; 971 cb->rdma_sq_wr.sg_list->length = cb->remote_len; 972 cb->rdma_sgl.lkey = krping_rdma_rkey(cb, cb->rdma_dma_addr, 1); 973 cb->rdma_sq_wr.next = NULL; 974 975 /* Issue RDMA Read. */ 976 if (cb->read_inv) 977 cb->rdma_sq_wr.opcode = IB_WR_RDMA_READ_WITH_INV; 978 else { 979 980 cb->rdma_sq_wr.opcode = IB_WR_RDMA_READ; 981 if (cb->mem == FASTREG) { 982 /* 983 * Immediately follow the read with a 984 * fenced LOCAL_INV. 985 */ 986 cb->rdma_sq_wr.next = &inv; 987 memset(&inv, 0, sizeof inv); 988 inv.opcode = IB_WR_LOCAL_INV; 989 inv.ex.invalidate_rkey = cb->fastreg_mr->rkey; 990 inv.send_flags = IB_SEND_FENCE; 991 } 992 } 993 994 ret = ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr); 995 if (ret) { 996 PRINTF(cb, "post send error %d\n", ret); 997 break; 998 } 999 cb->rdma_sq_wr.next = NULL; 1000 1001 DEBUG_LOG(cb, "server posted rdma read req \n"); 1002 1003 /* Wait for read completion */ 1004 wait_event_interruptible(cb->sem, 1005 cb->state >= RDMA_READ_COMPLETE); 1006 if (cb->state != RDMA_READ_COMPLETE) { 1007 PRINTF(cb, 1008 "wait for RDMA_READ_COMPLETE state %d\n", 1009 cb->state); 1010 break; 1011 } 1012 DEBUG_LOG(cb, "server received read complete\n"); 1013 1014 /* Display data in recv buf */ 1015 if (cb->verbose) 1016 PRINTF(cb, "server ping data: %s\n", 1017 cb->rdma_buf); 1018 1019 /* Tell client to continue */ 1020 if (cb->server && cb->server_invalidate) { 1021 cb->sq_wr.ex.invalidate_rkey = cb->remote_rkey; 1022 cb->sq_wr.opcode = IB_WR_SEND_WITH_INV; 1023 DEBUG_LOG(cb, "send-w-inv rkey 0x%x\n", cb->remote_rkey); 1024 } 1025 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1026 if (ret) { 1027 PRINTF(cb, "post send error %d\n", ret); 1028 break; 1029 } 1030 DEBUG_LOG(cb, "server posted go ahead\n"); 1031 1032 /* Wait for client's RDMA STAG/TO/Len */ 1033 wait_event_interruptible(cb->sem, cb->state >= RDMA_WRITE_ADV); 1034 if (cb->state != RDMA_WRITE_ADV) { 1035 PRINTF(cb, 1036 "wait for RDMA_WRITE_ADV state %d\n", 1037 cb->state); 1038 break; 1039 } 1040 DEBUG_LOG(cb, "server received sink adv\n"); 1041 1042 /* RDMA Write echo data */ 1043 cb->rdma_sq_wr.opcode = IB_WR_RDMA_WRITE; 1044 cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey; 1045 cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr; 1046 cb->rdma_sq_wr.sg_list->length = strlen(cb->rdma_buf) + 1; 1047 if (cb->local_dma_lkey) 1048 cb->rdma_sgl.lkey = cb->qp->device->local_dma_lkey; 1049 else 1050 cb->rdma_sgl.lkey = krping_rdma_rkey(cb, cb->rdma_dma_addr, 0); 1051 1052 DEBUG_LOG(cb, "rdma write from lkey %x laddr %llx len %d\n", 1053 cb->rdma_sq_wr.sg_list->lkey, 1054 (unsigned long long)cb->rdma_sq_wr.sg_list->addr, 1055 cb->rdma_sq_wr.sg_list->length); 1056 1057 ret = ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr); 1058 if (ret) { 1059 PRINTF(cb, "post send error %d\n", ret); 1060 break; 1061 } 1062 1063 /* Wait for completion */ 1064 ret = wait_event_interruptible(cb->sem, cb->state >= 1065 RDMA_WRITE_COMPLETE); 1066 if (cb->state != RDMA_WRITE_COMPLETE) { 1067 PRINTF(cb, 1068 "wait for RDMA_WRITE_COMPLETE state %d\n", 1069 cb->state); 1070 break; 1071 } 1072 DEBUG_LOG(cb, "server rdma write complete \n"); 1073 1074 cb->state = CONNECTED; 1075 1076 /* Tell client to begin again */ 1077 if (cb->server && cb->server_invalidate) { 1078 cb->sq_wr.ex.invalidate_rkey = cb->remote_rkey; 1079 cb->sq_wr.opcode = IB_WR_SEND_WITH_INV; 1080 DEBUG_LOG(cb, "send-w-inv rkey 0x%x\n", cb->remote_rkey); 1081 } 1082 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1083 if (ret) { 1084 PRINTF(cb, "post send error %d\n", ret); 1085 break; 1086 } 1087 DEBUG_LOG(cb, "server posted go ahead\n"); 1088 } 1089} 1090 1091static void rlat_test(struct krping_cb *cb) 1092{ 1093 int scnt; 1094 int iters = cb->count; 1095 struct timeval start_tv, stop_tv; 1096 int ret; 1097 struct ib_wc wc; 1098 struct ib_send_wr *bad_wr; 1099 int ne; 1100 1101 scnt = 0; 1102 cb->rdma_sq_wr.opcode = IB_WR_RDMA_READ; 1103 cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey; 1104 cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr; 1105 cb->rdma_sq_wr.sg_list->length = cb->size; 1106 1107 microtime(&start_tv); 1108 if (!cb->poll) { 1109 cb->state = RDMA_READ_ADV; 1110 ib_req_notify_cq(cb->cq, IB_CQ_NEXT_COMP); 1111 } 1112 while (scnt < iters) { 1113 1114 cb->state = RDMA_READ_ADV; 1115 ret = ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr); 1116 if (ret) { 1117 PRINTF(cb, 1118 "Couldn't post send: ret=%d scnt %d\n", 1119 ret, scnt); 1120 return; 1121 } 1122 1123 do { 1124 if (!cb->poll) { 1125 wait_event_interruptible(cb->sem, 1126 cb->state != RDMA_READ_ADV); 1127 if (cb->state == RDMA_READ_COMPLETE) { 1128 ne = 1; 1129 ib_req_notify_cq(cb->cq, 1130 IB_CQ_NEXT_COMP); 1131 } else { 1132 ne = -1; 1133 } 1134 } else 1135 ne = ib_poll_cq(cb->cq, 1, &wc); 1136 if (cb->state == ERROR) { 1137 PRINTF(cb, 1138 "state == ERROR...bailing scnt %d\n", 1139 scnt); 1140 return; 1141 } 1142 } while (ne == 0); 1143 1144 if (ne < 0) { 1145 PRINTF(cb, "poll CQ failed %d\n", ne); 1146 return; 1147 } 1148 if (cb->poll && wc.status != IB_WC_SUCCESS) { 1149 PRINTF(cb, "Completion wth error at %s:\n", 1150 cb->server ? "server" : "client"); 1151 PRINTF(cb, "Failed status %d: wr_id %d\n", 1152 wc.status, (int) wc.wr_id); 1153 return; 1154 } 1155 ++scnt; 1156 } 1157 microtime(&stop_tv); 1158 1159 if (stop_tv.tv_usec < start_tv.tv_usec) { 1160 stop_tv.tv_usec += 1000000; 1161 stop_tv.tv_sec -= 1; 1162 } 1163 1164 PRINTF(cb, "delta sec %lu delta usec %lu iter %d size %d\n", 1165 stop_tv.tv_sec - start_tv.tv_sec, 1166 stop_tv.tv_usec - start_tv.tv_usec, 1167 scnt, cb->size); 1168} 1169 1170static void wlat_test(struct krping_cb *cb) 1171{ 1172 int ccnt, scnt, rcnt; 1173 int iters=cb->count; 1174 volatile char *poll_buf = (char *) cb->start_buf; 1175 char *buf = (char *)cb->rdma_buf; 1176 struct timeval start_tv, stop_tv; 1177 cycles_t *post_cycles_start, *post_cycles_stop; 1178 cycles_t *poll_cycles_start, *poll_cycles_stop; 1179 cycles_t *last_poll_cycles_start; 1180 cycles_t sum_poll = 0, sum_post = 0, sum_last_poll = 0; 1181 int i; 1182 int cycle_iters = 1000; 1183 1184 ccnt = 0; 1185 scnt = 0; 1186 rcnt = 0; 1187 1188 post_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL); 1189 if (!post_cycles_start) { 1190 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1191 return; 1192 } 1193 post_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL); 1194 if (!post_cycles_stop) { 1195 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1196 return; 1197 } 1198 poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL); 1199 if (!poll_cycles_start) { 1200 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1201 return; 1202 } 1203 poll_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL); 1204 if (!poll_cycles_stop) { 1205 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1206 return; 1207 } 1208 last_poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), 1209 GFP_KERNEL); 1210 if (!last_poll_cycles_start) { 1211 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1212 return; 1213 } 1214 cb->rdma_sq_wr.opcode = IB_WR_RDMA_WRITE; 1215 cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey; 1216 cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr; 1217 cb->rdma_sq_wr.sg_list->length = cb->size; 1218 1219 if (cycle_iters > iters) 1220 cycle_iters = iters; 1221 microtime(&start_tv); 1222 while (scnt < iters || ccnt < iters || rcnt < iters) { 1223 1224 /* Wait till buffer changes. */ 1225 if (rcnt < iters && !(scnt < 1 && !cb->server)) { 1226 ++rcnt; 1227 while (*poll_buf != (char)rcnt) { 1228 if (cb->state == ERROR) { 1229 PRINTF(cb, 1230 "state = ERROR, bailing\n"); 1231 return; 1232 } 1233 } 1234 } 1235 1236 if (scnt < iters) { 1237 struct ib_send_wr *bad_wr; 1238 1239 *buf = (char)scnt+1; 1240 if (scnt < cycle_iters) 1241 post_cycles_start[scnt] = get_cycles(); 1242 if (ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr)) { 1243 PRINTF(cb, 1244 "Couldn't post send: scnt=%d\n", 1245 scnt); 1246 return; 1247 } 1248 if (scnt < cycle_iters) 1249 post_cycles_stop[scnt] = get_cycles(); 1250 scnt++; 1251 } 1252 1253 if (ccnt < iters) { 1254 struct ib_wc wc; 1255 int ne; 1256 1257 if (ccnt < cycle_iters) 1258 poll_cycles_start[ccnt] = get_cycles(); 1259 do { 1260 if (ccnt < cycle_iters) 1261 last_poll_cycles_start[ccnt] = 1262 get_cycles(); 1263 ne = ib_poll_cq(cb->cq, 1, &wc); 1264 } while (ne == 0); 1265 if (ccnt < cycle_iters) 1266 poll_cycles_stop[ccnt] = get_cycles(); 1267 ++ccnt; 1268 1269 if (ne < 0) { 1270 PRINTF(cb, "poll CQ failed %d\n", ne); 1271 return; 1272 } 1273 if (wc.status != IB_WC_SUCCESS) { 1274 PRINTF(cb, 1275 "Completion wth error at %s:\n", 1276 cb->server ? "server" : "client"); 1277 PRINTF(cb, 1278 "Failed status %d: wr_id %d\n", 1279 wc.status, (int) wc.wr_id); 1280 PRINTF(cb, 1281 "scnt=%d, rcnt=%d, ccnt=%d\n", 1282 scnt, rcnt, ccnt); 1283 return; 1284 } 1285 } 1286 } 1287 microtime(&stop_tv); 1288 1289 if (stop_tv.tv_usec < start_tv.tv_usec) { 1290 stop_tv.tv_usec += 1000000; 1291 stop_tv.tv_sec -= 1; 1292 } 1293 1294 for (i=0; i < cycle_iters; i++) { 1295 sum_post += post_cycles_stop[i] - post_cycles_start[i]; 1296 sum_poll += poll_cycles_stop[i] - poll_cycles_start[i]; 1297 sum_last_poll += poll_cycles_stop[i]-last_poll_cycles_start[i]; 1298 } 1299 PRINTF(cb, 1300 "delta sec %lu delta usec %lu iter %d size %d cycle_iters %d" 1301 " sum_post %llu sum_poll %llu sum_last_poll %llu\n", 1302 stop_tv.tv_sec - start_tv.tv_sec, 1303 stop_tv.tv_usec - start_tv.tv_usec, 1304 scnt, cb->size, cycle_iters, 1305 (unsigned long long)sum_post, (unsigned long long)sum_poll, 1306 (unsigned long long)sum_last_poll); 1307 kfree(post_cycles_start); 1308 kfree(post_cycles_stop); 1309 kfree(poll_cycles_start); 1310 kfree(poll_cycles_stop); 1311 kfree(last_poll_cycles_start); 1312} 1313 1314static void bw_test(struct krping_cb *cb) 1315{ 1316 int ccnt, scnt, rcnt; 1317 int iters=cb->count; 1318 struct timeval start_tv, stop_tv; 1319 cycles_t *post_cycles_start, *post_cycles_stop; 1320 cycles_t *poll_cycles_start, *poll_cycles_stop; 1321 cycles_t *last_poll_cycles_start; 1322 cycles_t sum_poll = 0, sum_post = 0, sum_last_poll = 0; 1323 int i; 1324 int cycle_iters = 1000; 1325 1326 ccnt = 0; 1327 scnt = 0; 1328 rcnt = 0; 1329 1330 post_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL); 1331 if (!post_cycles_start) { 1332 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1333 return; 1334 } 1335 post_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL); 1336 if (!post_cycles_stop) { 1337 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1338 return; 1339 } 1340 poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL); 1341 if (!poll_cycles_start) { 1342 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1343 return; 1344 } 1345 poll_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL); 1346 if (!poll_cycles_stop) { 1347 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1348 return; 1349 } 1350 last_poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), 1351 GFP_KERNEL); 1352 if (!last_poll_cycles_start) { 1353 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1354 return; 1355 } 1356 cb->rdma_sq_wr.opcode = IB_WR_RDMA_WRITE; 1357 cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey; 1358 cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr; 1359 cb->rdma_sq_wr.sg_list->length = cb->size; 1360 1361 if (cycle_iters > iters) 1362 cycle_iters = iters; 1363 microtime(&start_tv); 1364 while (scnt < iters || ccnt < iters) { 1365 1366 while (scnt < iters && scnt - ccnt < cb->txdepth) { 1367 struct ib_send_wr *bad_wr; 1368 1369 if (scnt < cycle_iters) 1370 post_cycles_start[scnt] = get_cycles(); 1371 if (ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr)) { 1372 PRINTF(cb, 1373 "Couldn't post send: scnt=%d\n", 1374 scnt); 1375 return; 1376 } 1377 if (scnt < cycle_iters) 1378 post_cycles_stop[scnt] = get_cycles(); 1379 ++scnt; 1380 } 1381 1382 if (ccnt < iters) { 1383 int ne; 1384 struct ib_wc wc; 1385 1386 if (ccnt < cycle_iters) 1387 poll_cycles_start[ccnt] = get_cycles(); 1388 do { 1389 if (ccnt < cycle_iters) 1390 last_poll_cycles_start[ccnt] = 1391 get_cycles(); 1392 ne = ib_poll_cq(cb->cq, 1, &wc); 1393 } while (ne == 0); 1394 if (ccnt < cycle_iters) 1395 poll_cycles_stop[ccnt] = get_cycles(); 1396 ccnt += 1; 1397 1398 if (ne < 0) { 1399 PRINTF(cb, "poll CQ failed %d\n", ne); 1400 return; 1401 } 1402 if (wc.status != IB_WC_SUCCESS) { 1403 PRINTF(cb, 1404 "Completion wth error at %s:\n", 1405 cb->server ? "server" : "client"); 1406 PRINTF(cb, 1407 "Failed status %d: wr_id %d\n", 1408 wc.status, (int) wc.wr_id); 1409 return; 1410 } 1411 } 1412 } 1413 microtime(&stop_tv); 1414 1415 if (stop_tv.tv_usec < start_tv.tv_usec) { 1416 stop_tv.tv_usec += 1000000; 1417 stop_tv.tv_sec -= 1; 1418 } 1419 1420 for (i=0; i < cycle_iters; i++) { 1421 sum_post += post_cycles_stop[i] - post_cycles_start[i]; 1422 sum_poll += poll_cycles_stop[i] - poll_cycles_start[i]; 1423 sum_last_poll += poll_cycles_stop[i]-last_poll_cycles_start[i]; 1424 } 1425 PRINTF(cb, 1426 "delta sec %lu delta usec %lu iter %d size %d cycle_iters %d" 1427 " sum_post %llu sum_poll %llu sum_last_poll %llu\n", 1428 stop_tv.tv_sec - start_tv.tv_sec, 1429 stop_tv.tv_usec - start_tv.tv_usec, 1430 scnt, cb->size, cycle_iters, 1431 (unsigned long long)sum_post, (unsigned long long)sum_poll, 1432 (unsigned long long)sum_last_poll); 1433 kfree(post_cycles_start); 1434 kfree(post_cycles_stop); 1435 kfree(poll_cycles_start); 1436 kfree(poll_cycles_stop); 1437 kfree(last_poll_cycles_start); 1438} 1439 1440static void krping_rlat_test_server(struct krping_cb *cb) 1441{ 1442 struct ib_send_wr *bad_wr; 1443 struct ib_wc wc; 1444 int ret; 1445 1446 /* Spin waiting for client's Start STAG/TO/Len */ 1447 while (cb->state < RDMA_READ_ADV) { 1448 krping_cq_event_handler(cb->cq, cb); 1449 } 1450 1451 /* Send STAG/TO/Len to client */ 1452 krping_format_send(cb, cb->start_dma_addr); 1453 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1454 if (ret) { 1455 PRINTF(cb, "post send error %d\n", ret); 1456 return; 1457 } 1458 1459 /* Spin waiting for send completion */ 1460 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0)); 1461 if (ret < 0) { 1462 PRINTF(cb, "poll error %d\n", ret); 1463 return; 1464 } 1465 if (wc.status) { 1466 PRINTF(cb, "send completiong error %d\n", wc.status); 1467 return; 1468 } 1469 1470 wait_event_interruptible(cb->sem, cb->state == ERROR); 1471} 1472 1473static void krping_wlat_test_server(struct krping_cb *cb) 1474{ 1475 struct ib_send_wr *bad_wr; 1476 struct ib_wc wc; 1477 int ret; 1478 1479 /* Spin waiting for client's Start STAG/TO/Len */ 1480 while (cb->state < RDMA_READ_ADV) { 1481 krping_cq_event_handler(cb->cq, cb); 1482 } 1483 1484 /* Send STAG/TO/Len to client */ 1485 krping_format_send(cb, cb->start_dma_addr); 1486 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1487 if (ret) { 1488 PRINTF(cb, "post send error %d\n", ret); 1489 return; 1490 } 1491 1492 /* Spin waiting for send completion */ 1493 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0)); 1494 if (ret < 0) { 1495 PRINTF(cb, "poll error %d\n", ret); 1496 return; 1497 } 1498 if (wc.status) { 1499 PRINTF(cb, "send completiong error %d\n", wc.status); 1500 return; 1501 } 1502 1503 wlat_test(cb); 1504 wait_event_interruptible(cb->sem, cb->state == ERROR); 1505} 1506 1507static void krping_bw_test_server(struct krping_cb *cb) 1508{ 1509 struct ib_send_wr *bad_wr; 1510 struct ib_wc wc; 1511 int ret; 1512 1513 /* Spin waiting for client's Start STAG/TO/Len */ 1514 while (cb->state < RDMA_READ_ADV) { 1515 krping_cq_event_handler(cb->cq, cb); 1516 } 1517 1518 /* Send STAG/TO/Len to client */ 1519 krping_format_send(cb, cb->start_dma_addr); 1520 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1521 if (ret) { 1522 PRINTF(cb, "post send error %d\n", ret); 1523 return; 1524 } 1525 1526 /* Spin waiting for send completion */ 1527 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0)); 1528 if (ret < 0) { 1529 PRINTF(cb, "poll error %d\n", ret); 1530 return; 1531 } 1532 if (wc.status) { 1533 PRINTF(cb, "send completiong error %d\n", wc.status); 1534 return; 1535 } 1536 1537 if (cb->duplex) 1538 bw_test(cb); 1539 wait_event_interruptible(cb->sem, cb->state == ERROR); 1540} 1541 1542static int fastreg_supported(struct krping_cb *cb) 1543{ 1544 struct ib_device *dev = cb->child_cm_id->device; 1545 struct ib_device_attr attr; 1546 int ret; 1547 1548 ret = ib_query_device(dev, &attr); 1549 if (ret) { 1550 PRINTF(cb, "ib_query_device failed ret %d\n", ret); 1551 return 0; 1552 } 1553 if (!(attr.device_cap_flags & IB_DEVICE_MEM_MGT_EXTENSIONS)) { 1554 PRINTF(cb, "Fastreg not supported - device_cap_flags 0x%x\n", 1555 attr.device_cap_flags); 1556 return 0; 1557 } 1558 DEBUG_LOG(cb, "Fastreg supported - device_cap_flags 0x%x\n", 1559 attr.device_cap_flags); 1560 return 1; 1561} 1562 1563static int krping_bind_server(struct krping_cb *cb) 1564{ 1565 struct sockaddr_in sin; 1566 int ret; 1567 1568 memset(&sin, 0, sizeof(sin)); 1569 sin.sin_len = sizeof sin; 1570 sin.sin_family = AF_INET; 1571 sin.sin_addr.s_addr = cb->addr.s_addr; 1572 sin.sin_port = cb->port; 1573 1574 ret = rdma_bind_addr(cb->cm_id, (struct sockaddr *) &sin); 1575 if (ret) { 1576 PRINTF(cb, "rdma_bind_addr error %d\n", ret); 1577 return ret; 1578 } 1579 DEBUG_LOG(cb, "rdma_bind_addr successful\n"); 1580 1581 DEBUG_LOG(cb, "rdma_listen\n"); 1582 ret = rdma_listen(cb->cm_id, 3); 1583 if (ret) { 1584 PRINTF(cb, "rdma_listen failed: %d\n", ret); 1585 return ret; 1586 } 1587 1588 wait_event_interruptible(cb->sem, cb->state >= CONNECT_REQUEST); 1589 if (cb->state != CONNECT_REQUEST) { 1590 PRINTF(cb, "wait for CONNECT_REQUEST state %d\n", 1591 cb->state); 1592 return -1; 1593 } 1594 1595 if (cb->mem == FASTREG && !fastreg_supported(cb)) 1596 return -EINVAL; 1597 1598 return 0; 1599} 1600 1601static void krping_run_server(struct krping_cb *cb) 1602{ 1603 struct ib_recv_wr *bad_wr; 1604 int ret; 1605 1606 ret = krping_bind_server(cb); 1607 if (ret) 1608 return; 1609 1610 ret = krping_setup_qp(cb, cb->child_cm_id); 1611 if (ret) { 1612 PRINTF(cb, "setup_qp failed: %d\n", ret); 1613 goto err0; 1614 } 1615 1616 ret = krping_setup_buffers(cb); 1617 if (ret) { 1618 PRINTF(cb, "krping_setup_buffers failed: %d\n", ret); 1619 goto err1; 1620 } 1621 1622 ret = ib_post_recv(cb->qp, &cb->rq_wr, &bad_wr); 1623 if (ret) { 1624 PRINTF(cb, "ib_post_recv failed: %d\n", ret); 1625 goto err2; 1626 } 1627 1628 ret = krping_accept(cb); 1629 if (ret) { 1630 PRINTF(cb, "connect error %d\n", ret); 1631 goto err2; 1632 } 1633 1634 if (cb->wlat) 1635 krping_wlat_test_server(cb); 1636 else if (cb->rlat) 1637 krping_rlat_test_server(cb); 1638 else if (cb->bw) 1639 krping_bw_test_server(cb); 1640 else 1641 krping_test_server(cb); 1642 rdma_disconnect(cb->child_cm_id); 1643err2: 1644 krping_free_buffers(cb); 1645err1: 1646 krping_free_qp(cb); 1647err0: 1648 rdma_destroy_id(cb->child_cm_id); 1649} 1650 1651static void krping_test_client(struct krping_cb *cb) 1652{ 1653 int ping, start, cc, i, ret; 1654 struct ib_send_wr *bad_wr; 1655 unsigned char c; 1656 1657 start = 65; 1658 for (ping = 0; !cb->count || ping < cb->count; ping++) { 1659 cb->state = RDMA_READ_ADV; 1660 1661 /* Put some ascii text in the buffer. */ 1662 cc = sprintf(cb->start_buf, "rdma-ping-%d: ", ping); 1663 for (i = cc, c = start; i < cb->size; i++) { 1664 cb->start_buf[i] = c; 1665 c++; 1666 if (c > 122) 1667 c = 65; 1668 } 1669 start++; 1670 if (start > 122) 1671 start = 65; 1672 cb->start_buf[cb->size - 1] = 0; 1673 1674 krping_format_send(cb, cb->start_dma_addr); 1675 if (cb->state == ERROR) { 1676 PRINTF(cb, "krping_format_send failed\n"); 1677 break; 1678 } 1679 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1680 if (ret) { 1681 PRINTF(cb, "post send error %d\n", ret); 1682 break; 1683 } 1684 1685 /* Wait for server to ACK */ 1686 wait_event_interruptible(cb->sem, cb->state >= RDMA_WRITE_ADV); 1687 if (cb->state != RDMA_WRITE_ADV) { 1688 PRINTF(cb, 1689 "wait for RDMA_WRITE_ADV state %d\n", 1690 cb->state); 1691 break; 1692 } 1693 1694 krping_format_send(cb, cb->rdma_dma_addr); 1695 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1696 if (ret) { 1697 PRINTF(cb, "post send error %d\n", ret); 1698 break; 1699 } 1700 1701 /* Wait for the server to say the RDMA Write is complete. */ 1702 wait_event_interruptible(cb->sem, 1703 cb->state >= RDMA_WRITE_COMPLETE); 1704 if (cb->state != RDMA_WRITE_COMPLETE) { 1705 PRINTF(cb, 1706 "wait for RDMA_WRITE_COMPLETE state %d\n", 1707 cb->state); 1708 break; 1709 } 1710 1711 if (cb->validate) 1712 if (memcmp(cb->start_buf, cb->rdma_buf, cb->size)) { 1713 PRINTF(cb, "data mismatch!\n"); 1714 break; 1715 } 1716 1717 if (cb->verbose) 1718 PRINTF(cb, "ping data: %s\n", cb->rdma_buf); 1719#ifdef SLOW_KRPING 1720 wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ); 1721#endif 1722 } 1723} 1724 1725static void krping_rlat_test_client(struct krping_cb *cb) 1726{ 1727 struct ib_send_wr *bad_wr; 1728 struct ib_wc wc; 1729 int ret; 1730 1731 cb->state = RDMA_READ_ADV; 1732 1733 /* Send STAG/TO/Len to client */ 1734 krping_format_send(cb, cb->start_dma_addr); 1735 if (cb->state == ERROR) { 1736 PRINTF(cb, "krping_format_send failed\n"); 1737 return; 1738 } 1739 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1740 if (ret) { 1741 PRINTF(cb, "post send error %d\n", ret); 1742 return; 1743 } 1744 1745 /* Spin waiting for send completion */ 1746 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0)); 1747 if (ret < 0) { 1748 PRINTF(cb, "poll error %d\n", ret); 1749 return; 1750 } 1751 if (wc.status) { 1752 PRINTF(cb, "send completion error %d\n", wc.status); 1753 return; 1754 } 1755 1756 /* Spin waiting for server's Start STAG/TO/Len */ 1757 while (cb->state < RDMA_WRITE_ADV) { 1758 krping_cq_event_handler(cb->cq, cb); 1759 } 1760 1761#if 0 1762{ 1763 int i; 1764 struct timeval start, stop; 1765 time_t sec; 1766 suseconds_t usec; 1767 unsigned long long elapsed; 1768 struct ib_wc wc; 1769 struct ib_send_wr *bad_wr; 1770 int ne; 1771 1772 cb->rdma_sq_wr.opcode = IB_WR_RDMA_WRITE; 1773 cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey; 1774 cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr; 1775 cb->rdma_sq_wr.sg_list->length = 0; 1776 cb->rdma_sq_wr.num_sge = 0; 1777 1778 microtime(&start); 1779 for (i=0; i < 100000; i++) { 1780 if (ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr)) { 1781 PRINTF(cb, "Couldn't post send\n"); 1782 return; 1783 } 1784 do { 1785 ne = ib_poll_cq(cb->cq, 1, &wc); 1786 } while (ne == 0); 1787 if (ne < 0) { 1788 PRINTF(cb, "poll CQ failed %d\n", ne); 1789 return; 1790 } 1791 if (wc.status != IB_WC_SUCCESS) { 1792 PRINTF(cb, "Completion wth error at %s:\n", 1793 cb->server ? "server" : "client"); 1794 PRINTF(cb, "Failed status %d: wr_id %d\n", 1795 wc.status, (int) wc.wr_id); 1796 return; 1797 } 1798 } 1799 microtime(&stop); 1800 1801 if (stop.tv_usec < start.tv_usec) { 1802 stop.tv_usec += 1000000; 1803 stop.tv_sec -= 1; 1804 } 1805 sec = stop.tv_sec - start.tv_sec; 1806 usec = stop.tv_usec - start.tv_usec; 1807 elapsed = sec * 1000000 + usec; 1808 PRINTF(cb, "0B-write-lat iters 100000 usec %llu\n", elapsed); 1809} 1810#endif 1811 1812 rlat_test(cb); 1813} 1814 1815static void krping_wlat_test_client(struct krping_cb *cb) 1816{ 1817 struct ib_send_wr *bad_wr; 1818 struct ib_wc wc; 1819 int ret; 1820 1821 cb->state = RDMA_READ_ADV; 1822 1823 /* Send STAG/TO/Len to client */ 1824 krping_format_send(cb, cb->start_dma_addr); 1825 if (cb->state == ERROR) { 1826 PRINTF(cb, "krping_format_send failed\n"); 1827 return; 1828 } 1829 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1830 if (ret) { 1831 PRINTF(cb, "post send error %d\n", ret); 1832 return; 1833 } 1834 1835 /* Spin waiting for send completion */ 1836 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0)); 1837 if (ret < 0) { 1838 PRINTF(cb, "poll error %d\n", ret); 1839 return; 1840 } 1841 if (wc.status) { 1842 PRINTF(cb, "send completion error %d\n", wc.status); 1843 return; 1844 } 1845 1846 /* Spin waiting for server's Start STAG/TO/Len */ 1847 while (cb->state < RDMA_WRITE_ADV) { 1848 krping_cq_event_handler(cb->cq, cb); 1849 } 1850 1851 wlat_test(cb); 1852} 1853 1854static void krping_bw_test_client(struct krping_cb *cb) 1855{ 1856 struct ib_send_wr *bad_wr; 1857 struct ib_wc wc; 1858 int ret; 1859 1860 cb->state = RDMA_READ_ADV; 1861 1862 /* Send STAG/TO/Len to client */ 1863 krping_format_send(cb, cb->start_dma_addr); 1864 if (cb->state == ERROR) { 1865 PRINTF(cb, "krping_format_send failed\n"); 1866 return; 1867 } 1868 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1869 if (ret) { 1870 PRINTF(cb, "post send error %d\n", ret); 1871 return; 1872 } 1873 1874 /* Spin waiting for send completion */ 1875 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0)); 1876 if (ret < 0) { 1877 PRINTF(cb, "poll error %d\n", ret); 1878 return; 1879 } 1880 if (wc.status) { 1881 PRINTF(cb, "send completion error %d\n", wc.status); 1882 return; 1883 } 1884 1885 /* Spin waiting for server's Start STAG/TO/Len */ 1886 while (cb->state < RDMA_WRITE_ADV) { 1887 krping_cq_event_handler(cb->cq, cb); 1888 } 1889 1890 bw_test(cb); 1891} 1892 1893static void krping_fr_test(struct krping_cb *cb) 1894{ 1895 struct ib_fast_reg_page_list *pl; 1896 struct ib_send_wr fr, inv, *bad; 1897 struct ib_wc wc; 1898 u8 key = 0; 1899 struct ib_mr *mr; 1900 int i; 1901 int ret; 1902 int size = cb->size; 1903 int plen = (((size - 1) & PAGE_MASK) + PAGE_SIZE) >> PAGE_SHIFT; 1904 time_t start; 1905 int count = 0; 1906 int scnt = 0; 1907 1908 pl = ib_alloc_fast_reg_page_list(cb->qp->device, plen); 1909 if (IS_ERR(pl)) { 1910 PRINTF(cb, "ib_alloc_fast_reg_page_list failed %ld\n", PTR_ERR(pl)); 1911 return; 1912 } 1913 1914 mr = ib_alloc_fast_reg_mr(cb->pd, plen); 1915 if (IS_ERR(mr)) { 1916 PRINTF(cb, "ib_alloc_fast_reg_mr failed %ld\n", PTR_ERR(pl)); 1917 goto err1; 1918 } 1919 1920 for (i=0; i<plen; i++) 1921 pl->page_list[i] = 0xcafebabe | i; 1922 1923 memset(&fr, 0, sizeof fr); 1924 fr.opcode = IB_WR_FAST_REG_MR; 1925 fr.wr.fast_reg.page_shift = PAGE_SHIFT; 1926 fr.wr.fast_reg.length = size; 1927 fr.wr.fast_reg.page_list = pl; 1928 fr.wr.fast_reg.page_list_len = plen; 1929 fr.wr.fast_reg.iova_start = 0; 1930 fr.wr.fast_reg.access_flags = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE; 1931 fr.next = &inv; 1932 memset(&inv, 0, sizeof inv); 1933 inv.opcode = IB_WR_LOCAL_INV; 1934 inv.send_flags = IB_SEND_SIGNALED; 1935 1936 DEBUG_LOG(cb, "fr_test: stag index 0x%x plen %u size %u depth %u\n", mr->rkey >> 8, plen, cb->size, cb->txdepth); 1937 start = time_uptime; 1938 while (1) { 1939 if ((time_uptime - start) >= 9) { 1940 DEBUG_LOG(cb, "fr_test: pausing 1 second! count %u latest size %u plen %u\n", count, size, plen); 1941 wait_event_interruptible(cb->sem, cb->state == ERROR); 1942 if (cb->state == ERROR) 1943 break; 1944 start = time_uptime; 1945 } 1946 while (scnt < (cb->txdepth>>1)) { 1947 ib_update_fast_reg_key(mr, ++key); 1948 fr.wr.fast_reg.rkey = mr->rkey; 1949 inv.ex.invalidate_rkey = mr->rkey; 1950 size = arc4random() % cb->size; 1951 if (size == 0) 1952 size = cb->size; 1953 plen = (((size - 1) & PAGE_MASK) + PAGE_SIZE) >> PAGE_SHIFT; 1954 fr.wr.fast_reg.length = size; 1955 fr.wr.fast_reg.page_list_len = plen; 1956 ret = ib_post_send(cb->qp, &fr, &bad); 1957 if (ret) { 1958 PRINTF(cb, "ib_post_send failed %d\n", ret); 1959 goto err2; 1960 } 1961 scnt++; 1962 } 1963 1964 do { 1965 ret = ib_poll_cq(cb->cq, 1, &wc); 1966 if (ret < 0) { 1967 PRINTF(cb, "ib_poll_cq failed %d\n", ret); 1968 goto err2; 1969 } 1970 if (ret == 1) { 1971 if (wc.status) { 1972 PRINTF(cb, "completion error %u\n", wc.status); 1973 goto err2; 1974 } 1975 count++; 1976 scnt--; 1977 } 1978 else if (krping_sigpending()) { 1979 PRINTF(cb, "signal!\n"); 1980 goto err2; 1981 } 1982 } while (ret == 1); 1983 } 1984err2: 1985#if 0 1986 DEBUG_LOG(cb, "sleeping 1 second\n"); 1987 wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ); 1988#endif 1989 DEBUG_LOG(cb, "draining the cq...\n"); 1990 do { 1991 ret = ib_poll_cq(cb->cq, 1, &wc); 1992 if (ret < 0) { 1993 PRINTF(cb, "ib_poll_cq failed %d\n", ret); 1994 break; 1995 } 1996 if (ret == 1) { 1997 if (wc.status) { 1998 PRINTF(cb, "completion error %u opcode %u\n", wc.status, wc.opcode); 1999 } 2000 } 2001 } while (ret == 1); 2002 DEBUG_LOG(cb, "fr_test: done!\n"); 2003 ib_dereg_mr(mr); 2004err1: 2005 ib_free_fast_reg_page_list(pl); 2006} 2007 2008static int krping_connect_client(struct krping_cb *cb) 2009{ 2010 struct rdma_conn_param conn_param; 2011 int ret; 2012 2013 memset(&conn_param, 0, sizeof conn_param); 2014 conn_param.responder_resources = 1; 2015 conn_param.initiator_depth = 1; 2016 conn_param.retry_count = 10; 2017 2018 ret = rdma_connect(cb->cm_id, &conn_param); 2019 if (ret) { 2020 PRINTF(cb, "rdma_connect error %d\n", ret); 2021 return ret; 2022 } 2023 2024 wait_event_interruptible(cb->sem, cb->state >= CONNECTED); 2025 if (cb->state == ERROR) { 2026 PRINTF(cb, "wait for CONNECTED state %d\n", cb->state); 2027 return -1; 2028 } 2029 2030 DEBUG_LOG(cb, "rdma_connect successful\n"); 2031 return 0; 2032} 2033 2034static int krping_bind_client(struct krping_cb *cb) 2035{ 2036 struct sockaddr_in sin; 2037 int ret; 2038 2039 memset(&sin, 0, sizeof(sin)); 2040 sin.sin_len = sizeof sin; 2041 sin.sin_family = AF_INET; 2042 sin.sin_addr.s_addr = cb->addr.s_addr; 2043 sin.sin_port = cb->port; 2044 2045 ret = rdma_resolve_addr(cb->cm_id, NULL, (struct sockaddr *) &sin, 2046 2000); 2047 if (ret) { 2048 PRINTF(cb, "rdma_resolve_addr error %d\n", ret); 2049 return ret; 2050 } 2051 2052 wait_event_interruptible(cb->sem, cb->state >= ROUTE_RESOLVED); 2053 if (cb->state != ROUTE_RESOLVED) { 2054 PRINTF(cb, 2055 "addr/route resolution did not resolve: state %d\n", 2056 cb->state); 2057 return -EINTR; 2058 } 2059 2060 if (cb->mem == FASTREG && !fastreg_supported(cb)) 2061 return -EINVAL; 2062 2063 DEBUG_LOG(cb, "rdma_resolve_addr - rdma_resolve_route successful\n"); 2064 return 0; 2065} 2066 2067static void krping_run_client(struct krping_cb *cb) 2068{ 2069 struct ib_recv_wr *bad_wr; 2070 int ret; 2071 2072 ret = krping_bind_client(cb); 2073 if (ret) 2074 return; 2075 2076 ret = krping_setup_qp(cb, cb->cm_id); 2077 if (ret) { 2078 PRINTF(cb, "setup_qp failed: %d\n", ret); 2079 return; 2080 } 2081 2082 ret = krping_setup_buffers(cb); 2083 if (ret) { 2084 PRINTF(cb, "krping_setup_buffers failed: %d\n", ret); 2085 goto err1; 2086 } 2087 2088 ret = ib_post_recv(cb->qp, &cb->rq_wr, &bad_wr); 2089 if (ret) { 2090 PRINTF(cb, "ib_post_recv failed: %d\n", ret); 2091 goto err2; 2092 } 2093 2094 ret = krping_connect_client(cb); 2095 if (ret) { 2096 PRINTF(cb, "connect error %d\n", ret); 2097 goto err2; 2098 } 2099 2100 if (cb->wlat) 2101 krping_wlat_test_client(cb); 2102 else if (cb->rlat) 2103 krping_rlat_test_client(cb); 2104 else if (cb->bw) 2105 krping_bw_test_client(cb); 2106 else if (cb->frtest) 2107 krping_fr_test(cb); 2108 else 2109 krping_test_client(cb); 2110 rdma_disconnect(cb->cm_id); 2111err2: 2112 krping_free_buffers(cb); 2113err1: 2114 krping_free_qp(cb); 2115} 2116 2117int krping_doit(char *cmd, void *cookie) 2118{ 2119 struct krping_cb *cb; 2120 int op; 2121 int ret = 0; 2122 char *optarg; 2123 unsigned long optint; 2124 2125 cb = kzalloc(sizeof(*cb), GFP_KERNEL); 2126 if (!cb) 2127 return -ENOMEM; 2128 2129 mutex_lock(&krping_mutex); 2130 list_add_tail(&cb->list, &krping_cbs); 2131 mutex_unlock(&krping_mutex); 2132 2133 cb->cookie = cookie; 2134 cb->server = -1; 2135 cb->state = IDLE; 2136 cb->size = 64; 2137 cb->txdepth = RPING_SQ_DEPTH; 2138 cb->mem = DMA; 2139 init_waitqueue_head(&cb->sem); 2140 2141 while ((op = krping_getopt("krping", &cmd, krping_opts, NULL, &optarg, 2142 &optint)) != 0) { 2143 switch (op) { 2144 case 'a': 2145 cb->addr_str = optarg; 2146 DEBUG_LOG(cb, "ipaddr (%s)\n", optarg); 2147 if (!inet_aton(optarg, &cb->addr)) { 2148 PRINTF(cb, "bad addr string %s\n", 2149 optarg); 2150 ret = EINVAL; 2151 } 2152 break; 2153 case 'p': 2154 cb->port = htons(optint); 2155 DEBUG_LOG(cb, "port %d\n", (int)optint); 2156 break; 2157 case 'P': 2158 cb->poll = 1; 2159 DEBUG_LOG(cb, "server\n"); 2160 break; 2161 case 's': 2162 cb->server = 1; 2163 DEBUG_LOG(cb, "server\n"); 2164 break; 2165 case 'c': 2166 cb->server = 0; 2167 DEBUG_LOG(cb, "client\n"); 2168 break; 2169 case 'S': 2170 cb->size = optint; 2171 if ((cb->size < 1) || 2172 (cb->size > RPING_BUFSIZE)) { 2173 PRINTF(cb, "Invalid size %d " 2174 "(valid range is 1 to %d)\n", 2175 cb->size, RPING_BUFSIZE); 2176 ret = EINVAL; 2177 } else 2178 DEBUG_LOG(cb, "size %d\n", (int)optint); 2179 break; 2180 case 'C': 2181 cb->count = optint; 2182 if (cb->count < 0) { 2183 PRINTF(cb, "Invalid count %d\n", 2184 cb->count); 2185 ret = EINVAL; 2186 } else 2187 DEBUG_LOG(cb, "count %d\n", (int) cb->count); 2188 break; 2189 case 'v': 2190 cb->verbose++; 2191 DEBUG_LOG(cb, "verbose\n"); 2192 break; 2193 case 'V': 2194 cb->validate++; 2195 DEBUG_LOG(cb, "validate data\n"); 2196 break; 2197 case 'l': 2198 cb->wlat++; 2199 break; 2200 case 'L': 2201 cb->rlat++; 2202 break; 2203 case 'B': 2204 cb->bw++; 2205 break; 2206 case 'd': 2207 cb->duplex++; 2208 break; 2209 case 'm': 2210 if (!strncmp(optarg, "dma", 3)) 2211 cb->mem = DMA; 2212 else if (!strncmp(optarg, "fastreg", 7)) 2213 cb->mem = FASTREG; 2214 else if (!strncmp(optarg, "mw", 2)) 2215 cb->mem = MW; 2216 else if (!strncmp(optarg, "mr", 2)) 2217 cb->mem = MR; 2218 else { 2219 PRINTF(cb, "unknown mem mode %s. " 2220 "Must be dma, fastreg, mw, or mr\n", 2221 optarg); 2222 ret = -EINVAL; 2223 break; 2224 } 2225 break; 2226 case 'I': 2227 cb->server_invalidate = 1; 2228 break; 2229 case 'T': 2230 cb->txdepth = optint; 2231 DEBUG_LOG(cb, "txdepth %d\n", (int) cb->txdepth); 2232 break; 2233 case 'Z': 2234 cb->local_dma_lkey = 1; 2235 DEBUG_LOG(cb, "using local dma lkey\n"); 2236 break; 2237 case 'R': 2238 cb->read_inv = 1; 2239 DEBUG_LOG(cb, "using read-with-inv\n"); 2240 break; 2241 case 'f': 2242 cb->frtest = 1; 2243 DEBUG_LOG(cb, "fast-reg test!\n"); 2244 break; 2245 default: 2246 PRINTF(cb, "unknown opt %s\n", optarg); 2247 ret = -EINVAL; 2248 break; 2249 } 2250 } 2251 if (ret) 2252 goto out; 2253 2254 if (cb->server == -1) { 2255 PRINTF(cb, "must be either client or server\n"); 2256 ret = -EINVAL; 2257 goto out; 2258 } 2259 2260 if (cb->server && cb->frtest) { 2261 PRINTF(cb, "must be client to run frtest\n"); 2262 ret = -EINVAL; 2263 goto out; 2264 } 2265 2266 if ((cb->frtest + cb->bw + cb->rlat + cb->wlat) > 1) { 2267 PRINTF(cb, "Pick only one test: fr, bw, rlat, wlat\n"); 2268 ret = -EINVAL; 2269 goto out; 2270 } 2271 2272 if (cb->server_invalidate && cb->mem != FASTREG) { 2273 PRINTF(cb, "server_invalidate only valid with fastreg mem_mode\n"); 2274 ret = -EINVAL; 2275 goto out; 2276 } 2277 2278 if (cb->read_inv && cb->mem != FASTREG) { 2279 PRINTF(cb, "read_inv only valid with fastreg mem_mode\n"); 2280 ret = -EINVAL; 2281 goto out; 2282 } 2283 2284 if (cb->mem != MR && (cb->wlat || cb->rlat || cb->bw)) { 2285 PRINTF(cb, "wlat, rlat, and bw tests only support mem_mode MR\n"); 2286 ret = -EINVAL; 2287 goto out; 2288 } 2289 2290 cb->cm_id = rdma_create_id(krping_cma_event_handler, cb, RDMA_PS_TCP); 2291 if (IS_ERR(cb->cm_id)) { 2292 ret = PTR_ERR(cb->cm_id); 2293 PRINTF(cb, "rdma_create_id error %d\n", ret); 2294 goto out; 2295 } 2296 DEBUG_LOG(cb, "created cm_id %p\n", cb->cm_id); 2297 2298 if (cb->server) 2299 krping_run_server(cb); 2300 else 2301 krping_run_client(cb); 2302 2303 DEBUG_LOG(cb, "destroy cm_id %p\n", cb->cm_id); 2304 rdma_destroy_id(cb->cm_id); 2305out: 2306 mutex_lock(&krping_mutex); 2307 list_del(&cb->list); 2308 mutex_unlock(&krping_mutex); 2309 kfree(cb); 2310 return ret; 2311} 2312 2313void 2314krping_walk_cb_list(void (*f)(struct krping_stats *, void *), void *arg) 2315{ 2316 struct krping_cb *cb; 2317 2318 mutex_lock(&krping_mutex); 2319 list_for_each_entry(cb, &krping_cbs, list) 2320 (*f)(cb->pd ? &cb->stats : NULL, arg); 2321 mutex_unlock(&krping_mutex); 2322} 2323 2324void krping_init(void) 2325{ 2326 2327 mutex_init(&krping_mutex); 2328}
| 42#include <linux/list.h> 43#include <linux/in.h> 44#include <linux/device.h> 45#include <linux/pci.h> 46#include <linux/sched.h> 47 48#include <asm/atomic.h> 49 50#include <rdma/ib_verbs.h> 51#include <rdma/rdma_cm.h> 52 53#include "krping.h" 54#include "getopt.h" 55 56extern int krping_debug; 57#define DEBUG_LOG(cb, x...) if (krping_debug) krping_printf((cb)->cookie, x) 58#define PRINTF(cb, x...) krping_printf((cb)->cookie, x) 59 60MODULE_AUTHOR("Steve Wise"); 61MODULE_DESCRIPTION("RDMA ping client/server"); 62MODULE_LICENSE("Dual BSD/GPL"); 63 64static __inline uint64_t 65get_cycles(void) 66{ 67 uint32_t low, high; 68 __asm __volatile("rdtsc" : "=a" (low), "=d" (high)); 69 return (low | ((u_int64_t)high << 32)); 70} 71 72typedef uint64_t cycles_t; 73 74enum mem_type { 75 DMA = 1, 76 FASTREG = 2, 77 MW = 3, 78 MR = 4 79}; 80 81static const struct krping_option krping_opts[] = { 82 {"count", OPT_INT, 'C'}, 83 {"size", OPT_INT, 'S'}, 84 {"addr", OPT_STRING, 'a'}, 85 {"port", OPT_INT, 'p'}, 86 {"verbose", OPT_NOPARAM, 'v'}, 87 {"validate", OPT_NOPARAM, 'V'}, 88 {"server", OPT_NOPARAM, 's'}, 89 {"client", OPT_NOPARAM, 'c'}, 90 {"mem_mode", OPT_STRING, 'm'}, 91 {"server_inv", OPT_NOPARAM, 'I'}, 92 {"wlat", OPT_NOPARAM, 'l'}, 93 {"rlat", OPT_NOPARAM, 'L'}, 94 {"bw", OPT_NOPARAM, 'B'}, 95 {"duplex", OPT_NOPARAM, 'd'}, 96 {"txdepth", OPT_INT, 'T'}, 97 {"poll", OPT_NOPARAM, 'P'}, 98 {"local_dma_lkey", OPT_NOPARAM, 'Z'}, 99 {"read_inv", OPT_NOPARAM, 'R'}, 100 {"fr", OPT_NOPARAM, 'f'}, 101 {NULL, 0, 0} 102}; 103 104#define htonll(x) cpu_to_be64((x)) 105#define ntohll(x) cpu_to_be64((x)) 106 107static struct mutex krping_mutex; 108 109/* 110 * List of running krping threads. 111 */ 112static LIST_HEAD(krping_cbs); 113 114/* 115 * krping "ping/pong" loop: 116 * client sends source rkey/addr/len 117 * server receives source rkey/add/len 118 * server rdma reads "ping" data from source 119 * server sends "go ahead" on rdma read completion 120 * client sends sink rkey/addr/len 121 * server receives sink rkey/addr/len 122 * server rdma writes "pong" data to sink 123 * server sends "go ahead" on rdma write completion 124 * <repeat loop> 125 */ 126 127/* 128 * These states are used to signal events between the completion handler 129 * and the main client or server thread. 130 * 131 * Once CONNECTED, they cycle through RDMA_READ_ADV, RDMA_WRITE_ADV, 132 * and RDMA_WRITE_COMPLETE for each ping. 133 */ 134enum test_state { 135 IDLE = 1, 136 CONNECT_REQUEST, 137 ADDR_RESOLVED, 138 ROUTE_RESOLVED, 139 CONNECTED, 140 RDMA_READ_ADV, 141 RDMA_READ_COMPLETE, 142 RDMA_WRITE_ADV, 143 RDMA_WRITE_COMPLETE, 144 ERROR 145}; 146 147struct krping_rdma_info { 148 uint64_t buf; 149 uint32_t rkey; 150 uint32_t size; 151}; 152 153/* 154 * Default max buffer size for IO... 155 */ 156#define RPING_BUFSIZE 128*1024 157#define RPING_SQ_DEPTH 64 158 159/* 160 * Control block struct. 161 */ 162struct krping_cb { 163 void *cookie; 164 int server; /* 0 iff client */ 165 struct ib_cq *cq; 166 struct ib_pd *pd; 167 struct ib_qp *qp; 168 169 enum mem_type mem; 170 struct ib_mr *dma_mr; 171 172 struct ib_fast_reg_page_list *page_list; 173 int page_list_len; 174 struct ib_send_wr fastreg_wr; 175 struct ib_send_wr invalidate_wr; 176 struct ib_mr *fastreg_mr; 177 int server_invalidate; 178 int read_inv; 179 u8 key; 180 181 struct ib_mw *mw; 182 struct ib_mw_bind bind_attr; 183 184 struct ib_recv_wr rq_wr; /* recv work request record */ 185 struct ib_sge recv_sgl; /* recv single SGE */ 186 struct krping_rdma_info recv_buf;/* malloc'd buffer */ 187 u64 recv_dma_addr; 188 DECLARE_PCI_UNMAP_ADDR(recv_mapping) 189 struct ib_mr *recv_mr; 190 191 struct ib_send_wr sq_wr; /* send work requrest record */ 192 struct ib_sge send_sgl; 193 struct krping_rdma_info send_buf;/* single send buf */ 194 u64 send_dma_addr; 195 DECLARE_PCI_UNMAP_ADDR(send_mapping) 196 struct ib_mr *send_mr; 197 198 struct ib_send_wr rdma_sq_wr; /* rdma work request record */ 199 struct ib_sge rdma_sgl; /* rdma single SGE */ 200 char *rdma_buf; /* used as rdma sink */ 201 u64 rdma_dma_addr; 202 DECLARE_PCI_UNMAP_ADDR(rdma_mapping) 203 struct ib_mr *rdma_mr; 204 205 uint32_t remote_rkey; /* remote guys RKEY */ 206 uint64_t remote_addr; /* remote guys TO */ 207 uint32_t remote_len; /* remote guys LEN */ 208 209 char *start_buf; /* rdma read src */ 210 u64 start_dma_addr; 211 DECLARE_PCI_UNMAP_ADDR(start_mapping) 212 struct ib_mr *start_mr; 213 214 enum test_state state; /* used for cond/signalling */ 215 wait_queue_head_t sem; 216 struct krping_stats stats; 217 218 uint16_t port; /* dst port in NBO */ 219 struct in_addr addr; /* dst addr in NBO */ 220 char *addr_str; /* dst addr string */ 221 int verbose; /* verbose logging */ 222 int count; /* ping count */ 223 int size; /* ping data size */ 224 int validate; /* validate ping data */ 225 int wlat; /* run wlat test */ 226 int rlat; /* run rlat test */ 227 int bw; /* run bw test */ 228 int duplex; /* run bw full duplex test */ 229 int poll; /* poll or block for rlat test */ 230 int txdepth; /* SQ depth */ 231 int local_dma_lkey; /* use 0 for lkey */ 232 int frtest; /* fastreg test */ 233 234 /* CM stuff */ 235 struct rdma_cm_id *cm_id; /* connection on client side,*/ 236 /* listener on server side. */ 237 struct rdma_cm_id *child_cm_id; /* connection on server side */ 238 struct list_head list; 239}; 240 241static int krping_cma_event_handler(struct rdma_cm_id *cma_id, 242 struct rdma_cm_event *event) 243{ 244 int ret; 245 struct krping_cb *cb = cma_id->context; 246 247 DEBUG_LOG(cb, "cma_event type %d cma_id %p (%s)\n", event->event, 248 cma_id, (cma_id == cb->cm_id) ? "parent" : "child"); 249 250 switch (event->event) { 251 case RDMA_CM_EVENT_ADDR_RESOLVED: 252 cb->state = ADDR_RESOLVED; 253 ret = rdma_resolve_route(cma_id, 2000); 254 if (ret) { 255 PRINTF(cb, "rdma_resolve_route error %d\n", ret); 256 wake_up_interruptible(&cb->sem); 257 } 258 break; 259 260 case RDMA_CM_EVENT_ROUTE_RESOLVED: 261 cb->state = ROUTE_RESOLVED; 262 wake_up_interruptible(&cb->sem); 263 break; 264 265 case RDMA_CM_EVENT_CONNECT_REQUEST: 266 cb->state = CONNECT_REQUEST; 267 cb->child_cm_id = cma_id; 268 DEBUG_LOG(cb, "child cma %p\n", cb->child_cm_id); 269 wake_up_interruptible(&cb->sem); 270 break; 271 272 case RDMA_CM_EVENT_ESTABLISHED: 273 DEBUG_LOG(cb, "ESTABLISHED\n"); 274 if (!cb->server) { 275 cb->state = CONNECTED; 276 } 277 wake_up_interruptible(&cb->sem); 278 break; 279 280 case RDMA_CM_EVENT_ADDR_ERROR: 281 case RDMA_CM_EVENT_ROUTE_ERROR: 282 case RDMA_CM_EVENT_CONNECT_ERROR: 283 case RDMA_CM_EVENT_UNREACHABLE: 284 case RDMA_CM_EVENT_REJECTED: 285 PRINTF(cb, "cma event %d, error %d\n", event->event, 286 event->status); 287 cb->state = ERROR; 288 wake_up_interruptible(&cb->sem); 289 break; 290 291 case RDMA_CM_EVENT_DISCONNECTED: 292 PRINTF(cb, "DISCONNECT EVENT...\n"); 293 cb->state = ERROR; 294 wake_up_interruptible(&cb->sem); 295 break; 296 297 case RDMA_CM_EVENT_DEVICE_REMOVAL: 298 PRINTF(cb, "cma detected device removal!!!!\n"); 299 break; 300 301 default: 302 PRINTF(cb, "oof bad type!\n"); 303 wake_up_interruptible(&cb->sem); 304 break; 305 } 306 return 0; 307} 308 309static int server_recv(struct krping_cb *cb, struct ib_wc *wc) 310{ 311 if (wc->byte_len != sizeof(cb->recv_buf)) { 312 PRINTF(cb, "Received bogus data, size %d\n", 313 wc->byte_len); 314 return -1; 315 } 316 317 cb->remote_rkey = ntohl(cb->recv_buf.rkey); 318 cb->remote_addr = ntohll(cb->recv_buf.buf); 319 cb->remote_len = ntohl(cb->recv_buf.size); 320 DEBUG_LOG(cb, "Received rkey %x addr %llx len %d from peer\n", 321 cb->remote_rkey, (unsigned long long)cb->remote_addr, 322 cb->remote_len); 323 324 if (cb->state <= CONNECTED || cb->state == RDMA_WRITE_COMPLETE) 325 cb->state = RDMA_READ_ADV; 326 else 327 cb->state = RDMA_WRITE_ADV; 328 329 return 0; 330} 331 332static int client_recv(struct krping_cb *cb, struct ib_wc *wc) 333{ 334 if (wc->byte_len != sizeof(cb->recv_buf)) { 335 PRINTF(cb, "Received bogus data, size %d\n", 336 wc->byte_len); 337 return -1; 338 } 339 340 if (cb->state == RDMA_READ_ADV) 341 cb->state = RDMA_WRITE_ADV; 342 else 343 cb->state = RDMA_WRITE_COMPLETE; 344 345 return 0; 346} 347 348static void krping_cq_event_handler(struct ib_cq *cq, void *ctx) 349{ 350 struct krping_cb *cb = ctx; 351 struct ib_wc wc; 352 struct ib_recv_wr *bad_wr; 353 int ret; 354 355 BUG_ON(cb->cq != cq); 356 if (cb->state == ERROR) { 357 PRINTF(cb, "cq completion in ERROR state\n"); 358 return; 359 } 360 if (cb->frtest) { 361 PRINTF(cb, "cq completion event in frtest!\n"); 362 return; 363 } 364 if (!cb->wlat && !cb->rlat && !cb->bw) 365 ib_req_notify_cq(cb->cq, IB_CQ_NEXT_COMP); 366 while ((ret = ib_poll_cq(cb->cq, 1, &wc)) == 1) { 367 if (wc.status) { 368 if (wc.status == IB_WC_WR_FLUSH_ERR) { 369 DEBUG_LOG(cb, "cq flushed\n"); 370 continue; 371 } else { 372 PRINTF(cb, "cq completion failed with " 373 "wr_id %Lx status %d opcode %d vender_err %x\n", 374 wc.wr_id, wc.status, wc.opcode, wc.vendor_err); 375 goto error; 376 } 377 } 378 379 switch (wc.opcode) { 380 case IB_WC_SEND: 381 DEBUG_LOG(cb, "send completion\n"); 382 cb->stats.send_bytes += cb->send_sgl.length; 383 cb->stats.send_msgs++; 384 break; 385 386 case IB_WC_RDMA_WRITE: 387 DEBUG_LOG(cb, "rdma write completion\n"); 388 cb->stats.write_bytes += cb->rdma_sq_wr.sg_list->length; 389 cb->stats.write_msgs++; 390 cb->state = RDMA_WRITE_COMPLETE; 391 wake_up_interruptible(&cb->sem); 392 break; 393 394 case IB_WC_RDMA_READ: 395 DEBUG_LOG(cb, "rdma read completion\n"); 396 cb->stats.read_bytes += cb->rdma_sq_wr.sg_list->length; 397 cb->stats.read_msgs++; 398 cb->state = RDMA_READ_COMPLETE; 399 wake_up_interruptible(&cb->sem); 400 break; 401 402 case IB_WC_RECV: 403 DEBUG_LOG(cb, "recv completion\n"); 404 cb->stats.recv_bytes += sizeof(cb->recv_buf); 405 cb->stats.recv_msgs++; 406 if (cb->wlat || cb->rlat || cb->bw) 407 ret = server_recv(cb, &wc); 408 else 409 ret = cb->server ? server_recv(cb, &wc) : 410 client_recv(cb, &wc); 411 if (ret) { 412 PRINTF(cb, "recv wc error: %d\n", ret); 413 goto error; 414 } 415 416 ret = ib_post_recv(cb->qp, &cb->rq_wr, &bad_wr); 417 if (ret) { 418 PRINTF(cb, "post recv error: %d\n", 419 ret); 420 goto error; 421 } 422 wake_up_interruptible(&cb->sem); 423 break; 424 425 default: 426 PRINTF(cb, 427 "%s:%d Unexpected opcode %d, Shutting down\n", 428 __func__, __LINE__, wc.opcode); 429 goto error; 430 } 431 } 432 if (ret) { 433 PRINTF(cb, "poll error %d\n", ret); 434 goto error; 435 } 436 return; 437error: 438 cb->state = ERROR; 439 wake_up_interruptible(&cb->sem); 440} 441 442static int krping_accept(struct krping_cb *cb) 443{ 444 struct rdma_conn_param conn_param; 445 int ret; 446 447 DEBUG_LOG(cb, "accepting client connection request\n"); 448 449 memset(&conn_param, 0, sizeof conn_param); 450 conn_param.responder_resources = 1; 451 conn_param.initiator_depth = 1; 452 453 ret = rdma_accept(cb->child_cm_id, &conn_param); 454 if (ret) { 455 PRINTF(cb, "rdma_accept error: %d\n", ret); 456 return ret; 457 } 458 459 if (!cb->wlat && !cb->rlat && !cb->bw) { 460 wait_event_interruptible(cb->sem, cb->state >= CONNECTED); 461 if (cb->state == ERROR) { 462 PRINTF(cb, "wait for CONNECTED state %d\n", 463 cb->state); 464 return -1; 465 } 466 } 467 return 0; 468} 469 470static void krping_setup_wr(struct krping_cb *cb) 471{ 472 cb->recv_sgl.addr = cb->recv_dma_addr; 473 cb->recv_sgl.length = sizeof cb->recv_buf; 474 if (cb->local_dma_lkey) 475 cb->recv_sgl.lkey = cb->qp->device->local_dma_lkey; 476 else if (cb->mem == DMA) 477 cb->recv_sgl.lkey = cb->dma_mr->lkey; 478 else 479 cb->recv_sgl.lkey = cb->recv_mr->lkey; 480 cb->rq_wr.sg_list = &cb->recv_sgl; 481 cb->rq_wr.num_sge = 1; 482 483 cb->send_sgl.addr = cb->send_dma_addr; 484 cb->send_sgl.length = sizeof cb->send_buf; 485 if (cb->local_dma_lkey) 486 cb->send_sgl.lkey = cb->qp->device->local_dma_lkey; 487 else if (cb->mem == DMA) 488 cb->send_sgl.lkey = cb->dma_mr->lkey; 489 else 490 cb->send_sgl.lkey = cb->send_mr->lkey; 491 492 cb->sq_wr.opcode = IB_WR_SEND; 493 cb->sq_wr.send_flags = IB_SEND_SIGNALED; 494 cb->sq_wr.sg_list = &cb->send_sgl; 495 cb->sq_wr.num_sge = 1; 496 497 if (cb->server || cb->wlat || cb->rlat || cb->bw) { 498 cb->rdma_sgl.addr = cb->rdma_dma_addr; 499 if (cb->mem == MR) 500 cb->rdma_sgl.lkey = cb->rdma_mr->lkey; 501 cb->rdma_sq_wr.send_flags = IB_SEND_SIGNALED; 502 cb->rdma_sq_wr.sg_list = &cb->rdma_sgl; 503 cb->rdma_sq_wr.num_sge = 1; 504 } 505 506 switch(cb->mem) { 507 case FASTREG: 508 509 /* 510 * A chain of 2 WRs, INVALDATE_MR + FAST_REG_MR. 511 * both unsignaled. The client uses them to reregister 512 * the rdma buffers with a new key each iteration. 513 */ 514 cb->fastreg_wr.opcode = IB_WR_FAST_REG_MR; 515 cb->fastreg_wr.wr.fast_reg.page_shift = PAGE_SHIFT; 516 cb->fastreg_wr.wr.fast_reg.length = cb->size; 517 cb->fastreg_wr.wr.fast_reg.page_list = cb->page_list; 518 cb->fastreg_wr.wr.fast_reg.page_list_len = cb->page_list_len; 519 520 cb->invalidate_wr.next = &cb->fastreg_wr; 521 cb->invalidate_wr.opcode = IB_WR_LOCAL_INV; 522 break; 523 case MW: 524 cb->bind_attr.wr_id = 0xabbaabba; 525 cb->bind_attr.send_flags = 0; /* unsignaled */ 526 cb->bind_attr.length = cb->size; 527 break; 528 default: 529 break; 530 } 531} 532 533static int krping_setup_buffers(struct krping_cb *cb) 534{ 535 int ret; 536 struct ib_phys_buf buf; 537 u64 iovbase; 538 539 DEBUG_LOG(cb, "krping_setup_buffers called on cb %p\n", cb); 540 541 cb->recv_dma_addr = dma_map_single(cb->pd->device->dma_device, 542 &cb->recv_buf, 543 sizeof(cb->recv_buf), DMA_BIDIRECTIONAL); 544 pci_unmap_addr_set(cb, recv_mapping, cb->recv_dma_addr); 545 cb->send_dma_addr = dma_map_single(cb->pd->device->dma_device, 546 &cb->send_buf, sizeof(cb->send_buf), 547 DMA_BIDIRECTIONAL); 548 pci_unmap_addr_set(cb, send_mapping, cb->send_dma_addr); 549 550 if (cb->mem == DMA) { 551 cb->dma_mr = ib_get_dma_mr(cb->pd, IB_ACCESS_LOCAL_WRITE| 552 IB_ACCESS_REMOTE_READ| 553 IB_ACCESS_REMOTE_WRITE); 554 if (IS_ERR(cb->dma_mr)) { 555 DEBUG_LOG(cb, "reg_dmamr failed\n"); 556 ret = PTR_ERR(cb->dma_mr); 557 goto bail; 558 } 559 } else { 560 if (!cb->local_dma_lkey) { 561 buf.addr = cb->recv_dma_addr; 562 buf.size = sizeof cb->recv_buf; 563 DEBUG_LOG(cb, "recv buf dma_addr %llx size %d\n", buf.addr, 564 (int)buf.size); 565 iovbase = cb->recv_dma_addr; 566 cb->recv_mr = ib_reg_phys_mr(cb->pd, &buf, 1, 567 IB_ACCESS_LOCAL_WRITE, 568 &iovbase); 569 570 if (IS_ERR(cb->recv_mr)) { 571 DEBUG_LOG(cb, "recv_buf reg_mr failed\n"); 572 ret = PTR_ERR(cb->recv_mr); 573 goto bail; 574 } 575 576 buf.addr = cb->send_dma_addr; 577 buf.size = sizeof cb->send_buf; 578 DEBUG_LOG(cb, "send buf dma_addr %llx size %d\n", buf.addr, 579 (int)buf.size); 580 iovbase = cb->send_dma_addr; 581 cb->send_mr = ib_reg_phys_mr(cb->pd, &buf, 1, 582 0, &iovbase); 583 584 if (IS_ERR(cb->send_mr)) { 585 DEBUG_LOG(cb, "send_buf reg_mr failed\n"); 586 ret = PTR_ERR(cb->send_mr); 587 goto bail; 588 } 589 } 590 } 591 592 cb->rdma_buf = kmalloc(cb->size, GFP_KERNEL); 593 if (!cb->rdma_buf) { 594 DEBUG_LOG(cb, "rdma_buf malloc failed\n"); 595 ret = -ENOMEM; 596 goto bail; 597 } 598 599 cb->rdma_dma_addr = dma_map_single(cb->pd->device->dma_device, 600 cb->rdma_buf, cb->size, 601 DMA_BIDIRECTIONAL); 602 pci_unmap_addr_set(cb, rdma_mapping, cb->rdma_dma_addr); 603 if (cb->mem != DMA) { 604 switch (cb->mem) { 605 case FASTREG: 606 cb->page_list_len = (((cb->size - 1) & PAGE_MASK) + 607 PAGE_SIZE) >> PAGE_SHIFT; 608 cb->page_list = ib_alloc_fast_reg_page_list( 609 cb->pd->device, 610 cb->page_list_len); 611 if (IS_ERR(cb->page_list)) { 612 DEBUG_LOG(cb, "recv_buf reg_mr failed\n"); 613 ret = PTR_ERR(cb->page_list); 614 goto bail; 615 } 616 cb->fastreg_mr = ib_alloc_fast_reg_mr(cb->pd, 617 cb->page_list->max_page_list_len); 618 if (IS_ERR(cb->fastreg_mr)) { 619 DEBUG_LOG(cb, "recv_buf reg_mr failed\n"); 620 ret = PTR_ERR(cb->fastreg_mr); 621 goto bail; 622 } 623 DEBUG_LOG(cb, "fastreg rkey 0x%x page_list %p" 624 " page_list_len %u\n", cb->fastreg_mr->rkey, 625 cb->page_list, cb->page_list_len); 626 break; 627 case MW: 628 cb->mw = ib_alloc_mw(cb->pd); 629 if (IS_ERR(cb->mw)) { 630 DEBUG_LOG(cb, "recv_buf alloc_mw failed\n"); 631 ret = PTR_ERR(cb->mw); 632 goto bail; 633 } 634 DEBUG_LOG(cb, "mw rkey 0x%x\n", cb->mw->rkey); 635 /*FALLTHROUGH*/ 636 case MR: 637 buf.addr = cb->rdma_dma_addr; 638 buf.size = cb->size; 639 iovbase = cb->rdma_dma_addr; 640 cb->rdma_mr = ib_reg_phys_mr(cb->pd, &buf, 1, 641 IB_ACCESS_REMOTE_READ| 642 IB_ACCESS_REMOTE_WRITE, 643 &iovbase); 644 if (IS_ERR(cb->rdma_mr)) { 645 DEBUG_LOG(cb, "rdma_buf reg_mr failed\n"); 646 ret = PTR_ERR(cb->rdma_mr); 647 goto bail; 648 } 649 DEBUG_LOG(cb, "rdma buf dma_addr %llx size %d mr rkey 0x%x\n", 650 buf.addr, (int)buf.size, cb->rdma_mr->rkey); 651 break; 652 default: 653 ret = -EINVAL; 654 goto bail; 655 break; 656 } 657 } 658 659 if (!cb->server || cb->wlat || cb->rlat || cb->bw) { 660 661 cb->start_buf = kmalloc(cb->size, GFP_KERNEL); 662 if (!cb->start_buf) { 663 DEBUG_LOG(cb, "start_buf malloc failed\n"); 664 ret = -ENOMEM; 665 goto bail; 666 } 667 668 cb->start_dma_addr = dma_map_single(cb->pd->device->dma_device, 669 cb->start_buf, cb->size, 670 DMA_BIDIRECTIONAL); 671 pci_unmap_addr_set(cb, start_mapping, cb->start_dma_addr); 672 673 if (cb->mem == MR || cb->mem == MW) { 674 unsigned flags = IB_ACCESS_REMOTE_READ; 675 676 if (cb->wlat || cb->rlat || cb->bw) 677 flags |= IB_ACCESS_REMOTE_WRITE; 678 679 buf.addr = cb->start_dma_addr; 680 buf.size = cb->size; 681 DEBUG_LOG(cb, "start buf dma_addr %llx size %d\n", 682 buf.addr, (int)buf.size); 683 iovbase = cb->start_dma_addr; 684 cb->start_mr = ib_reg_phys_mr(cb->pd, &buf, 1, 685 flags, 686 &iovbase); 687 688 if (IS_ERR(cb->start_mr)) { 689 DEBUG_LOG(cb, "start_buf reg_mr failed\n"); 690 ret = PTR_ERR(cb->start_mr); 691 goto bail; 692 } 693 } 694 } 695 696 krping_setup_wr(cb); 697 DEBUG_LOG(cb, "allocated & registered buffers...\n"); 698 return 0; 699bail: 700 if (cb->fastreg_mr && !IS_ERR(cb->fastreg_mr)) 701 ib_dereg_mr(cb->fastreg_mr); 702 if (cb->mw && !IS_ERR(cb->mw)) 703 ib_dealloc_mw(cb->mw); 704 if (cb->rdma_mr && !IS_ERR(cb->rdma_mr)) 705 ib_dereg_mr(cb->rdma_mr); 706 if (cb->page_list && !IS_ERR(cb->page_list)) 707 ib_free_fast_reg_page_list(cb->page_list); 708 if (cb->dma_mr && !IS_ERR(cb->dma_mr)) 709 ib_dereg_mr(cb->dma_mr); 710 if (cb->recv_mr && !IS_ERR(cb->recv_mr)) 711 ib_dereg_mr(cb->recv_mr); 712 if (cb->send_mr && !IS_ERR(cb->send_mr)) 713 ib_dereg_mr(cb->send_mr); 714 if (cb->rdma_buf) 715 kfree(cb->rdma_buf); 716 if (cb->start_buf) 717 kfree(cb->start_buf); 718 return ret; 719} 720 721static void krping_free_buffers(struct krping_cb *cb) 722{ 723 DEBUG_LOG(cb, "krping_free_buffers called on cb %p\n", cb); 724 725 if (cb->dma_mr) 726 ib_dereg_mr(cb->dma_mr); 727 if (cb->send_mr) 728 ib_dereg_mr(cb->send_mr); 729 if (cb->recv_mr) 730 ib_dereg_mr(cb->recv_mr); 731 if (cb->rdma_mr) 732 ib_dereg_mr(cb->rdma_mr); 733 if (cb->start_mr) 734 ib_dereg_mr(cb->start_mr); 735 if (cb->fastreg_mr) 736 ib_dereg_mr(cb->fastreg_mr); 737 if (cb->mw) 738 ib_dealloc_mw(cb->mw); 739 740 dma_unmap_single(cb->pd->device->dma_device, 741 pci_unmap_addr(cb, recv_mapping), 742 sizeof(cb->recv_buf), DMA_BIDIRECTIONAL); 743 dma_unmap_single(cb->pd->device->dma_device, 744 pci_unmap_addr(cb, send_mapping), 745 sizeof(cb->send_buf), DMA_BIDIRECTIONAL); 746 dma_unmap_single(cb->pd->device->dma_device, 747 pci_unmap_addr(cb, rdma_mapping), 748 cb->size, DMA_BIDIRECTIONAL); 749 kfree(cb->rdma_buf); 750 if (cb->start_buf) { 751 dma_unmap_single(cb->pd->device->dma_device, 752 pci_unmap_addr(cb, start_mapping), 753 cb->size, DMA_BIDIRECTIONAL); 754 kfree(cb->start_buf); 755 } 756} 757 758static int krping_create_qp(struct krping_cb *cb) 759{ 760 struct ib_qp_init_attr init_attr; 761 int ret; 762 763 memset(&init_attr, 0, sizeof(init_attr)); 764 init_attr.cap.max_send_wr = cb->txdepth; 765 init_attr.cap.max_recv_wr = 2; 766 init_attr.cap.max_recv_sge = 1; 767 init_attr.cap.max_send_sge = 1; 768 init_attr.qp_type = IB_QPT_RC; 769 init_attr.send_cq = cb->cq; 770 init_attr.recv_cq = cb->cq; 771 init_attr.sq_sig_type = IB_SIGNAL_REQ_WR; 772 773 if (cb->server) { 774 ret = rdma_create_qp(cb->child_cm_id, cb->pd, &init_attr); 775 if (!ret) 776 cb->qp = cb->child_cm_id->qp; 777 } else { 778 ret = rdma_create_qp(cb->cm_id, cb->pd, &init_attr); 779 if (!ret) 780 cb->qp = cb->cm_id->qp; 781 } 782 783 return ret; 784} 785 786static void krping_free_qp(struct krping_cb *cb) 787{ 788 ib_destroy_qp(cb->qp); 789 ib_destroy_cq(cb->cq); 790 ib_dealloc_pd(cb->pd); 791} 792 793static int krping_setup_qp(struct krping_cb *cb, struct rdma_cm_id *cm_id) 794{ 795 int ret; 796 cb->pd = ib_alloc_pd(cm_id->device); 797 if (IS_ERR(cb->pd)) { 798 PRINTF(cb, "ib_alloc_pd failed\n"); 799 return PTR_ERR(cb->pd); 800 } 801 DEBUG_LOG(cb, "created pd %p\n", cb->pd); 802 803 strlcpy(cb->stats.name, cb->pd->device->name, sizeof(cb->stats.name)); 804 805 cb->cq = ib_create_cq(cm_id->device, krping_cq_event_handler, NULL, 806 cb, cb->txdepth * 2, 0); 807 if (IS_ERR(cb->cq)) { 808 PRINTF(cb, "ib_create_cq failed\n"); 809 ret = PTR_ERR(cb->cq); 810 goto err1; 811 } 812 DEBUG_LOG(cb, "created cq %p\n", cb->cq); 813 814 if (!cb->wlat && !cb->rlat && !cb->bw && !cb->frtest) { 815 ret = ib_req_notify_cq(cb->cq, IB_CQ_NEXT_COMP); 816 if (ret) { 817 PRINTF(cb, "ib_create_cq failed\n"); 818 goto err2; 819 } 820 } 821 822 ret = krping_create_qp(cb); 823 if (ret) { 824 PRINTF(cb, "krping_create_qp failed: %d\n", ret); 825 goto err2; 826 } 827 DEBUG_LOG(cb, "created qp %p\n", cb->qp); 828 return 0; 829err2: 830 ib_destroy_cq(cb->cq); 831err1: 832 ib_dealloc_pd(cb->pd); 833 return ret; 834} 835 836/* 837 * return the (possibly rebound) rkey for the rdma buffer. 838 * FASTREG mode: invalidate and rebind via fastreg wr. 839 * MW mode: rebind the MW. 840 * other modes: just return the mr rkey. 841 */ 842static u32 krping_rdma_rkey(struct krping_cb *cb, u64 buf, int post_inv) 843{ 844 u32 rkey = 0xffffffff; 845 u64 p; 846 struct ib_send_wr *bad_wr; 847 int i; 848 int ret; 849 850 switch (cb->mem) { 851 case FASTREG: 852 cb->invalidate_wr.ex.invalidate_rkey = cb->fastreg_mr->rkey; 853 854 /* 855 * Update the fastreg key. 856 */ 857 ib_update_fast_reg_key(cb->fastreg_mr, ++cb->key); 858 cb->fastreg_wr.wr.fast_reg.rkey = cb->fastreg_mr->rkey; 859 860 /* 861 * Update the fastreg WR with new buf info. 862 */ 863 if (buf == (u64)cb->start_dma_addr) 864 cb->fastreg_wr.wr.fast_reg.access_flags = IB_ACCESS_REMOTE_READ; 865 else 866 cb->fastreg_wr.wr.fast_reg.access_flags = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE; 867 cb->fastreg_wr.wr.fast_reg.iova_start = buf; 868 p = (u64)(buf & PAGE_MASK); 869 for (i=0; i < cb->fastreg_wr.wr.fast_reg.page_list_len; 870 i++, p += PAGE_SIZE) { 871 cb->page_list->page_list[i] = p; 872 DEBUG_LOG(cb, "page_list[%d] 0x%llx\n", i, p); 873 } 874 875 DEBUG_LOG(cb, "post_inv = %d, fastreg new rkey 0x%x shift %u len %u" 876 " iova_start %llx page_list_len %u\n", 877 post_inv, 878 cb->fastreg_wr.wr.fast_reg.rkey, 879 cb->fastreg_wr.wr.fast_reg.page_shift, 880 cb->fastreg_wr.wr.fast_reg.length, 881 cb->fastreg_wr.wr.fast_reg.iova_start, 882 cb->fastreg_wr.wr.fast_reg.page_list_len); 883 884 if (post_inv) 885 ret = ib_post_send(cb->qp, &cb->invalidate_wr, &bad_wr); 886 else 887 ret = ib_post_send(cb->qp, &cb->fastreg_wr, &bad_wr); 888 if (ret) { 889 PRINTF(cb, "post send error %d\n", ret); 890 cb->state = ERROR; 891 } 892 rkey = cb->fastreg_mr->rkey; 893 break; 894 case MW: 895 /* 896 * Update the MW with new buf info. 897 */ 898 if (buf == (u64)cb->start_dma_addr) { 899 cb->bind_attr.mw_access_flags = IB_ACCESS_REMOTE_READ; 900 cb->bind_attr.mr = cb->start_mr; 901 } else { 902 cb->bind_attr.mw_access_flags = IB_ACCESS_REMOTE_WRITE; 903 cb->bind_attr.mr = cb->rdma_mr; 904 } 905 cb->bind_attr.addr = buf; 906 DEBUG_LOG(cb, "binding mw rkey 0x%x to buf %llx mr rkey 0x%x\n", 907 cb->mw->rkey, buf, cb->bind_attr.mr->rkey); 908 ret = ib_bind_mw(cb->qp, cb->mw, &cb->bind_attr); 909 if (ret) { 910 PRINTF(cb, "bind mw error %d\n", ret); 911 cb->state = ERROR; 912 } else 913 rkey = cb->mw->rkey; 914 break; 915 case MR: 916 if (buf == (u64)cb->start_dma_addr) 917 rkey = cb->start_mr->rkey; 918 else 919 rkey = cb->rdma_mr->rkey; 920 break; 921 case DMA: 922 rkey = cb->dma_mr->rkey; 923 break; 924 default: 925 PRINTF(cb, "%s:%d case ERROR\n", __func__, __LINE__); 926 cb->state = ERROR; 927 break; 928 } 929 return rkey; 930} 931 932static void krping_format_send(struct krping_cb *cb, u64 buf) 933{ 934 struct krping_rdma_info *info = &cb->send_buf; 935 u32 rkey; 936 937 /* 938 * Client side will do fastreg or mw bind before 939 * advertising the rdma buffer. Server side 940 * sends have no data. 941 */ 942 if (!cb->server || cb->wlat || cb->rlat || cb->bw) { 943 rkey = krping_rdma_rkey(cb, buf, !cb->server_invalidate); 944 info->buf = htonll(buf); 945 info->rkey = htonl(rkey); 946 info->size = htonl(cb->size); 947 DEBUG_LOG(cb, "RDMA addr %llx rkey %x len %d\n", 948 (unsigned long long)buf, rkey, cb->size); 949 } 950} 951 952static void krping_test_server(struct krping_cb *cb) 953{ 954 struct ib_send_wr *bad_wr, inv; 955 int ret; 956 957 while (1) { 958 /* Wait for client's Start STAG/TO/Len */ 959 wait_event_interruptible(cb->sem, cb->state >= RDMA_READ_ADV); 960 if (cb->state != RDMA_READ_ADV) { 961 PRINTF(cb, "wait for RDMA_READ_ADV state %d\n", 962 cb->state); 963 break; 964 } 965 966 DEBUG_LOG(cb, "server received sink adv\n"); 967 968 cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey; 969 cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr; 970 cb->rdma_sq_wr.sg_list->length = cb->remote_len; 971 cb->rdma_sgl.lkey = krping_rdma_rkey(cb, cb->rdma_dma_addr, 1); 972 cb->rdma_sq_wr.next = NULL; 973 974 /* Issue RDMA Read. */ 975 if (cb->read_inv) 976 cb->rdma_sq_wr.opcode = IB_WR_RDMA_READ_WITH_INV; 977 else { 978 979 cb->rdma_sq_wr.opcode = IB_WR_RDMA_READ; 980 if (cb->mem == FASTREG) { 981 /* 982 * Immediately follow the read with a 983 * fenced LOCAL_INV. 984 */ 985 cb->rdma_sq_wr.next = &inv; 986 memset(&inv, 0, sizeof inv); 987 inv.opcode = IB_WR_LOCAL_INV; 988 inv.ex.invalidate_rkey = cb->fastreg_mr->rkey; 989 inv.send_flags = IB_SEND_FENCE; 990 } 991 } 992 993 ret = ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr); 994 if (ret) { 995 PRINTF(cb, "post send error %d\n", ret); 996 break; 997 } 998 cb->rdma_sq_wr.next = NULL; 999 1000 DEBUG_LOG(cb, "server posted rdma read req \n"); 1001 1002 /* Wait for read completion */ 1003 wait_event_interruptible(cb->sem, 1004 cb->state >= RDMA_READ_COMPLETE); 1005 if (cb->state != RDMA_READ_COMPLETE) { 1006 PRINTF(cb, 1007 "wait for RDMA_READ_COMPLETE state %d\n", 1008 cb->state); 1009 break; 1010 } 1011 DEBUG_LOG(cb, "server received read complete\n"); 1012 1013 /* Display data in recv buf */ 1014 if (cb->verbose) 1015 PRINTF(cb, "server ping data: %s\n", 1016 cb->rdma_buf); 1017 1018 /* Tell client to continue */ 1019 if (cb->server && cb->server_invalidate) { 1020 cb->sq_wr.ex.invalidate_rkey = cb->remote_rkey; 1021 cb->sq_wr.opcode = IB_WR_SEND_WITH_INV; 1022 DEBUG_LOG(cb, "send-w-inv rkey 0x%x\n", cb->remote_rkey); 1023 } 1024 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1025 if (ret) { 1026 PRINTF(cb, "post send error %d\n", ret); 1027 break; 1028 } 1029 DEBUG_LOG(cb, "server posted go ahead\n"); 1030 1031 /* Wait for client's RDMA STAG/TO/Len */ 1032 wait_event_interruptible(cb->sem, cb->state >= RDMA_WRITE_ADV); 1033 if (cb->state != RDMA_WRITE_ADV) { 1034 PRINTF(cb, 1035 "wait for RDMA_WRITE_ADV state %d\n", 1036 cb->state); 1037 break; 1038 } 1039 DEBUG_LOG(cb, "server received sink adv\n"); 1040 1041 /* RDMA Write echo data */ 1042 cb->rdma_sq_wr.opcode = IB_WR_RDMA_WRITE; 1043 cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey; 1044 cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr; 1045 cb->rdma_sq_wr.sg_list->length = strlen(cb->rdma_buf) + 1; 1046 if (cb->local_dma_lkey) 1047 cb->rdma_sgl.lkey = cb->qp->device->local_dma_lkey; 1048 else 1049 cb->rdma_sgl.lkey = krping_rdma_rkey(cb, cb->rdma_dma_addr, 0); 1050 1051 DEBUG_LOG(cb, "rdma write from lkey %x laddr %llx len %d\n", 1052 cb->rdma_sq_wr.sg_list->lkey, 1053 (unsigned long long)cb->rdma_sq_wr.sg_list->addr, 1054 cb->rdma_sq_wr.sg_list->length); 1055 1056 ret = ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr); 1057 if (ret) { 1058 PRINTF(cb, "post send error %d\n", ret); 1059 break; 1060 } 1061 1062 /* Wait for completion */ 1063 ret = wait_event_interruptible(cb->sem, cb->state >= 1064 RDMA_WRITE_COMPLETE); 1065 if (cb->state != RDMA_WRITE_COMPLETE) { 1066 PRINTF(cb, 1067 "wait for RDMA_WRITE_COMPLETE state %d\n", 1068 cb->state); 1069 break; 1070 } 1071 DEBUG_LOG(cb, "server rdma write complete \n"); 1072 1073 cb->state = CONNECTED; 1074 1075 /* Tell client to begin again */ 1076 if (cb->server && cb->server_invalidate) { 1077 cb->sq_wr.ex.invalidate_rkey = cb->remote_rkey; 1078 cb->sq_wr.opcode = IB_WR_SEND_WITH_INV; 1079 DEBUG_LOG(cb, "send-w-inv rkey 0x%x\n", cb->remote_rkey); 1080 } 1081 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1082 if (ret) { 1083 PRINTF(cb, "post send error %d\n", ret); 1084 break; 1085 } 1086 DEBUG_LOG(cb, "server posted go ahead\n"); 1087 } 1088} 1089 1090static void rlat_test(struct krping_cb *cb) 1091{ 1092 int scnt; 1093 int iters = cb->count; 1094 struct timeval start_tv, stop_tv; 1095 int ret; 1096 struct ib_wc wc; 1097 struct ib_send_wr *bad_wr; 1098 int ne; 1099 1100 scnt = 0; 1101 cb->rdma_sq_wr.opcode = IB_WR_RDMA_READ; 1102 cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey; 1103 cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr; 1104 cb->rdma_sq_wr.sg_list->length = cb->size; 1105 1106 microtime(&start_tv); 1107 if (!cb->poll) { 1108 cb->state = RDMA_READ_ADV; 1109 ib_req_notify_cq(cb->cq, IB_CQ_NEXT_COMP); 1110 } 1111 while (scnt < iters) { 1112 1113 cb->state = RDMA_READ_ADV; 1114 ret = ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr); 1115 if (ret) { 1116 PRINTF(cb, 1117 "Couldn't post send: ret=%d scnt %d\n", 1118 ret, scnt); 1119 return; 1120 } 1121 1122 do { 1123 if (!cb->poll) { 1124 wait_event_interruptible(cb->sem, 1125 cb->state != RDMA_READ_ADV); 1126 if (cb->state == RDMA_READ_COMPLETE) { 1127 ne = 1; 1128 ib_req_notify_cq(cb->cq, 1129 IB_CQ_NEXT_COMP); 1130 } else { 1131 ne = -1; 1132 } 1133 } else 1134 ne = ib_poll_cq(cb->cq, 1, &wc); 1135 if (cb->state == ERROR) { 1136 PRINTF(cb, 1137 "state == ERROR...bailing scnt %d\n", 1138 scnt); 1139 return; 1140 } 1141 } while (ne == 0); 1142 1143 if (ne < 0) { 1144 PRINTF(cb, "poll CQ failed %d\n", ne); 1145 return; 1146 } 1147 if (cb->poll && wc.status != IB_WC_SUCCESS) { 1148 PRINTF(cb, "Completion wth error at %s:\n", 1149 cb->server ? "server" : "client"); 1150 PRINTF(cb, "Failed status %d: wr_id %d\n", 1151 wc.status, (int) wc.wr_id); 1152 return; 1153 } 1154 ++scnt; 1155 } 1156 microtime(&stop_tv); 1157 1158 if (stop_tv.tv_usec < start_tv.tv_usec) { 1159 stop_tv.tv_usec += 1000000; 1160 stop_tv.tv_sec -= 1; 1161 } 1162 1163 PRINTF(cb, "delta sec %lu delta usec %lu iter %d size %d\n", 1164 stop_tv.tv_sec - start_tv.tv_sec, 1165 stop_tv.tv_usec - start_tv.tv_usec, 1166 scnt, cb->size); 1167} 1168 1169static void wlat_test(struct krping_cb *cb) 1170{ 1171 int ccnt, scnt, rcnt; 1172 int iters=cb->count; 1173 volatile char *poll_buf = (char *) cb->start_buf; 1174 char *buf = (char *)cb->rdma_buf; 1175 struct timeval start_tv, stop_tv; 1176 cycles_t *post_cycles_start, *post_cycles_stop; 1177 cycles_t *poll_cycles_start, *poll_cycles_stop; 1178 cycles_t *last_poll_cycles_start; 1179 cycles_t sum_poll = 0, sum_post = 0, sum_last_poll = 0; 1180 int i; 1181 int cycle_iters = 1000; 1182 1183 ccnt = 0; 1184 scnt = 0; 1185 rcnt = 0; 1186 1187 post_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL); 1188 if (!post_cycles_start) { 1189 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1190 return; 1191 } 1192 post_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL); 1193 if (!post_cycles_stop) { 1194 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1195 return; 1196 } 1197 poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL); 1198 if (!poll_cycles_start) { 1199 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1200 return; 1201 } 1202 poll_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL); 1203 if (!poll_cycles_stop) { 1204 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1205 return; 1206 } 1207 last_poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), 1208 GFP_KERNEL); 1209 if (!last_poll_cycles_start) { 1210 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1211 return; 1212 } 1213 cb->rdma_sq_wr.opcode = IB_WR_RDMA_WRITE; 1214 cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey; 1215 cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr; 1216 cb->rdma_sq_wr.sg_list->length = cb->size; 1217 1218 if (cycle_iters > iters) 1219 cycle_iters = iters; 1220 microtime(&start_tv); 1221 while (scnt < iters || ccnt < iters || rcnt < iters) { 1222 1223 /* Wait till buffer changes. */ 1224 if (rcnt < iters && !(scnt < 1 && !cb->server)) { 1225 ++rcnt; 1226 while (*poll_buf != (char)rcnt) { 1227 if (cb->state == ERROR) { 1228 PRINTF(cb, 1229 "state = ERROR, bailing\n"); 1230 return; 1231 } 1232 } 1233 } 1234 1235 if (scnt < iters) { 1236 struct ib_send_wr *bad_wr; 1237 1238 *buf = (char)scnt+1; 1239 if (scnt < cycle_iters) 1240 post_cycles_start[scnt] = get_cycles(); 1241 if (ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr)) { 1242 PRINTF(cb, 1243 "Couldn't post send: scnt=%d\n", 1244 scnt); 1245 return; 1246 } 1247 if (scnt < cycle_iters) 1248 post_cycles_stop[scnt] = get_cycles(); 1249 scnt++; 1250 } 1251 1252 if (ccnt < iters) { 1253 struct ib_wc wc; 1254 int ne; 1255 1256 if (ccnt < cycle_iters) 1257 poll_cycles_start[ccnt] = get_cycles(); 1258 do { 1259 if (ccnt < cycle_iters) 1260 last_poll_cycles_start[ccnt] = 1261 get_cycles(); 1262 ne = ib_poll_cq(cb->cq, 1, &wc); 1263 } while (ne == 0); 1264 if (ccnt < cycle_iters) 1265 poll_cycles_stop[ccnt] = get_cycles(); 1266 ++ccnt; 1267 1268 if (ne < 0) { 1269 PRINTF(cb, "poll CQ failed %d\n", ne); 1270 return; 1271 } 1272 if (wc.status != IB_WC_SUCCESS) { 1273 PRINTF(cb, 1274 "Completion wth error at %s:\n", 1275 cb->server ? "server" : "client"); 1276 PRINTF(cb, 1277 "Failed status %d: wr_id %d\n", 1278 wc.status, (int) wc.wr_id); 1279 PRINTF(cb, 1280 "scnt=%d, rcnt=%d, ccnt=%d\n", 1281 scnt, rcnt, ccnt); 1282 return; 1283 } 1284 } 1285 } 1286 microtime(&stop_tv); 1287 1288 if (stop_tv.tv_usec < start_tv.tv_usec) { 1289 stop_tv.tv_usec += 1000000; 1290 stop_tv.tv_sec -= 1; 1291 } 1292 1293 for (i=0; i < cycle_iters; i++) { 1294 sum_post += post_cycles_stop[i] - post_cycles_start[i]; 1295 sum_poll += poll_cycles_stop[i] - poll_cycles_start[i]; 1296 sum_last_poll += poll_cycles_stop[i]-last_poll_cycles_start[i]; 1297 } 1298 PRINTF(cb, 1299 "delta sec %lu delta usec %lu iter %d size %d cycle_iters %d" 1300 " sum_post %llu sum_poll %llu sum_last_poll %llu\n", 1301 stop_tv.tv_sec - start_tv.tv_sec, 1302 stop_tv.tv_usec - start_tv.tv_usec, 1303 scnt, cb->size, cycle_iters, 1304 (unsigned long long)sum_post, (unsigned long long)sum_poll, 1305 (unsigned long long)sum_last_poll); 1306 kfree(post_cycles_start); 1307 kfree(post_cycles_stop); 1308 kfree(poll_cycles_start); 1309 kfree(poll_cycles_stop); 1310 kfree(last_poll_cycles_start); 1311} 1312 1313static void bw_test(struct krping_cb *cb) 1314{ 1315 int ccnt, scnt, rcnt; 1316 int iters=cb->count; 1317 struct timeval start_tv, stop_tv; 1318 cycles_t *post_cycles_start, *post_cycles_stop; 1319 cycles_t *poll_cycles_start, *poll_cycles_stop; 1320 cycles_t *last_poll_cycles_start; 1321 cycles_t sum_poll = 0, sum_post = 0, sum_last_poll = 0; 1322 int i; 1323 int cycle_iters = 1000; 1324 1325 ccnt = 0; 1326 scnt = 0; 1327 rcnt = 0; 1328 1329 post_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL); 1330 if (!post_cycles_start) { 1331 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1332 return; 1333 } 1334 post_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL); 1335 if (!post_cycles_stop) { 1336 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1337 return; 1338 } 1339 poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL); 1340 if (!poll_cycles_start) { 1341 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1342 return; 1343 } 1344 poll_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL); 1345 if (!poll_cycles_stop) { 1346 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1347 return; 1348 } 1349 last_poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), 1350 GFP_KERNEL); 1351 if (!last_poll_cycles_start) { 1352 PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__); 1353 return; 1354 } 1355 cb->rdma_sq_wr.opcode = IB_WR_RDMA_WRITE; 1356 cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey; 1357 cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr; 1358 cb->rdma_sq_wr.sg_list->length = cb->size; 1359 1360 if (cycle_iters > iters) 1361 cycle_iters = iters; 1362 microtime(&start_tv); 1363 while (scnt < iters || ccnt < iters) { 1364 1365 while (scnt < iters && scnt - ccnt < cb->txdepth) { 1366 struct ib_send_wr *bad_wr; 1367 1368 if (scnt < cycle_iters) 1369 post_cycles_start[scnt] = get_cycles(); 1370 if (ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr)) { 1371 PRINTF(cb, 1372 "Couldn't post send: scnt=%d\n", 1373 scnt); 1374 return; 1375 } 1376 if (scnt < cycle_iters) 1377 post_cycles_stop[scnt] = get_cycles(); 1378 ++scnt; 1379 } 1380 1381 if (ccnt < iters) { 1382 int ne; 1383 struct ib_wc wc; 1384 1385 if (ccnt < cycle_iters) 1386 poll_cycles_start[ccnt] = get_cycles(); 1387 do { 1388 if (ccnt < cycle_iters) 1389 last_poll_cycles_start[ccnt] = 1390 get_cycles(); 1391 ne = ib_poll_cq(cb->cq, 1, &wc); 1392 } while (ne == 0); 1393 if (ccnt < cycle_iters) 1394 poll_cycles_stop[ccnt] = get_cycles(); 1395 ccnt += 1; 1396 1397 if (ne < 0) { 1398 PRINTF(cb, "poll CQ failed %d\n", ne); 1399 return; 1400 } 1401 if (wc.status != IB_WC_SUCCESS) { 1402 PRINTF(cb, 1403 "Completion wth error at %s:\n", 1404 cb->server ? "server" : "client"); 1405 PRINTF(cb, 1406 "Failed status %d: wr_id %d\n", 1407 wc.status, (int) wc.wr_id); 1408 return; 1409 } 1410 } 1411 } 1412 microtime(&stop_tv); 1413 1414 if (stop_tv.tv_usec < start_tv.tv_usec) { 1415 stop_tv.tv_usec += 1000000; 1416 stop_tv.tv_sec -= 1; 1417 } 1418 1419 for (i=0; i < cycle_iters; i++) { 1420 sum_post += post_cycles_stop[i] - post_cycles_start[i]; 1421 sum_poll += poll_cycles_stop[i] - poll_cycles_start[i]; 1422 sum_last_poll += poll_cycles_stop[i]-last_poll_cycles_start[i]; 1423 } 1424 PRINTF(cb, 1425 "delta sec %lu delta usec %lu iter %d size %d cycle_iters %d" 1426 " sum_post %llu sum_poll %llu sum_last_poll %llu\n", 1427 stop_tv.tv_sec - start_tv.tv_sec, 1428 stop_tv.tv_usec - start_tv.tv_usec, 1429 scnt, cb->size, cycle_iters, 1430 (unsigned long long)sum_post, (unsigned long long)sum_poll, 1431 (unsigned long long)sum_last_poll); 1432 kfree(post_cycles_start); 1433 kfree(post_cycles_stop); 1434 kfree(poll_cycles_start); 1435 kfree(poll_cycles_stop); 1436 kfree(last_poll_cycles_start); 1437} 1438 1439static void krping_rlat_test_server(struct krping_cb *cb) 1440{ 1441 struct ib_send_wr *bad_wr; 1442 struct ib_wc wc; 1443 int ret; 1444 1445 /* Spin waiting for client's Start STAG/TO/Len */ 1446 while (cb->state < RDMA_READ_ADV) { 1447 krping_cq_event_handler(cb->cq, cb); 1448 } 1449 1450 /* Send STAG/TO/Len to client */ 1451 krping_format_send(cb, cb->start_dma_addr); 1452 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1453 if (ret) { 1454 PRINTF(cb, "post send error %d\n", ret); 1455 return; 1456 } 1457 1458 /* Spin waiting for send completion */ 1459 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0)); 1460 if (ret < 0) { 1461 PRINTF(cb, "poll error %d\n", ret); 1462 return; 1463 } 1464 if (wc.status) { 1465 PRINTF(cb, "send completiong error %d\n", wc.status); 1466 return; 1467 } 1468 1469 wait_event_interruptible(cb->sem, cb->state == ERROR); 1470} 1471 1472static void krping_wlat_test_server(struct krping_cb *cb) 1473{ 1474 struct ib_send_wr *bad_wr; 1475 struct ib_wc wc; 1476 int ret; 1477 1478 /* Spin waiting for client's Start STAG/TO/Len */ 1479 while (cb->state < RDMA_READ_ADV) { 1480 krping_cq_event_handler(cb->cq, cb); 1481 } 1482 1483 /* Send STAG/TO/Len to client */ 1484 krping_format_send(cb, cb->start_dma_addr); 1485 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1486 if (ret) { 1487 PRINTF(cb, "post send error %d\n", ret); 1488 return; 1489 } 1490 1491 /* Spin waiting for send completion */ 1492 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0)); 1493 if (ret < 0) { 1494 PRINTF(cb, "poll error %d\n", ret); 1495 return; 1496 } 1497 if (wc.status) { 1498 PRINTF(cb, "send completiong error %d\n", wc.status); 1499 return; 1500 } 1501 1502 wlat_test(cb); 1503 wait_event_interruptible(cb->sem, cb->state == ERROR); 1504} 1505 1506static void krping_bw_test_server(struct krping_cb *cb) 1507{ 1508 struct ib_send_wr *bad_wr; 1509 struct ib_wc wc; 1510 int ret; 1511 1512 /* Spin waiting for client's Start STAG/TO/Len */ 1513 while (cb->state < RDMA_READ_ADV) { 1514 krping_cq_event_handler(cb->cq, cb); 1515 } 1516 1517 /* Send STAG/TO/Len to client */ 1518 krping_format_send(cb, cb->start_dma_addr); 1519 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1520 if (ret) { 1521 PRINTF(cb, "post send error %d\n", ret); 1522 return; 1523 } 1524 1525 /* Spin waiting for send completion */ 1526 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0)); 1527 if (ret < 0) { 1528 PRINTF(cb, "poll error %d\n", ret); 1529 return; 1530 } 1531 if (wc.status) { 1532 PRINTF(cb, "send completiong error %d\n", wc.status); 1533 return; 1534 } 1535 1536 if (cb->duplex) 1537 bw_test(cb); 1538 wait_event_interruptible(cb->sem, cb->state == ERROR); 1539} 1540 1541static int fastreg_supported(struct krping_cb *cb) 1542{ 1543 struct ib_device *dev = cb->child_cm_id->device; 1544 struct ib_device_attr attr; 1545 int ret; 1546 1547 ret = ib_query_device(dev, &attr); 1548 if (ret) { 1549 PRINTF(cb, "ib_query_device failed ret %d\n", ret); 1550 return 0; 1551 } 1552 if (!(attr.device_cap_flags & IB_DEVICE_MEM_MGT_EXTENSIONS)) { 1553 PRINTF(cb, "Fastreg not supported - device_cap_flags 0x%x\n", 1554 attr.device_cap_flags); 1555 return 0; 1556 } 1557 DEBUG_LOG(cb, "Fastreg supported - device_cap_flags 0x%x\n", 1558 attr.device_cap_flags); 1559 return 1; 1560} 1561 1562static int krping_bind_server(struct krping_cb *cb) 1563{ 1564 struct sockaddr_in sin; 1565 int ret; 1566 1567 memset(&sin, 0, sizeof(sin)); 1568 sin.sin_len = sizeof sin; 1569 sin.sin_family = AF_INET; 1570 sin.sin_addr.s_addr = cb->addr.s_addr; 1571 sin.sin_port = cb->port; 1572 1573 ret = rdma_bind_addr(cb->cm_id, (struct sockaddr *) &sin); 1574 if (ret) { 1575 PRINTF(cb, "rdma_bind_addr error %d\n", ret); 1576 return ret; 1577 } 1578 DEBUG_LOG(cb, "rdma_bind_addr successful\n"); 1579 1580 DEBUG_LOG(cb, "rdma_listen\n"); 1581 ret = rdma_listen(cb->cm_id, 3); 1582 if (ret) { 1583 PRINTF(cb, "rdma_listen failed: %d\n", ret); 1584 return ret; 1585 } 1586 1587 wait_event_interruptible(cb->sem, cb->state >= CONNECT_REQUEST); 1588 if (cb->state != CONNECT_REQUEST) { 1589 PRINTF(cb, "wait for CONNECT_REQUEST state %d\n", 1590 cb->state); 1591 return -1; 1592 } 1593 1594 if (cb->mem == FASTREG && !fastreg_supported(cb)) 1595 return -EINVAL; 1596 1597 return 0; 1598} 1599 1600static void krping_run_server(struct krping_cb *cb) 1601{ 1602 struct ib_recv_wr *bad_wr; 1603 int ret; 1604 1605 ret = krping_bind_server(cb); 1606 if (ret) 1607 return; 1608 1609 ret = krping_setup_qp(cb, cb->child_cm_id); 1610 if (ret) { 1611 PRINTF(cb, "setup_qp failed: %d\n", ret); 1612 goto err0; 1613 } 1614 1615 ret = krping_setup_buffers(cb); 1616 if (ret) { 1617 PRINTF(cb, "krping_setup_buffers failed: %d\n", ret); 1618 goto err1; 1619 } 1620 1621 ret = ib_post_recv(cb->qp, &cb->rq_wr, &bad_wr); 1622 if (ret) { 1623 PRINTF(cb, "ib_post_recv failed: %d\n", ret); 1624 goto err2; 1625 } 1626 1627 ret = krping_accept(cb); 1628 if (ret) { 1629 PRINTF(cb, "connect error %d\n", ret); 1630 goto err2; 1631 } 1632 1633 if (cb->wlat) 1634 krping_wlat_test_server(cb); 1635 else if (cb->rlat) 1636 krping_rlat_test_server(cb); 1637 else if (cb->bw) 1638 krping_bw_test_server(cb); 1639 else 1640 krping_test_server(cb); 1641 rdma_disconnect(cb->child_cm_id); 1642err2: 1643 krping_free_buffers(cb); 1644err1: 1645 krping_free_qp(cb); 1646err0: 1647 rdma_destroy_id(cb->child_cm_id); 1648} 1649 1650static void krping_test_client(struct krping_cb *cb) 1651{ 1652 int ping, start, cc, i, ret; 1653 struct ib_send_wr *bad_wr; 1654 unsigned char c; 1655 1656 start = 65; 1657 for (ping = 0; !cb->count || ping < cb->count; ping++) { 1658 cb->state = RDMA_READ_ADV; 1659 1660 /* Put some ascii text in the buffer. */ 1661 cc = sprintf(cb->start_buf, "rdma-ping-%d: ", ping); 1662 for (i = cc, c = start; i < cb->size; i++) { 1663 cb->start_buf[i] = c; 1664 c++; 1665 if (c > 122) 1666 c = 65; 1667 } 1668 start++; 1669 if (start > 122) 1670 start = 65; 1671 cb->start_buf[cb->size - 1] = 0; 1672 1673 krping_format_send(cb, cb->start_dma_addr); 1674 if (cb->state == ERROR) { 1675 PRINTF(cb, "krping_format_send failed\n"); 1676 break; 1677 } 1678 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1679 if (ret) { 1680 PRINTF(cb, "post send error %d\n", ret); 1681 break; 1682 } 1683 1684 /* Wait for server to ACK */ 1685 wait_event_interruptible(cb->sem, cb->state >= RDMA_WRITE_ADV); 1686 if (cb->state != RDMA_WRITE_ADV) { 1687 PRINTF(cb, 1688 "wait for RDMA_WRITE_ADV state %d\n", 1689 cb->state); 1690 break; 1691 } 1692 1693 krping_format_send(cb, cb->rdma_dma_addr); 1694 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1695 if (ret) { 1696 PRINTF(cb, "post send error %d\n", ret); 1697 break; 1698 } 1699 1700 /* Wait for the server to say the RDMA Write is complete. */ 1701 wait_event_interruptible(cb->sem, 1702 cb->state >= RDMA_WRITE_COMPLETE); 1703 if (cb->state != RDMA_WRITE_COMPLETE) { 1704 PRINTF(cb, 1705 "wait for RDMA_WRITE_COMPLETE state %d\n", 1706 cb->state); 1707 break; 1708 } 1709 1710 if (cb->validate) 1711 if (memcmp(cb->start_buf, cb->rdma_buf, cb->size)) { 1712 PRINTF(cb, "data mismatch!\n"); 1713 break; 1714 } 1715 1716 if (cb->verbose) 1717 PRINTF(cb, "ping data: %s\n", cb->rdma_buf); 1718#ifdef SLOW_KRPING 1719 wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ); 1720#endif 1721 } 1722} 1723 1724static void krping_rlat_test_client(struct krping_cb *cb) 1725{ 1726 struct ib_send_wr *bad_wr; 1727 struct ib_wc wc; 1728 int ret; 1729 1730 cb->state = RDMA_READ_ADV; 1731 1732 /* Send STAG/TO/Len to client */ 1733 krping_format_send(cb, cb->start_dma_addr); 1734 if (cb->state == ERROR) { 1735 PRINTF(cb, "krping_format_send failed\n"); 1736 return; 1737 } 1738 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1739 if (ret) { 1740 PRINTF(cb, "post send error %d\n", ret); 1741 return; 1742 } 1743 1744 /* Spin waiting for send completion */ 1745 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0)); 1746 if (ret < 0) { 1747 PRINTF(cb, "poll error %d\n", ret); 1748 return; 1749 } 1750 if (wc.status) { 1751 PRINTF(cb, "send completion error %d\n", wc.status); 1752 return; 1753 } 1754 1755 /* Spin waiting for server's Start STAG/TO/Len */ 1756 while (cb->state < RDMA_WRITE_ADV) { 1757 krping_cq_event_handler(cb->cq, cb); 1758 } 1759 1760#if 0 1761{ 1762 int i; 1763 struct timeval start, stop; 1764 time_t sec; 1765 suseconds_t usec; 1766 unsigned long long elapsed; 1767 struct ib_wc wc; 1768 struct ib_send_wr *bad_wr; 1769 int ne; 1770 1771 cb->rdma_sq_wr.opcode = IB_WR_RDMA_WRITE; 1772 cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey; 1773 cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr; 1774 cb->rdma_sq_wr.sg_list->length = 0; 1775 cb->rdma_sq_wr.num_sge = 0; 1776 1777 microtime(&start); 1778 for (i=0; i < 100000; i++) { 1779 if (ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr)) { 1780 PRINTF(cb, "Couldn't post send\n"); 1781 return; 1782 } 1783 do { 1784 ne = ib_poll_cq(cb->cq, 1, &wc); 1785 } while (ne == 0); 1786 if (ne < 0) { 1787 PRINTF(cb, "poll CQ failed %d\n", ne); 1788 return; 1789 } 1790 if (wc.status != IB_WC_SUCCESS) { 1791 PRINTF(cb, "Completion wth error at %s:\n", 1792 cb->server ? "server" : "client"); 1793 PRINTF(cb, "Failed status %d: wr_id %d\n", 1794 wc.status, (int) wc.wr_id); 1795 return; 1796 } 1797 } 1798 microtime(&stop); 1799 1800 if (stop.tv_usec < start.tv_usec) { 1801 stop.tv_usec += 1000000; 1802 stop.tv_sec -= 1; 1803 } 1804 sec = stop.tv_sec - start.tv_sec; 1805 usec = stop.tv_usec - start.tv_usec; 1806 elapsed = sec * 1000000 + usec; 1807 PRINTF(cb, "0B-write-lat iters 100000 usec %llu\n", elapsed); 1808} 1809#endif 1810 1811 rlat_test(cb); 1812} 1813 1814static void krping_wlat_test_client(struct krping_cb *cb) 1815{ 1816 struct ib_send_wr *bad_wr; 1817 struct ib_wc wc; 1818 int ret; 1819 1820 cb->state = RDMA_READ_ADV; 1821 1822 /* Send STAG/TO/Len to client */ 1823 krping_format_send(cb, cb->start_dma_addr); 1824 if (cb->state == ERROR) { 1825 PRINTF(cb, "krping_format_send failed\n"); 1826 return; 1827 } 1828 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1829 if (ret) { 1830 PRINTF(cb, "post send error %d\n", ret); 1831 return; 1832 } 1833 1834 /* Spin waiting for send completion */ 1835 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0)); 1836 if (ret < 0) { 1837 PRINTF(cb, "poll error %d\n", ret); 1838 return; 1839 } 1840 if (wc.status) { 1841 PRINTF(cb, "send completion error %d\n", wc.status); 1842 return; 1843 } 1844 1845 /* Spin waiting for server's Start STAG/TO/Len */ 1846 while (cb->state < RDMA_WRITE_ADV) { 1847 krping_cq_event_handler(cb->cq, cb); 1848 } 1849 1850 wlat_test(cb); 1851} 1852 1853static void krping_bw_test_client(struct krping_cb *cb) 1854{ 1855 struct ib_send_wr *bad_wr; 1856 struct ib_wc wc; 1857 int ret; 1858 1859 cb->state = RDMA_READ_ADV; 1860 1861 /* Send STAG/TO/Len to client */ 1862 krping_format_send(cb, cb->start_dma_addr); 1863 if (cb->state == ERROR) { 1864 PRINTF(cb, "krping_format_send failed\n"); 1865 return; 1866 } 1867 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr); 1868 if (ret) { 1869 PRINTF(cb, "post send error %d\n", ret); 1870 return; 1871 } 1872 1873 /* Spin waiting for send completion */ 1874 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0)); 1875 if (ret < 0) { 1876 PRINTF(cb, "poll error %d\n", ret); 1877 return; 1878 } 1879 if (wc.status) { 1880 PRINTF(cb, "send completion error %d\n", wc.status); 1881 return; 1882 } 1883 1884 /* Spin waiting for server's Start STAG/TO/Len */ 1885 while (cb->state < RDMA_WRITE_ADV) { 1886 krping_cq_event_handler(cb->cq, cb); 1887 } 1888 1889 bw_test(cb); 1890} 1891 1892static void krping_fr_test(struct krping_cb *cb) 1893{ 1894 struct ib_fast_reg_page_list *pl; 1895 struct ib_send_wr fr, inv, *bad; 1896 struct ib_wc wc; 1897 u8 key = 0; 1898 struct ib_mr *mr; 1899 int i; 1900 int ret; 1901 int size = cb->size; 1902 int plen = (((size - 1) & PAGE_MASK) + PAGE_SIZE) >> PAGE_SHIFT; 1903 time_t start; 1904 int count = 0; 1905 int scnt = 0; 1906 1907 pl = ib_alloc_fast_reg_page_list(cb->qp->device, plen); 1908 if (IS_ERR(pl)) { 1909 PRINTF(cb, "ib_alloc_fast_reg_page_list failed %ld\n", PTR_ERR(pl)); 1910 return; 1911 } 1912 1913 mr = ib_alloc_fast_reg_mr(cb->pd, plen); 1914 if (IS_ERR(mr)) { 1915 PRINTF(cb, "ib_alloc_fast_reg_mr failed %ld\n", PTR_ERR(pl)); 1916 goto err1; 1917 } 1918 1919 for (i=0; i<plen; i++) 1920 pl->page_list[i] = 0xcafebabe | i; 1921 1922 memset(&fr, 0, sizeof fr); 1923 fr.opcode = IB_WR_FAST_REG_MR; 1924 fr.wr.fast_reg.page_shift = PAGE_SHIFT; 1925 fr.wr.fast_reg.length = size; 1926 fr.wr.fast_reg.page_list = pl; 1927 fr.wr.fast_reg.page_list_len = plen; 1928 fr.wr.fast_reg.iova_start = 0; 1929 fr.wr.fast_reg.access_flags = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE; 1930 fr.next = &inv; 1931 memset(&inv, 0, sizeof inv); 1932 inv.opcode = IB_WR_LOCAL_INV; 1933 inv.send_flags = IB_SEND_SIGNALED; 1934 1935 DEBUG_LOG(cb, "fr_test: stag index 0x%x plen %u size %u depth %u\n", mr->rkey >> 8, plen, cb->size, cb->txdepth); 1936 start = time_uptime; 1937 while (1) { 1938 if ((time_uptime - start) >= 9) { 1939 DEBUG_LOG(cb, "fr_test: pausing 1 second! count %u latest size %u plen %u\n", count, size, plen); 1940 wait_event_interruptible(cb->sem, cb->state == ERROR); 1941 if (cb->state == ERROR) 1942 break; 1943 start = time_uptime; 1944 } 1945 while (scnt < (cb->txdepth>>1)) { 1946 ib_update_fast_reg_key(mr, ++key); 1947 fr.wr.fast_reg.rkey = mr->rkey; 1948 inv.ex.invalidate_rkey = mr->rkey; 1949 size = arc4random() % cb->size; 1950 if (size == 0) 1951 size = cb->size; 1952 plen = (((size - 1) & PAGE_MASK) + PAGE_SIZE) >> PAGE_SHIFT; 1953 fr.wr.fast_reg.length = size; 1954 fr.wr.fast_reg.page_list_len = plen; 1955 ret = ib_post_send(cb->qp, &fr, &bad); 1956 if (ret) { 1957 PRINTF(cb, "ib_post_send failed %d\n", ret); 1958 goto err2; 1959 } 1960 scnt++; 1961 } 1962 1963 do { 1964 ret = ib_poll_cq(cb->cq, 1, &wc); 1965 if (ret < 0) { 1966 PRINTF(cb, "ib_poll_cq failed %d\n", ret); 1967 goto err2; 1968 } 1969 if (ret == 1) { 1970 if (wc.status) { 1971 PRINTF(cb, "completion error %u\n", wc.status); 1972 goto err2; 1973 } 1974 count++; 1975 scnt--; 1976 } 1977 else if (krping_sigpending()) { 1978 PRINTF(cb, "signal!\n"); 1979 goto err2; 1980 } 1981 } while (ret == 1); 1982 } 1983err2: 1984#if 0 1985 DEBUG_LOG(cb, "sleeping 1 second\n"); 1986 wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ); 1987#endif 1988 DEBUG_LOG(cb, "draining the cq...\n"); 1989 do { 1990 ret = ib_poll_cq(cb->cq, 1, &wc); 1991 if (ret < 0) { 1992 PRINTF(cb, "ib_poll_cq failed %d\n", ret); 1993 break; 1994 } 1995 if (ret == 1) { 1996 if (wc.status) { 1997 PRINTF(cb, "completion error %u opcode %u\n", wc.status, wc.opcode); 1998 } 1999 } 2000 } while (ret == 1); 2001 DEBUG_LOG(cb, "fr_test: done!\n"); 2002 ib_dereg_mr(mr); 2003err1: 2004 ib_free_fast_reg_page_list(pl); 2005} 2006 2007static int krping_connect_client(struct krping_cb *cb) 2008{ 2009 struct rdma_conn_param conn_param; 2010 int ret; 2011 2012 memset(&conn_param, 0, sizeof conn_param); 2013 conn_param.responder_resources = 1; 2014 conn_param.initiator_depth = 1; 2015 conn_param.retry_count = 10; 2016 2017 ret = rdma_connect(cb->cm_id, &conn_param); 2018 if (ret) { 2019 PRINTF(cb, "rdma_connect error %d\n", ret); 2020 return ret; 2021 } 2022 2023 wait_event_interruptible(cb->sem, cb->state >= CONNECTED); 2024 if (cb->state == ERROR) { 2025 PRINTF(cb, "wait for CONNECTED state %d\n", cb->state); 2026 return -1; 2027 } 2028 2029 DEBUG_LOG(cb, "rdma_connect successful\n"); 2030 return 0; 2031} 2032 2033static int krping_bind_client(struct krping_cb *cb) 2034{ 2035 struct sockaddr_in sin; 2036 int ret; 2037 2038 memset(&sin, 0, sizeof(sin)); 2039 sin.sin_len = sizeof sin; 2040 sin.sin_family = AF_INET; 2041 sin.sin_addr.s_addr = cb->addr.s_addr; 2042 sin.sin_port = cb->port; 2043 2044 ret = rdma_resolve_addr(cb->cm_id, NULL, (struct sockaddr *) &sin, 2045 2000); 2046 if (ret) { 2047 PRINTF(cb, "rdma_resolve_addr error %d\n", ret); 2048 return ret; 2049 } 2050 2051 wait_event_interruptible(cb->sem, cb->state >= ROUTE_RESOLVED); 2052 if (cb->state != ROUTE_RESOLVED) { 2053 PRINTF(cb, 2054 "addr/route resolution did not resolve: state %d\n", 2055 cb->state); 2056 return -EINTR; 2057 } 2058 2059 if (cb->mem == FASTREG && !fastreg_supported(cb)) 2060 return -EINVAL; 2061 2062 DEBUG_LOG(cb, "rdma_resolve_addr - rdma_resolve_route successful\n"); 2063 return 0; 2064} 2065 2066static void krping_run_client(struct krping_cb *cb) 2067{ 2068 struct ib_recv_wr *bad_wr; 2069 int ret; 2070 2071 ret = krping_bind_client(cb); 2072 if (ret) 2073 return; 2074 2075 ret = krping_setup_qp(cb, cb->cm_id); 2076 if (ret) { 2077 PRINTF(cb, "setup_qp failed: %d\n", ret); 2078 return; 2079 } 2080 2081 ret = krping_setup_buffers(cb); 2082 if (ret) { 2083 PRINTF(cb, "krping_setup_buffers failed: %d\n", ret); 2084 goto err1; 2085 } 2086 2087 ret = ib_post_recv(cb->qp, &cb->rq_wr, &bad_wr); 2088 if (ret) { 2089 PRINTF(cb, "ib_post_recv failed: %d\n", ret); 2090 goto err2; 2091 } 2092 2093 ret = krping_connect_client(cb); 2094 if (ret) { 2095 PRINTF(cb, "connect error %d\n", ret); 2096 goto err2; 2097 } 2098 2099 if (cb->wlat) 2100 krping_wlat_test_client(cb); 2101 else if (cb->rlat) 2102 krping_rlat_test_client(cb); 2103 else if (cb->bw) 2104 krping_bw_test_client(cb); 2105 else if (cb->frtest) 2106 krping_fr_test(cb); 2107 else 2108 krping_test_client(cb); 2109 rdma_disconnect(cb->cm_id); 2110err2: 2111 krping_free_buffers(cb); 2112err1: 2113 krping_free_qp(cb); 2114} 2115 2116int krping_doit(char *cmd, void *cookie) 2117{ 2118 struct krping_cb *cb; 2119 int op; 2120 int ret = 0; 2121 char *optarg; 2122 unsigned long optint; 2123 2124 cb = kzalloc(sizeof(*cb), GFP_KERNEL); 2125 if (!cb) 2126 return -ENOMEM; 2127 2128 mutex_lock(&krping_mutex); 2129 list_add_tail(&cb->list, &krping_cbs); 2130 mutex_unlock(&krping_mutex); 2131 2132 cb->cookie = cookie; 2133 cb->server = -1; 2134 cb->state = IDLE; 2135 cb->size = 64; 2136 cb->txdepth = RPING_SQ_DEPTH; 2137 cb->mem = DMA; 2138 init_waitqueue_head(&cb->sem); 2139 2140 while ((op = krping_getopt("krping", &cmd, krping_opts, NULL, &optarg, 2141 &optint)) != 0) { 2142 switch (op) { 2143 case 'a': 2144 cb->addr_str = optarg; 2145 DEBUG_LOG(cb, "ipaddr (%s)\n", optarg); 2146 if (!inet_aton(optarg, &cb->addr)) { 2147 PRINTF(cb, "bad addr string %s\n", 2148 optarg); 2149 ret = EINVAL; 2150 } 2151 break; 2152 case 'p': 2153 cb->port = htons(optint); 2154 DEBUG_LOG(cb, "port %d\n", (int)optint); 2155 break; 2156 case 'P': 2157 cb->poll = 1; 2158 DEBUG_LOG(cb, "server\n"); 2159 break; 2160 case 's': 2161 cb->server = 1; 2162 DEBUG_LOG(cb, "server\n"); 2163 break; 2164 case 'c': 2165 cb->server = 0; 2166 DEBUG_LOG(cb, "client\n"); 2167 break; 2168 case 'S': 2169 cb->size = optint; 2170 if ((cb->size < 1) || 2171 (cb->size > RPING_BUFSIZE)) { 2172 PRINTF(cb, "Invalid size %d " 2173 "(valid range is 1 to %d)\n", 2174 cb->size, RPING_BUFSIZE); 2175 ret = EINVAL; 2176 } else 2177 DEBUG_LOG(cb, "size %d\n", (int)optint); 2178 break; 2179 case 'C': 2180 cb->count = optint; 2181 if (cb->count < 0) { 2182 PRINTF(cb, "Invalid count %d\n", 2183 cb->count); 2184 ret = EINVAL; 2185 } else 2186 DEBUG_LOG(cb, "count %d\n", (int) cb->count); 2187 break; 2188 case 'v': 2189 cb->verbose++; 2190 DEBUG_LOG(cb, "verbose\n"); 2191 break; 2192 case 'V': 2193 cb->validate++; 2194 DEBUG_LOG(cb, "validate data\n"); 2195 break; 2196 case 'l': 2197 cb->wlat++; 2198 break; 2199 case 'L': 2200 cb->rlat++; 2201 break; 2202 case 'B': 2203 cb->bw++; 2204 break; 2205 case 'd': 2206 cb->duplex++; 2207 break; 2208 case 'm': 2209 if (!strncmp(optarg, "dma", 3)) 2210 cb->mem = DMA; 2211 else if (!strncmp(optarg, "fastreg", 7)) 2212 cb->mem = FASTREG; 2213 else if (!strncmp(optarg, "mw", 2)) 2214 cb->mem = MW; 2215 else if (!strncmp(optarg, "mr", 2)) 2216 cb->mem = MR; 2217 else { 2218 PRINTF(cb, "unknown mem mode %s. " 2219 "Must be dma, fastreg, mw, or mr\n", 2220 optarg); 2221 ret = -EINVAL; 2222 break; 2223 } 2224 break; 2225 case 'I': 2226 cb->server_invalidate = 1; 2227 break; 2228 case 'T': 2229 cb->txdepth = optint; 2230 DEBUG_LOG(cb, "txdepth %d\n", (int) cb->txdepth); 2231 break; 2232 case 'Z': 2233 cb->local_dma_lkey = 1; 2234 DEBUG_LOG(cb, "using local dma lkey\n"); 2235 break; 2236 case 'R': 2237 cb->read_inv = 1; 2238 DEBUG_LOG(cb, "using read-with-inv\n"); 2239 break; 2240 case 'f': 2241 cb->frtest = 1; 2242 DEBUG_LOG(cb, "fast-reg test!\n"); 2243 break; 2244 default: 2245 PRINTF(cb, "unknown opt %s\n", optarg); 2246 ret = -EINVAL; 2247 break; 2248 } 2249 } 2250 if (ret) 2251 goto out; 2252 2253 if (cb->server == -1) { 2254 PRINTF(cb, "must be either client or server\n"); 2255 ret = -EINVAL; 2256 goto out; 2257 } 2258 2259 if (cb->server && cb->frtest) { 2260 PRINTF(cb, "must be client to run frtest\n"); 2261 ret = -EINVAL; 2262 goto out; 2263 } 2264 2265 if ((cb->frtest + cb->bw + cb->rlat + cb->wlat) > 1) { 2266 PRINTF(cb, "Pick only one test: fr, bw, rlat, wlat\n"); 2267 ret = -EINVAL; 2268 goto out; 2269 } 2270 2271 if (cb->server_invalidate && cb->mem != FASTREG) { 2272 PRINTF(cb, "server_invalidate only valid with fastreg mem_mode\n"); 2273 ret = -EINVAL; 2274 goto out; 2275 } 2276 2277 if (cb->read_inv && cb->mem != FASTREG) { 2278 PRINTF(cb, "read_inv only valid with fastreg mem_mode\n"); 2279 ret = -EINVAL; 2280 goto out; 2281 } 2282 2283 if (cb->mem != MR && (cb->wlat || cb->rlat || cb->bw)) { 2284 PRINTF(cb, "wlat, rlat, and bw tests only support mem_mode MR\n"); 2285 ret = -EINVAL; 2286 goto out; 2287 } 2288 2289 cb->cm_id = rdma_create_id(krping_cma_event_handler, cb, RDMA_PS_TCP); 2290 if (IS_ERR(cb->cm_id)) { 2291 ret = PTR_ERR(cb->cm_id); 2292 PRINTF(cb, "rdma_create_id error %d\n", ret); 2293 goto out; 2294 } 2295 DEBUG_LOG(cb, "created cm_id %p\n", cb->cm_id); 2296 2297 if (cb->server) 2298 krping_run_server(cb); 2299 else 2300 krping_run_client(cb); 2301 2302 DEBUG_LOG(cb, "destroy cm_id %p\n", cb->cm_id); 2303 rdma_destroy_id(cb->cm_id); 2304out: 2305 mutex_lock(&krping_mutex); 2306 list_del(&cb->list); 2307 mutex_unlock(&krping_mutex); 2308 kfree(cb); 2309 return ret; 2310} 2311 2312void 2313krping_walk_cb_list(void (*f)(struct krping_stats *, void *), void *arg) 2314{ 2315 struct krping_cb *cb; 2316 2317 mutex_lock(&krping_mutex); 2318 list_for_each_entry(cb, &krping_cbs, list) 2319 (*f)(cb->pd ? &cb->stats : NULL, arg); 2320 mutex_unlock(&krping_mutex); 2321} 2322 2323void krping_init(void) 2324{ 2325 2326 mutex_init(&krping_mutex); 2327}
|