1 /*
2 * Copyright (c) 2005 Ammasso, Inc. All rights reserved.
3 * Copyright (c) 2006-2009 Open Grid Computing, Inc. All rights reserved.
4 *
5 * This software is available to you under a choice of one of two
6 * licenses. You may choose to be licensed under the terms of the GNU
7 * General Public License (GPL) Version 2, available from the file
8 * COPYING in the main directory of this source tree, or the
9 * OpenIB.org BSD license below:
10 *
11 * Redistribution and use in source and binary forms, with or
12 * without modification, are permitted provided that the following
13 * conditions are met:
14 *
15 * - Redistributions of source code must retain the above
16 * copyright notice, this list of conditions and the following
17 * disclaimer.
18 *
19 * - Redistributions in binary form must reproduce the above
20 * copyright notice, this list of conditions and the following
21 * disclaimer in the documentation and/or other materials
22 * provided with the distribution.
23 *
24 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
25 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
26 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
27 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
28 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
29 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
30 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
31 * SOFTWARE.
32 */
33
34 #include <sys/cdefs.h>
35 __FBSDID("$FreeBSD$");
36
37 #include <linux/module.h>
38 #include <linux/moduleparam.h>
39 #include <linux/slab.h>
40 #include <linux/err.h>
41 #include <linux/string.h>
42 #include <linux/list.h>
43 #include <linux/in.h>
44 #include <linux/device.h>
45 #include <linux/pci.h>
46 #include <linux/sched.h>
47 #include <linux/wait.h>
48
49 #include <asm/atomic.h>
50
51 #include <rdma/ib_verbs.h>
52 #include <rdma/rdma_cm.h>
53
54 #include "krping.h"
55 #include "getopt.h"
56
57 #define PFX "krping: "
58
59 extern int krping_debug;
60 #define DEBUG_LOG(...) do { if (krping_debug) log(LOG_INFO, __VA_ARGS__); } while (0)
61 #define BIND_INFO 1
62
63 MODULE_AUTHOR("Steve Wise");
64 MODULE_DESCRIPTION("RDMA ping server");
65 MODULE_LICENSE("Dual BSD/GPL");
66 MODULE_VERSION(krping, 1);
67 MODULE_DEPEND(krping, linuxkpi, 1, 1, 1);
68
69 static __inline uint64_t
get_cycles(void)70 get_cycles(void)
71 {
72 uint32_t low, high;
73 __asm __volatile("rdtsc" : "=a" (low), "=d" (high));
74 return (low | ((u_int64_t)high << 32));
75 }
76
77 typedef uint64_t cycles_t;
78
79 enum mem_type {
80 DMA = 1,
81 REG = 2,
82 };
83
84 static const struct krping_option krping_opts[] = {
85 {"count", OPT_INT, 'C'},
86 {"size", OPT_INT, 'S'},
87 {"addr", OPT_STRING, 'a'},
88 {"addr6", OPT_STRING, 'A'},
89 {"port", OPT_INT, 'p'},
90 {"verbose", OPT_NOPARAM, 'v'},
91 {"validate", OPT_NOPARAM, 'V'},
92 {"server", OPT_NOPARAM, 's'},
93 {"client", OPT_NOPARAM, 'c'},
94 {"server_inv", OPT_NOPARAM, 'I'},
95 {"wlat", OPT_NOPARAM, 'l'},
96 {"rlat", OPT_NOPARAM, 'L'},
97 {"bw", OPT_NOPARAM, 'B'},
98 {"duplex", OPT_NOPARAM, 'd'},
99 {"tos", OPT_INT, 't'},
100 {"txdepth", OPT_INT, 'T'},
101 {"poll", OPT_NOPARAM, 'P'},
102 {"local_dma_lkey", OPT_NOPARAM, 'Z'},
103 {"read_inv", OPT_NOPARAM, 'R'},
104 {"fr", OPT_NOPARAM, 'f'},
105 {NULL, 0, 0}
106 };
107
108 #define htonll(x) cpu_to_be64((x))
109 #define ntohll(x) cpu_to_be64((x))
110
111 static DEFINE_MUTEX(krping_mutex);
112
113 /*
114 * List of running krping threads.
115 */
116 static LIST_HEAD(krping_cbs);
117
118 /*
119 * Invoke like this, one on each side, using the server's address on
120 * the RDMA device (iw%d):
121 *
122 * /bin/echo server,port=9999,addr=192.168.69.142,validate > /proc/krping
123 * /bin/echo client,port=9999,addr=192.168.69.142,validate > /proc/krping
124 * /bin/echo client,port=9999,addr6=2001:db8:0:f101::1,validate > /proc/krping
125 *
126 * krping "ping/pong" loop:
127 * client sends source rkey/addr/len
128 * server receives source rkey/add/len
129 * server rdma reads "ping" data from source
130 * server sends "go ahead" on rdma read completion
131 * client sends sink rkey/addr/len
132 * server receives sink rkey/addr/len
133 * server rdma writes "pong" data to sink
134 * server sends "go ahead" on rdma write completion
135 * <repeat loop>
136 */
137
138 /*
139 * These states are used to signal events between the completion handler
140 * and the main client or server thread.
141 *
142 * Once CONNECTED, they cycle through RDMA_READ_ADV, RDMA_WRITE_ADV,
143 * and RDMA_WRITE_COMPLETE for each ping.
144 */
145 enum test_state {
146 IDLE = 1,
147 CONNECT_REQUEST,
148 ADDR_RESOLVED,
149 ROUTE_RESOLVED,
150 CONNECTED,
151 RDMA_READ_ADV,
152 RDMA_READ_COMPLETE,
153 RDMA_WRITE_ADV,
154 RDMA_WRITE_COMPLETE,
155 ERROR
156 };
157
158 struct krping_rdma_info {
159 uint64_t buf;
160 uint32_t rkey;
161 uint32_t size;
162 };
163
164 /*
165 * Default max buffer size for IO...
166 */
167 #define RPING_BUFSIZE 128*1024
168 #define RPING_SQ_DEPTH 64
169
170 /*
171 * Control block struct.
172 */
173 struct krping_cb {
174 int server; /* 0 iff client */
175 struct ib_cq *cq;
176 struct ib_pd *pd;
177 struct ib_qp *qp;
178
179 struct ib_mr *dma_mr;
180
181 struct ib_fast_reg_page_list *page_list;
182 int page_list_len;
183 struct ib_reg_wr reg_mr_wr;
184 struct ib_send_wr invalidate_wr;
185 struct ib_mr *reg_mr;
186 int server_invalidate;
187 int read_inv;
188 u8 key;
189
190 struct ib_recv_wr rq_wr; /* recv work request record */
191 struct ib_sge recv_sgl; /* recv single SGE */
192 struct krping_rdma_info recv_buf __aligned(16); /* malloc'd buffer */
193 u64 recv_dma_addr;
194 DECLARE_PCI_UNMAP_ADDR(recv_mapping)
195
196 struct ib_send_wr sq_wr; /* send work requrest record */
197 struct ib_sge send_sgl;
198 struct krping_rdma_info send_buf __aligned(16); /* single send buf */
199 u64 send_dma_addr;
200 DECLARE_PCI_UNMAP_ADDR(send_mapping)
201
202 struct ib_rdma_wr rdma_sq_wr; /* rdma work request record */
203 struct ib_sge rdma_sgl; /* rdma single SGE */
204 char *rdma_buf; /* used as rdma sink */
205 u64 rdma_dma_addr;
206 DECLARE_PCI_UNMAP_ADDR(rdma_mapping)
207 struct ib_mr *rdma_mr;
208
209 uint32_t remote_rkey; /* remote guys RKEY */
210 uint64_t remote_addr; /* remote guys TO */
211 uint32_t remote_len; /* remote guys LEN */
212
213 char *start_buf; /* rdma read src */
214 u64 start_dma_addr;
215 DECLARE_PCI_UNMAP_ADDR(start_mapping)
216 struct ib_mr *start_mr;
217
218 enum test_state state; /* used for cond/signalling */
219 wait_queue_head_t sem;
220 struct krping_stats stats;
221
222 uint16_t port; /* dst port in NBO */
223 u8 addr[16] __aligned(8); /* dst addr in NBO */
224 char *addr_str; /* dst addr string */
225 uint8_t addr_type; /* ADDR_FAMILY - IPv4/V6 */
226 int verbose; /* verbose logging */
227 int count; /* ping count */
228 int size; /* ping data size */
229 int validate; /* validate ping data */
230 int wlat; /* run wlat test */
231 int rlat; /* run rlat test */
232 int bw; /* run bw test */
233 int duplex; /* run bw full duplex test */
234 int poll; /* poll or block for rlat test */
235 int txdepth; /* SQ depth */
236 int local_dma_lkey; /* use 0 for lkey */
237 int frtest; /* reg test */
238 int tos; /* type of service */
239
240 /* CM stuff */
241 struct rdma_cm_id *cm_id; /* connection on client side,*/
242 /* listener on server side. */
243 struct rdma_cm_id *child_cm_id; /* connection on server side */
244 struct list_head list;
245 };
246
krping_cma_event_handler(struct rdma_cm_id * cma_id,struct rdma_cm_event * event)247 static int krping_cma_event_handler(struct rdma_cm_id *cma_id,
248 struct rdma_cm_event *event)
249 {
250 int ret;
251 struct krping_cb *cb = cma_id->context;
252
253 DEBUG_LOG("cma_event type %d cma_id %p (%s)\n", event->event, cma_id,
254 (cma_id == cb->cm_id) ? "parent" : "child");
255
256 switch (event->event) {
257 case RDMA_CM_EVENT_ADDR_RESOLVED:
258 cb->state = ADDR_RESOLVED;
259 ret = rdma_resolve_route(cma_id, 2000);
260 if (ret) {
261 printk(KERN_ERR PFX "rdma_resolve_route error %d\n",
262 ret);
263 wake_up_interruptible(&cb->sem);
264 }
265 break;
266
267 case RDMA_CM_EVENT_ROUTE_RESOLVED:
268 cb->state = ROUTE_RESOLVED;
269 wake_up_interruptible(&cb->sem);
270 break;
271
272 case RDMA_CM_EVENT_CONNECT_REQUEST:
273 cb->state = CONNECT_REQUEST;
274 cb->child_cm_id = cma_id;
275 DEBUG_LOG("child cma %p\n", cb->child_cm_id);
276 wake_up_interruptible(&cb->sem);
277 break;
278
279 case RDMA_CM_EVENT_ESTABLISHED:
280 DEBUG_LOG("ESTABLISHED\n");
281 if (!cb->server) {
282 cb->state = CONNECTED;
283 }
284 wake_up_interruptible(&cb->sem);
285 break;
286
287 case RDMA_CM_EVENT_ADDR_ERROR:
288 case RDMA_CM_EVENT_ROUTE_ERROR:
289 case RDMA_CM_EVENT_CONNECT_ERROR:
290 case RDMA_CM_EVENT_UNREACHABLE:
291 case RDMA_CM_EVENT_REJECTED:
292 printk(KERN_ERR PFX "cma event %d, error %d\n", event->event,
293 event->status);
294 cb->state = ERROR;
295 wake_up_interruptible(&cb->sem);
296 break;
297
298 case RDMA_CM_EVENT_DISCONNECTED:
299 printk(KERN_ERR PFX "DISCONNECT EVENT...\n");
300 cb->state = ERROR;
301 wake_up_interruptible(&cb->sem);
302 break;
303
304 case RDMA_CM_EVENT_DEVICE_REMOVAL:
305 printk(KERN_ERR PFX "cma detected device removal!!!!\n");
306 cb->state = ERROR;
307 wake_up_interruptible(&cb->sem);
308 break;
309
310 default:
311 printk(KERN_ERR PFX "oof bad type!\n");
312 wake_up_interruptible(&cb->sem);
313 break;
314 }
315 return 0;
316 }
317
server_recv(struct krping_cb * cb,struct ib_wc * wc)318 static int server_recv(struct krping_cb *cb, struct ib_wc *wc)
319 {
320 if (wc->byte_len != sizeof(cb->recv_buf)) {
321 printk(KERN_ERR PFX "Received bogus data, size %d\n",
322 wc->byte_len);
323 return -1;
324 }
325
326 cb->remote_rkey = ntohl(cb->recv_buf.rkey);
327 cb->remote_addr = ntohll(cb->recv_buf.buf);
328 cb->remote_len = ntohl(cb->recv_buf.size);
329 DEBUG_LOG("Received rkey %x addr %llx len %d from peer\n",
330 cb->remote_rkey, (unsigned long long)cb->remote_addr,
331 cb->remote_len);
332
333 if (cb->state <= CONNECTED || cb->state == RDMA_WRITE_COMPLETE)
334 cb->state = RDMA_READ_ADV;
335 else
336 cb->state = RDMA_WRITE_ADV;
337
338 return 0;
339 }
340
client_recv(struct krping_cb * cb,struct ib_wc * wc)341 static int client_recv(struct krping_cb *cb, struct ib_wc *wc)
342 {
343 if (wc->byte_len != sizeof(cb->recv_buf)) {
344 printk(KERN_ERR PFX "Received bogus data, size %d\n",
345 wc->byte_len);
346 return -1;
347 }
348
349 if (cb->state == RDMA_READ_ADV)
350 cb->state = RDMA_WRITE_ADV;
351 else
352 cb->state = RDMA_WRITE_COMPLETE;
353
354 return 0;
355 }
356
krping_cq_event_handler(struct ib_cq * cq,void * ctx)357 static void krping_cq_event_handler(struct ib_cq *cq, void *ctx)
358 {
359 struct krping_cb *cb = ctx;
360 struct ib_wc wc;
361 struct ib_recv_wr *bad_wr;
362 int ret;
363
364 BUG_ON(cb->cq != cq);
365 if (cb->state == ERROR) {
366 printk(KERN_ERR PFX "cq completion in ERROR state\n");
367 return;
368 }
369 if (cb->frtest) {
370 printk(KERN_ERR PFX "cq completion event in frtest!\n");
371 return;
372 }
373 if (!cb->wlat && !cb->rlat && !cb->bw)
374 ib_req_notify_cq(cb->cq, IB_CQ_NEXT_COMP);
375 while ((ret = ib_poll_cq(cb->cq, 1, &wc)) == 1) {
376 if (wc.status) {
377 if (wc.status == IB_WC_WR_FLUSH_ERR) {
378 DEBUG_LOG("cq flushed\n");
379 continue;
380 } else {
381 printk(KERN_ERR PFX "cq completion failed with "
382 "wr_id %jx status %d opcode %d vender_err %x\n",
383 (uintmax_t)wc.wr_id, wc.status, wc.opcode, wc.vendor_err);
384 goto error;
385 }
386 }
387
388 switch (wc.opcode) {
389 case IB_WC_SEND:
390 DEBUG_LOG("send completion\n");
391 cb->stats.send_bytes += cb->send_sgl.length;
392 cb->stats.send_msgs++;
393 break;
394
395 case IB_WC_RDMA_WRITE:
396 DEBUG_LOG("rdma write completion\n");
397 cb->stats.write_bytes += cb->rdma_sq_wr.wr.sg_list->length;
398 cb->stats.write_msgs++;
399 cb->state = RDMA_WRITE_COMPLETE;
400 wake_up_interruptible(&cb->sem);
401 break;
402
403 case IB_WC_RDMA_READ:
404 DEBUG_LOG("rdma read completion\n");
405 cb->stats.read_bytes += cb->rdma_sq_wr.wr.sg_list->length;
406 cb->stats.read_msgs++;
407 cb->state = RDMA_READ_COMPLETE;
408 wake_up_interruptible(&cb->sem);
409 break;
410
411 case IB_WC_RECV:
412 DEBUG_LOG("recv completion\n");
413 cb->stats.recv_bytes += sizeof(cb->recv_buf);
414 cb->stats.recv_msgs++;
415 if (cb->wlat || cb->rlat || cb->bw)
416 ret = server_recv(cb, &wc);
417 else
418 ret = cb->server ? server_recv(cb, &wc) :
419 client_recv(cb, &wc);
420 if (ret) {
421 printk(KERN_ERR PFX "recv wc error: %d\n", ret);
422 goto error;
423 }
424
425 ret = ib_post_recv(cb->qp, &cb->rq_wr, &bad_wr);
426 if (ret) {
427 printk(KERN_ERR PFX "post recv error: %d\n",
428 ret);
429 goto error;
430 }
431 wake_up_interruptible(&cb->sem);
432 break;
433
434 default:
435 printk(KERN_ERR PFX
436 "%s:%d Unexpected opcode %d, Shutting down\n",
437 __func__, __LINE__, wc.opcode);
438 goto error;
439 }
440 }
441 if (ret) {
442 printk(KERN_ERR PFX "poll error %d\n", ret);
443 goto error;
444 }
445 return;
446 error:
447 cb->state = ERROR;
448 wake_up_interruptible(&cb->sem);
449 }
450
krping_accept(struct krping_cb * cb)451 static int krping_accept(struct krping_cb *cb)
452 {
453 struct rdma_conn_param conn_param;
454 int ret;
455
456 DEBUG_LOG("accepting client connection request\n");
457
458 memset(&conn_param, 0, sizeof conn_param);
459 conn_param.responder_resources = 1;
460 conn_param.initiator_depth = 1;
461
462 ret = rdma_accept(cb->child_cm_id, &conn_param);
463 if (ret) {
464 printk(KERN_ERR PFX "rdma_accept error: %d\n", ret);
465 return ret;
466 }
467
468 if (!cb->wlat && !cb->rlat && !cb->bw) {
469 wait_event_interruptible(cb->sem, cb->state >= CONNECTED);
470 if (cb->state == ERROR) {
471 printk(KERN_ERR PFX "wait for CONNECTED state %d\n",
472 cb->state);
473 return -1;
474 }
475 }
476 return 0;
477 }
478
krping_setup_wr(struct krping_cb * cb)479 static void krping_setup_wr(struct krping_cb *cb)
480 {
481 cb->recv_sgl.addr = cb->recv_dma_addr;
482 cb->recv_sgl.length = sizeof cb->recv_buf;
483 cb->recv_sgl.lkey = cb->pd->local_dma_lkey;
484 cb->rq_wr.sg_list = &cb->recv_sgl;
485 cb->rq_wr.num_sge = 1;
486
487 cb->send_sgl.addr = cb->send_dma_addr;
488 cb->send_sgl.length = sizeof cb->send_buf;
489 cb->send_sgl.lkey = cb->pd->local_dma_lkey;
490
491 cb->sq_wr.opcode = IB_WR_SEND;
492 cb->sq_wr.send_flags = IB_SEND_SIGNALED;
493 cb->sq_wr.sg_list = &cb->send_sgl;
494 cb->sq_wr.num_sge = 1;
495
496 if (cb->server || cb->wlat || cb->rlat || cb->bw) {
497 cb->rdma_sgl.addr = cb->rdma_dma_addr;
498 cb->rdma_sq_wr.wr.send_flags = IB_SEND_SIGNALED;
499 cb->rdma_sq_wr.wr.sg_list = &cb->rdma_sgl;
500 cb->rdma_sq_wr.wr.num_sge = 1;
501 }
502
503 /*
504 * A chain of 2 WRs, INVALDATE_MR + REG_MR.
505 * both unsignaled. The client uses them to reregister
506 * the rdma buffers with a new key each iteration.
507 */
508 cb->reg_mr_wr.wr.opcode = IB_WR_REG_MR;
509 cb->reg_mr_wr.mr = cb->reg_mr;
510
511 cb->invalidate_wr.next = &cb->reg_mr_wr.wr;
512 cb->invalidate_wr.opcode = IB_WR_LOCAL_INV;
513 }
514
krping_setup_buffers(struct krping_cb * cb)515 static int krping_setup_buffers(struct krping_cb *cb)
516 {
517 int ret;
518
519 DEBUG_LOG(PFX "krping_setup_buffers called on cb %p\n", cb);
520
521 cb->recv_dma_addr = ib_dma_map_single(cb->pd->device,
522 &cb->recv_buf,
523 sizeof(cb->recv_buf), DMA_BIDIRECTIONAL);
524 pci_unmap_addr_set(cb, recv_mapping, cb->recv_dma_addr);
525 cb->send_dma_addr = ib_dma_map_single(cb->pd->device,
526 &cb->send_buf, sizeof(cb->send_buf),
527 DMA_BIDIRECTIONAL);
528 pci_unmap_addr_set(cb, send_mapping, cb->send_dma_addr);
529
530 cb->rdma_buf = ib_dma_alloc_coherent(cb->pd->device, cb->size,
531 &cb->rdma_dma_addr,
532 GFP_KERNEL);
533 if (!cb->rdma_buf) {
534 DEBUG_LOG(PFX "rdma_buf allocation failed\n");
535 ret = -ENOMEM;
536 goto bail;
537 }
538 pci_unmap_addr_set(cb, rdma_mapping, cb->rdma_dma_addr);
539 cb->page_list_len = (((cb->size - 1) & PAGE_MASK) + PAGE_SIZE)
540 >> PAGE_SHIFT;
541 cb->reg_mr = ib_alloc_mr(cb->pd, IB_MR_TYPE_MEM_REG,
542 cb->page_list_len);
543 if (IS_ERR(cb->reg_mr)) {
544 ret = PTR_ERR(cb->reg_mr);
545 DEBUG_LOG(PFX "recv_buf reg_mr failed %d\n", ret);
546 goto bail;
547 }
548 DEBUG_LOG(PFX "reg rkey 0x%x page_list_len %u\n",
549 cb->reg_mr->rkey, cb->page_list_len);
550
551 if (!cb->server || cb->wlat || cb->rlat || cb->bw) {
552
553 cb->start_buf = ib_dma_alloc_coherent(cb->pd->device, cb->size,
554 &cb->start_dma_addr,
555 GFP_KERNEL);
556 if (!cb->start_buf) {
557 DEBUG_LOG(PFX "start_buf malloc failed\n");
558 ret = -ENOMEM;
559 goto bail;
560 }
561 pci_unmap_addr_set(cb, start_mapping, cb->start_dma_addr);
562 }
563
564 krping_setup_wr(cb);
565 DEBUG_LOG(PFX "allocated & registered buffers...\n");
566 return 0;
567 bail:
568 if (cb->reg_mr && !IS_ERR(cb->reg_mr))
569 ib_dereg_mr(cb->reg_mr);
570 if (cb->rdma_mr && !IS_ERR(cb->rdma_mr))
571 ib_dereg_mr(cb->rdma_mr);
572 if (cb->dma_mr && !IS_ERR(cb->dma_mr))
573 ib_dereg_mr(cb->dma_mr);
574 if (cb->rdma_buf) {
575 ib_dma_free_coherent(cb->pd->device, cb->size, cb->rdma_buf,
576 cb->rdma_dma_addr);
577 }
578 if (cb->start_buf) {
579 ib_dma_free_coherent(cb->pd->device, cb->size, cb->start_buf,
580 cb->start_dma_addr);
581 }
582 return ret;
583 }
584
krping_free_buffers(struct krping_cb * cb)585 static void krping_free_buffers(struct krping_cb *cb)
586 {
587 DEBUG_LOG("krping_free_buffers called on cb %p\n", cb);
588
589 if (cb->dma_mr)
590 ib_dereg_mr(cb->dma_mr);
591 if (cb->rdma_mr)
592 ib_dereg_mr(cb->rdma_mr);
593 if (cb->start_mr)
594 ib_dereg_mr(cb->start_mr);
595 if (cb->reg_mr)
596 ib_dereg_mr(cb->reg_mr);
597
598 dma_unmap_single(cb->pd->device->dma_device,
599 pci_unmap_addr(cb, recv_mapping),
600 sizeof(cb->recv_buf), DMA_BIDIRECTIONAL);
601 dma_unmap_single(cb->pd->device->dma_device,
602 pci_unmap_addr(cb, send_mapping),
603 sizeof(cb->send_buf), DMA_BIDIRECTIONAL);
604
605 ib_dma_free_coherent(cb->pd->device, cb->size, cb->rdma_buf,
606 cb->rdma_dma_addr);
607
608 if (cb->start_buf) {
609 ib_dma_free_coherent(cb->pd->device, cb->size, cb->start_buf,
610 cb->start_dma_addr);
611 }
612 }
613
krping_create_qp(struct krping_cb * cb)614 static int krping_create_qp(struct krping_cb *cb)
615 {
616 struct ib_qp_init_attr init_attr;
617 int ret;
618
619 memset(&init_attr, 0, sizeof(init_attr));
620 init_attr.cap.max_send_wr = cb->txdepth;
621 init_attr.cap.max_recv_wr = 2;
622
623 /* For flush_qp() */
624 init_attr.cap.max_send_wr++;
625 init_attr.cap.max_recv_wr++;
626
627 init_attr.cap.max_recv_sge = 1;
628 init_attr.cap.max_send_sge = 1;
629 init_attr.qp_type = IB_QPT_RC;
630 init_attr.send_cq = cb->cq;
631 init_attr.recv_cq = cb->cq;
632 init_attr.sq_sig_type = IB_SIGNAL_REQ_WR;
633
634 if (cb->server) {
635 ret = rdma_create_qp(cb->child_cm_id, cb->pd, &init_attr);
636 if (!ret)
637 cb->qp = cb->child_cm_id->qp;
638 } else {
639 ret = rdma_create_qp(cb->cm_id, cb->pd, &init_attr);
640 if (!ret)
641 cb->qp = cb->cm_id->qp;
642 }
643
644 return ret;
645 }
646
krping_free_qp(struct krping_cb * cb)647 static void krping_free_qp(struct krping_cb *cb)
648 {
649 ib_destroy_qp(cb->qp);
650 ib_destroy_cq(cb->cq);
651 ib_dealloc_pd(cb->pd);
652 }
653
krping_setup_qp(struct krping_cb * cb,struct rdma_cm_id * cm_id)654 static int krping_setup_qp(struct krping_cb *cb, struct rdma_cm_id *cm_id)
655 {
656 int ret;
657 struct ib_cq_init_attr attr = {0};
658
659 cb->pd = ib_alloc_pd(cm_id->device, 0);
660 if (IS_ERR(cb->pd)) {
661 printk(KERN_ERR PFX "ib_alloc_pd failed\n");
662 return PTR_ERR(cb->pd);
663 }
664 DEBUG_LOG("created pd %p\n", cb->pd);
665
666 strlcpy(cb->stats.name, cb->pd->device->name, sizeof(cb->stats.name));
667
668 attr.cqe = cb->txdepth * 2;
669 attr.comp_vector = 0;
670 cb->cq = ib_create_cq(cm_id->device, krping_cq_event_handler, NULL,
671 cb, &attr);
672 if (IS_ERR(cb->cq)) {
673 printk(KERN_ERR PFX "ib_create_cq failed\n");
674 ret = PTR_ERR(cb->cq);
675 goto err1;
676 }
677 DEBUG_LOG("created cq %p\n", cb->cq);
678
679 if (!cb->wlat && !cb->rlat && !cb->bw && !cb->frtest) {
680 ret = ib_req_notify_cq(cb->cq, IB_CQ_NEXT_COMP);
681 if (ret) {
682 printk(KERN_ERR PFX "ib_create_cq failed\n");
683 goto err2;
684 }
685 }
686
687 ret = krping_create_qp(cb);
688 if (ret) {
689 printk(KERN_ERR PFX "krping_create_qp failed: %d\n", ret);
690 goto err2;
691 }
692 DEBUG_LOG("created qp %p\n", cb->qp);
693 return 0;
694 err2:
695 ib_destroy_cq(cb->cq);
696 err1:
697 ib_dealloc_pd(cb->pd);
698 return ret;
699 }
700
701 /*
702 * return the (possibly rebound) rkey for the rdma buffer.
703 * REG mode: invalidate and rebind via reg wr.
704 * other modes: just return the mr rkey.
705 */
krping_rdma_rkey(struct krping_cb * cb,u64 buf,int post_inv)706 static u32 krping_rdma_rkey(struct krping_cb *cb, u64 buf, int post_inv)
707 {
708 u32 rkey;
709 struct ib_send_wr *bad_wr;
710 int ret;
711 struct scatterlist sg = {0};
712
713 cb->invalidate_wr.ex.invalidate_rkey = cb->reg_mr->rkey;
714
715 /*
716 * Update the reg key.
717 */
718 ib_update_fast_reg_key(cb->reg_mr, ++cb->key);
719 cb->reg_mr_wr.key = cb->reg_mr->rkey;
720
721 /*
722 * Update the reg WR with new buf info.
723 */
724 if (buf == (u64)cb->start_dma_addr)
725 cb->reg_mr_wr.access = IB_ACCESS_REMOTE_READ;
726 else
727 cb->reg_mr_wr.access = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE;
728 sg_dma_address(&sg) = buf;
729 sg_dma_len(&sg) = cb->size;
730
731 ret = ib_map_mr_sg(cb->reg_mr, &sg, 1, NULL, PAGE_SIZE);
732 BUG_ON(ret <= 0 || ret > cb->page_list_len);
733
734 DEBUG_LOG(PFX "post_inv = %d, reg_mr new rkey 0x%x pgsz %u len %u"
735 " iova_start %llx\n",
736 post_inv,
737 cb->reg_mr_wr.key,
738 cb->reg_mr->page_size,
739 cb->reg_mr->length,
740 (unsigned long long)cb->reg_mr->iova);
741
742 if (post_inv)
743 ret = ib_post_send(cb->qp, &cb->invalidate_wr, &bad_wr);
744 else
745 ret = ib_post_send(cb->qp, &cb->reg_mr_wr.wr, &bad_wr);
746 if (ret) {
747 printk(KERN_ERR PFX "post send error %d\n", ret);
748 cb->state = ERROR;
749 }
750 rkey = cb->reg_mr->rkey;
751 return rkey;
752 }
753
krping_format_send(struct krping_cb * cb,u64 buf)754 static void krping_format_send(struct krping_cb *cb, u64 buf)
755 {
756 struct krping_rdma_info *info = &cb->send_buf;
757 u32 rkey;
758
759 /*
760 * Client side will do reg or mw bind before
761 * advertising the rdma buffer. Server side
762 * sends have no data.
763 */
764 if (!cb->server || cb->wlat || cb->rlat || cb->bw) {
765 rkey = krping_rdma_rkey(cb, buf, !cb->server_invalidate);
766 info->buf = htonll(buf);
767 info->rkey = htonl(rkey);
768 info->size = htonl(cb->size);
769 DEBUG_LOG("RDMA addr %llx rkey %x len %d\n",
770 (unsigned long long)buf, rkey, cb->size);
771 }
772 }
773
krping_test_server(struct krping_cb * cb)774 static void krping_test_server(struct krping_cb *cb)
775 {
776 struct ib_send_wr *bad_wr, inv;
777 int ret;
778
779 while (1) {
780 /* Wait for client's Start STAG/TO/Len */
781 wait_event_interruptible(cb->sem, cb->state >= RDMA_READ_ADV);
782 if (cb->state != RDMA_READ_ADV) {
783 printk(KERN_ERR PFX "wait for RDMA_READ_ADV state %d\n",
784 cb->state);
785 break;
786 }
787
788 DEBUG_LOG("server received sink adv\n");
789
790 cb->rdma_sq_wr.rkey = cb->remote_rkey;
791 cb->rdma_sq_wr.remote_addr = cb->remote_addr;
792 cb->rdma_sq_wr.wr.sg_list->length = cb->remote_len;
793 cb->rdma_sgl.lkey = krping_rdma_rkey(cb, cb->rdma_dma_addr, !cb->read_inv);
794 cb->rdma_sq_wr.wr.next = NULL;
795
796 /* Issue RDMA Read. */
797 if (cb->read_inv)
798 cb->rdma_sq_wr.wr.opcode = IB_WR_RDMA_READ_WITH_INV;
799 else {
800
801 cb->rdma_sq_wr.wr.opcode = IB_WR_RDMA_READ;
802 /*
803 * Immediately follow the read with a
804 * fenced LOCAL_INV.
805 */
806 cb->rdma_sq_wr.wr.next = &inv;
807 memset(&inv, 0, sizeof inv);
808 inv.opcode = IB_WR_LOCAL_INV;
809 inv.ex.invalidate_rkey = cb->reg_mr->rkey;
810 inv.send_flags = IB_SEND_FENCE;
811 }
812
813 ret = ib_post_send(cb->qp, &cb->rdma_sq_wr.wr, &bad_wr);
814 if (ret) {
815 printk(KERN_ERR PFX "post send error %d\n", ret);
816 break;
817 }
818 cb->rdma_sq_wr.wr.next = NULL;
819
820 DEBUG_LOG("server posted rdma read req \n");
821
822 /* Wait for read completion */
823 wait_event_interruptible(cb->sem,
824 cb->state >= RDMA_READ_COMPLETE);
825 if (cb->state != RDMA_READ_COMPLETE) {
826 printk(KERN_ERR PFX
827 "wait for RDMA_READ_COMPLETE state %d\n",
828 cb->state);
829 break;
830 }
831 DEBUG_LOG("server received read complete\n");
832
833 /* Display data in recv buf */
834 if (cb->verbose)
835 printk(KERN_INFO PFX "server ping data: %s\n",
836 cb->rdma_buf);
837
838 /* Tell client to continue */
839 if (cb->server && cb->server_invalidate) {
840 cb->sq_wr.ex.invalidate_rkey = cb->remote_rkey;
841 cb->sq_wr.opcode = IB_WR_SEND_WITH_INV;
842 DEBUG_LOG("send-w-inv rkey 0x%x\n", cb->remote_rkey);
843 }
844 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
845 if (ret) {
846 printk(KERN_ERR PFX "post send error %d\n", ret);
847 break;
848 }
849 DEBUG_LOG("server posted go ahead\n");
850
851 /* Wait for client's RDMA STAG/TO/Len */
852 wait_event_interruptible(cb->sem, cb->state >= RDMA_WRITE_ADV);
853 if (cb->state != RDMA_WRITE_ADV) {
854 printk(KERN_ERR PFX
855 "wait for RDMA_WRITE_ADV state %d\n",
856 cb->state);
857 break;
858 }
859 DEBUG_LOG("server received sink adv\n");
860
861 /* RDMA Write echo data */
862 cb->rdma_sq_wr.wr.opcode = IB_WR_RDMA_WRITE;
863 cb->rdma_sq_wr.rkey = cb->remote_rkey;
864 cb->rdma_sq_wr.remote_addr = cb->remote_addr;
865 cb->rdma_sq_wr.wr.sg_list->length = strlen(cb->rdma_buf) + 1;
866 if (cb->local_dma_lkey)
867 cb->rdma_sgl.lkey = cb->pd->local_dma_lkey;
868 else
869 cb->rdma_sgl.lkey = krping_rdma_rkey(cb, cb->rdma_dma_addr, 0);
870
871 DEBUG_LOG("rdma write from lkey %x laddr %llx len %d\n",
872 cb->rdma_sq_wr.wr.sg_list->lkey,
873 (unsigned long long)cb->rdma_sq_wr.wr.sg_list->addr,
874 cb->rdma_sq_wr.wr.sg_list->length);
875
876 ret = ib_post_send(cb->qp, &cb->rdma_sq_wr.wr, &bad_wr);
877 if (ret) {
878 printk(KERN_ERR PFX "post send error %d\n", ret);
879 break;
880 }
881
882 /* Wait for completion */
883 ret = wait_event_interruptible(cb->sem, cb->state >=
884 RDMA_WRITE_COMPLETE);
885 if (cb->state != RDMA_WRITE_COMPLETE) {
886 printk(KERN_ERR PFX
887 "wait for RDMA_WRITE_COMPLETE state %d\n",
888 cb->state);
889 break;
890 }
891 DEBUG_LOG("server rdma write complete \n");
892
893 cb->state = CONNECTED;
894
895 /* Tell client to begin again */
896 if (cb->server && cb->server_invalidate) {
897 cb->sq_wr.ex.invalidate_rkey = cb->remote_rkey;
898 cb->sq_wr.opcode = IB_WR_SEND_WITH_INV;
899 DEBUG_LOG("send-w-inv rkey 0x%x\n", cb->remote_rkey);
900 }
901 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
902 if (ret) {
903 printk(KERN_ERR PFX "post send error %d\n", ret);
904 break;
905 }
906 DEBUG_LOG("server posted go ahead\n");
907 }
908 }
909
rlat_test(struct krping_cb * cb)910 static void rlat_test(struct krping_cb *cb)
911 {
912 int scnt;
913 int iters = cb->count;
914 struct timeval start_tv, stop_tv;
915 int ret;
916 struct ib_wc wc;
917 struct ib_send_wr *bad_wr;
918 int ne;
919
920 scnt = 0;
921 cb->rdma_sq_wr.wr.opcode = IB_WR_RDMA_READ;
922 cb->rdma_sq_wr.rkey = cb->remote_rkey;
923 cb->rdma_sq_wr.remote_addr = cb->remote_addr;
924 cb->rdma_sq_wr.wr.sg_list->length = cb->size;
925
926 microtime(&start_tv);
927 if (!cb->poll) {
928 cb->state = RDMA_READ_ADV;
929 ib_req_notify_cq(cb->cq, IB_CQ_NEXT_COMP);
930 }
931 while (scnt < iters) {
932
933 cb->state = RDMA_READ_ADV;
934 ret = ib_post_send(cb->qp, &cb->rdma_sq_wr.wr, &bad_wr);
935 if (ret) {
936 printk(KERN_ERR PFX
937 "Couldn't post send: ret=%d scnt %d\n",
938 ret, scnt);
939 return;
940 }
941
942 do {
943 if (!cb->poll) {
944 wait_event_interruptible(cb->sem,
945 cb->state != RDMA_READ_ADV);
946 if (cb->state == RDMA_READ_COMPLETE) {
947 ne = 1;
948 ib_req_notify_cq(cb->cq,
949 IB_CQ_NEXT_COMP);
950 } else {
951 ne = -1;
952 }
953 } else
954 ne = ib_poll_cq(cb->cq, 1, &wc);
955 if (cb->state == ERROR) {
956 printk(KERN_ERR PFX
957 "state == ERROR...bailing scnt %d\n",
958 scnt);
959 return;
960 }
961 } while (ne == 0);
962
963 if (ne < 0) {
964 printk(KERN_ERR PFX "poll CQ failed %d\n", ne);
965 return;
966 }
967 if (cb->poll && wc.status != IB_WC_SUCCESS) {
968 printk(KERN_ERR PFX "Completion wth error at %s:\n",
969 cb->server ? "server" : "client");
970 printk(KERN_ERR PFX "Failed status %d: wr_id %d\n",
971 wc.status, (int) wc.wr_id);
972 return;
973 }
974 ++scnt;
975 }
976 microtime(&stop_tv);
977
978 if (stop_tv.tv_usec < start_tv.tv_usec) {
979 stop_tv.tv_usec += 1000000;
980 stop_tv.tv_sec -= 1;
981 }
982
983 printk(KERN_ERR PFX "delta sec %lu delta usec %lu iter %d size %d\n",
984 (unsigned long)(stop_tv.tv_sec - start_tv.tv_sec),
985 (unsigned long)(stop_tv.tv_usec - start_tv.tv_usec),
986 scnt, cb->size);
987 }
988
wlat_test(struct krping_cb * cb)989 static void wlat_test(struct krping_cb *cb)
990 {
991 int ccnt, scnt, rcnt;
992 int iters=cb->count;
993 volatile char *poll_buf = (char *) cb->start_buf;
994 char *buf = (char *)cb->rdma_buf;
995 struct timeval start_tv, stop_tv;
996 cycles_t *post_cycles_start, *post_cycles_stop;
997 cycles_t *poll_cycles_start, *poll_cycles_stop;
998 cycles_t *last_poll_cycles_start;
999 cycles_t sum_poll = 0, sum_post = 0, sum_last_poll = 0;
1000 int i;
1001 int cycle_iters = 1000;
1002
1003 ccnt = 0;
1004 scnt = 0;
1005 rcnt = 0;
1006
1007 post_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1008 if (!post_cycles_start) {
1009 printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1010 return;
1011 }
1012 post_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1013 if (!post_cycles_stop) {
1014 printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1015 return;
1016 }
1017 poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1018 if (!poll_cycles_start) {
1019 printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1020 return;
1021 }
1022 poll_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1023 if (!poll_cycles_stop) {
1024 printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1025 return;
1026 }
1027 last_poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t),
1028 GFP_KERNEL);
1029 if (!last_poll_cycles_start) {
1030 printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1031 return;
1032 }
1033 cb->rdma_sq_wr.wr.opcode = IB_WR_RDMA_WRITE;
1034 cb->rdma_sq_wr.rkey = cb->remote_rkey;
1035 cb->rdma_sq_wr.remote_addr = cb->remote_addr;
1036 cb->rdma_sq_wr.wr.sg_list->length = cb->size;
1037
1038 if (cycle_iters > iters)
1039 cycle_iters = iters;
1040 microtime(&start_tv);
1041 while (scnt < iters || ccnt < iters || rcnt < iters) {
1042
1043 /* Wait till buffer changes. */
1044 if (rcnt < iters && !(scnt < 1 && !cb->server)) {
1045 ++rcnt;
1046 while (*poll_buf != (char)rcnt) {
1047 if (cb->state == ERROR) {
1048 printk(KERN_ERR PFX
1049 "state = ERROR, bailing\n");
1050 return;
1051 }
1052 }
1053 }
1054
1055 if (scnt < iters) {
1056 struct ib_send_wr *bad_wr;
1057
1058 *buf = (char)scnt+1;
1059 if (scnt < cycle_iters)
1060 post_cycles_start[scnt] = get_cycles();
1061 if (ib_post_send(cb->qp, &cb->rdma_sq_wr.wr, &bad_wr)) {
1062 printk(KERN_ERR PFX
1063 "Couldn't post send: scnt=%d\n",
1064 scnt);
1065 return;
1066 }
1067 if (scnt < cycle_iters)
1068 post_cycles_stop[scnt] = get_cycles();
1069 scnt++;
1070 }
1071
1072 if (ccnt < iters) {
1073 struct ib_wc wc;
1074 int ne;
1075
1076 if (ccnt < cycle_iters)
1077 poll_cycles_start[ccnt] = get_cycles();
1078 do {
1079 if (ccnt < cycle_iters)
1080 last_poll_cycles_start[ccnt] =
1081 get_cycles();
1082 ne = ib_poll_cq(cb->cq, 1, &wc);
1083 } while (ne == 0);
1084 if (ccnt < cycle_iters)
1085 poll_cycles_stop[ccnt] = get_cycles();
1086 ++ccnt;
1087
1088 if (ne < 0) {
1089 printk(KERN_ERR PFX "poll CQ failed %d\n", ne);
1090 return;
1091 }
1092 if (wc.status != IB_WC_SUCCESS) {
1093 printk(KERN_ERR PFX
1094 "Completion wth error at %s:\n",
1095 cb->server ? "server" : "client");
1096 printk(KERN_ERR PFX
1097 "Failed status %d: wr_id %d\n",
1098 wc.status, (int) wc.wr_id);
1099 printk(KERN_ERR PFX
1100 "scnt=%d, rcnt=%d, ccnt=%d\n",
1101 scnt, rcnt, ccnt);
1102 return;
1103 }
1104 }
1105 }
1106 microtime(&stop_tv);
1107
1108 if (stop_tv.tv_usec < start_tv.tv_usec) {
1109 stop_tv.tv_usec += 1000000;
1110 stop_tv.tv_sec -= 1;
1111 }
1112
1113 for (i=0; i < cycle_iters; i++) {
1114 sum_post += post_cycles_stop[i] - post_cycles_start[i];
1115 sum_poll += poll_cycles_stop[i] - poll_cycles_start[i];
1116 sum_last_poll += poll_cycles_stop[i]-last_poll_cycles_start[i];
1117 }
1118 printk(KERN_ERR PFX
1119 "delta sec %lu delta usec %lu iter %d size %d cycle_iters %d"
1120 " sum_post %llu sum_poll %llu sum_last_poll %llu\n",
1121 (unsigned long)(stop_tv.tv_sec - start_tv.tv_sec),
1122 (unsigned long)(stop_tv.tv_usec - start_tv.tv_usec),
1123 scnt, cb->size, cycle_iters,
1124 (unsigned long long)sum_post, (unsigned long long)sum_poll,
1125 (unsigned long long)sum_last_poll);
1126 kfree(post_cycles_start);
1127 kfree(post_cycles_stop);
1128 kfree(poll_cycles_start);
1129 kfree(poll_cycles_stop);
1130 kfree(last_poll_cycles_start);
1131 }
1132
bw_test(struct krping_cb * cb)1133 static void bw_test(struct krping_cb *cb)
1134 {
1135 int ccnt, scnt, rcnt;
1136 int iters=cb->count;
1137 struct timeval start_tv, stop_tv;
1138 cycles_t *post_cycles_start, *post_cycles_stop;
1139 cycles_t *poll_cycles_start, *poll_cycles_stop;
1140 cycles_t *last_poll_cycles_start;
1141 cycles_t sum_poll = 0, sum_post = 0, sum_last_poll = 0;
1142 int i;
1143 int cycle_iters = 1000;
1144
1145 ccnt = 0;
1146 scnt = 0;
1147 rcnt = 0;
1148
1149 post_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1150 if (!post_cycles_start) {
1151 printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1152 return;
1153 }
1154 post_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1155 if (!post_cycles_stop) {
1156 printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1157 return;
1158 }
1159 poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1160 if (!poll_cycles_start) {
1161 printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1162 return;
1163 }
1164 poll_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1165 if (!poll_cycles_stop) {
1166 printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1167 return;
1168 }
1169 last_poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t),
1170 GFP_KERNEL);
1171 if (!last_poll_cycles_start) {
1172 printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1173 return;
1174 }
1175 cb->rdma_sq_wr.wr.opcode = IB_WR_RDMA_WRITE;
1176 cb->rdma_sq_wr.rkey = cb->remote_rkey;
1177 cb->rdma_sq_wr.remote_addr = cb->remote_addr;
1178 cb->rdma_sq_wr.wr.sg_list->length = cb->size;
1179
1180 if (cycle_iters > iters)
1181 cycle_iters = iters;
1182 microtime(&start_tv);
1183 while (scnt < iters || ccnt < iters) {
1184
1185 while (scnt < iters && scnt - ccnt < cb->txdepth) {
1186 struct ib_send_wr *bad_wr;
1187
1188 if (scnt < cycle_iters)
1189 post_cycles_start[scnt] = get_cycles();
1190 if (ib_post_send(cb->qp, &cb->rdma_sq_wr.wr, &bad_wr)) {
1191 printk(KERN_ERR PFX
1192 "Couldn't post send: scnt=%d\n",
1193 scnt);
1194 return;
1195 }
1196 if (scnt < cycle_iters)
1197 post_cycles_stop[scnt] = get_cycles();
1198 ++scnt;
1199 }
1200
1201 if (ccnt < iters) {
1202 int ne;
1203 struct ib_wc wc;
1204
1205 if (ccnt < cycle_iters)
1206 poll_cycles_start[ccnt] = get_cycles();
1207 do {
1208 if (ccnt < cycle_iters)
1209 last_poll_cycles_start[ccnt] =
1210 get_cycles();
1211 ne = ib_poll_cq(cb->cq, 1, &wc);
1212 } while (ne == 0);
1213 if (ccnt < cycle_iters)
1214 poll_cycles_stop[ccnt] = get_cycles();
1215 ccnt += 1;
1216
1217 if (ne < 0) {
1218 printk(KERN_ERR PFX "poll CQ failed %d\n", ne);
1219 return;
1220 }
1221 if (wc.status != IB_WC_SUCCESS) {
1222 printk(KERN_ERR PFX
1223 "Completion wth error at %s:\n",
1224 cb->server ? "server" : "client");
1225 printk(KERN_ERR PFX
1226 "Failed status %d: wr_id %d\n",
1227 wc.status, (int) wc.wr_id);
1228 return;
1229 }
1230 }
1231 }
1232 microtime(&stop_tv);
1233
1234 if (stop_tv.tv_usec < start_tv.tv_usec) {
1235 stop_tv.tv_usec += 1000000;
1236 stop_tv.tv_sec -= 1;
1237 }
1238
1239 for (i=0; i < cycle_iters; i++) {
1240 sum_post += post_cycles_stop[i] - post_cycles_start[i];
1241 sum_poll += poll_cycles_stop[i] - poll_cycles_start[i];
1242 sum_last_poll += poll_cycles_stop[i]-last_poll_cycles_start[i];
1243 }
1244 printk(KERN_ERR PFX
1245 "delta sec %lu delta usec %lu iter %d size %d cycle_iters %d"
1246 " sum_post %llu sum_poll %llu sum_last_poll %llu\n",
1247 (unsigned long)(stop_tv.tv_sec - start_tv.tv_sec),
1248 (unsigned long)(stop_tv.tv_usec - start_tv.tv_usec),
1249 scnt, cb->size, cycle_iters,
1250 (unsigned long long)sum_post, (unsigned long long)sum_poll,
1251 (unsigned long long)sum_last_poll);
1252 kfree(post_cycles_start);
1253 kfree(post_cycles_stop);
1254 kfree(poll_cycles_start);
1255 kfree(poll_cycles_stop);
1256 kfree(last_poll_cycles_start);
1257 }
1258
krping_rlat_test_server(struct krping_cb * cb)1259 static void krping_rlat_test_server(struct krping_cb *cb)
1260 {
1261 struct ib_send_wr *bad_wr;
1262 struct ib_wc wc;
1263 int ret;
1264
1265 /* Spin waiting for client's Start STAG/TO/Len */
1266 while (cb->state < RDMA_READ_ADV) {
1267 krping_cq_event_handler(cb->cq, cb);
1268 }
1269
1270 /* Send STAG/TO/Len to client */
1271 krping_format_send(cb, cb->start_dma_addr);
1272 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1273 if (ret) {
1274 printk(KERN_ERR PFX "post send error %d\n", ret);
1275 return;
1276 }
1277
1278 /* Spin waiting for send completion */
1279 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1280 if (ret < 0) {
1281 printk(KERN_ERR PFX "poll error %d\n", ret);
1282 return;
1283 }
1284 if (wc.status) {
1285 printk(KERN_ERR PFX "send completiong error %d\n", wc.status);
1286 return;
1287 }
1288
1289 wait_event_interruptible(cb->sem, cb->state == ERROR);
1290 }
1291
krping_wlat_test_server(struct krping_cb * cb)1292 static void krping_wlat_test_server(struct krping_cb *cb)
1293 {
1294 struct ib_send_wr *bad_wr;
1295 struct ib_wc wc;
1296 int ret;
1297
1298 /* Spin waiting for client's Start STAG/TO/Len */
1299 while (cb->state < RDMA_READ_ADV) {
1300 krping_cq_event_handler(cb->cq, cb);
1301 }
1302
1303 /* Send STAG/TO/Len to client */
1304 krping_format_send(cb, cb->start_dma_addr);
1305 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1306 if (ret) {
1307 printk(KERN_ERR PFX "post send error %d\n", ret);
1308 return;
1309 }
1310
1311 /* Spin waiting for send completion */
1312 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1313 if (ret < 0) {
1314 printk(KERN_ERR PFX "poll error %d\n", ret);
1315 return;
1316 }
1317 if (wc.status) {
1318 printk(KERN_ERR PFX "send completiong error %d\n", wc.status);
1319 return;
1320 }
1321
1322 wlat_test(cb);
1323 wait_event_interruptible(cb->sem, cb->state == ERROR);
1324 }
1325
krping_bw_test_server(struct krping_cb * cb)1326 static void krping_bw_test_server(struct krping_cb *cb)
1327 {
1328 struct ib_send_wr *bad_wr;
1329 struct ib_wc wc;
1330 int ret;
1331
1332 /* Spin waiting for client's Start STAG/TO/Len */
1333 while (cb->state < RDMA_READ_ADV) {
1334 krping_cq_event_handler(cb->cq, cb);
1335 }
1336
1337 /* Send STAG/TO/Len to client */
1338 krping_format_send(cb, cb->start_dma_addr);
1339 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1340 if (ret) {
1341 printk(KERN_ERR PFX "post send error %d\n", ret);
1342 return;
1343 }
1344
1345 /* Spin waiting for send completion */
1346 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1347 if (ret < 0) {
1348 printk(KERN_ERR PFX "poll error %d\n", ret);
1349 return;
1350 }
1351 if (wc.status) {
1352 printk(KERN_ERR PFX "send completiong error %d\n", wc.status);
1353 return;
1354 }
1355
1356 if (cb->duplex)
1357 bw_test(cb);
1358 wait_event_interruptible(cb->sem, cb->state == ERROR);
1359 }
1360
reg_supported(struct ib_device * dev)1361 static int reg_supported(struct ib_device *dev)
1362 {
1363 u64 needed_flags = IB_DEVICE_MEM_MGT_EXTENSIONS;
1364
1365 if ((dev->attrs.device_cap_flags & needed_flags) != needed_flags) {
1366 printk(KERN_ERR PFX
1367 "Fastreg not supported - device_cap_flags 0x%llx\n",
1368 (unsigned long long)dev->attrs.device_cap_flags);
1369 return 0;
1370 }
1371 DEBUG_LOG("Fastreg supported - device_cap_flags 0x%llx\n",
1372 (unsigned long long)dev->attrs.device_cap_flags);
1373 return 1;
1374 }
1375
fill_sockaddr(struct sockaddr_storage * sin,struct krping_cb * cb)1376 static void fill_sockaddr(struct sockaddr_storage *sin, struct krping_cb *cb)
1377 {
1378 memset(sin, 0, sizeof(*sin));
1379
1380 if (cb->addr_type == AF_INET) {
1381 struct sockaddr_in *sin4 = (struct sockaddr_in *)sin;
1382 sin4->sin_len = sizeof(*sin4);
1383 sin4->sin_family = AF_INET;
1384 memcpy((void *)&sin4->sin_addr.s_addr, cb->addr, 4);
1385 sin4->sin_port = cb->port;
1386 } else if (cb->addr_type == AF_INET6) {
1387 struct sockaddr_in6 *sin6 = (struct sockaddr_in6 *)sin;
1388 sin6->sin6_len = sizeof(*sin6);
1389 sin6->sin6_family = AF_INET6;
1390 memcpy((void *)&sin6->sin6_addr, cb->addr, 16);
1391 sin6->sin6_port = cb->port;
1392 }
1393 }
1394
krping_bind_server(struct krping_cb * cb)1395 static int krping_bind_server(struct krping_cb *cb)
1396 {
1397 struct sockaddr_storage sin;
1398 int ret;
1399
1400
1401 fill_sockaddr(&sin, cb);
1402
1403 ret = rdma_bind_addr(cb->cm_id, (struct sockaddr *)&sin);
1404 if (ret) {
1405 printk(KERN_ERR PFX "rdma_bind_addr error %d\n", ret);
1406 return ret;
1407 }
1408 DEBUG_LOG("rdma_bind_addr successful\n");
1409
1410 DEBUG_LOG("rdma_listen\n");
1411 ret = rdma_listen(cb->cm_id, 3);
1412 if (ret) {
1413 printk(KERN_ERR PFX "rdma_listen failed: %d\n", ret);
1414 return ret;
1415 }
1416
1417 wait_event_interruptible(cb->sem, cb->state >= CONNECT_REQUEST);
1418 if (cb->state != CONNECT_REQUEST) {
1419 printk(KERN_ERR PFX "wait for CONNECT_REQUEST state %d\n",
1420 cb->state);
1421 return -1;
1422 }
1423
1424 if (!reg_supported(cb->child_cm_id->device))
1425 return -EINVAL;
1426
1427 return 0;
1428 }
1429
krping_run_server(struct krping_cb * cb)1430 static void krping_run_server(struct krping_cb *cb)
1431 {
1432 struct ib_recv_wr *bad_wr;
1433 int ret;
1434
1435 ret = krping_bind_server(cb);
1436 if (ret)
1437 return;
1438
1439 ret = krping_setup_qp(cb, cb->child_cm_id);
1440 if (ret) {
1441 printk(KERN_ERR PFX "setup_qp failed: %d\n", ret);
1442 goto err0;
1443 }
1444
1445 ret = krping_setup_buffers(cb);
1446 if (ret) {
1447 printk(KERN_ERR PFX "krping_setup_buffers failed: %d\n", ret);
1448 goto err1;
1449 }
1450
1451 ret = ib_post_recv(cb->qp, &cb->rq_wr, &bad_wr);
1452 if (ret) {
1453 printk(KERN_ERR PFX "ib_post_recv failed: %d\n", ret);
1454 goto err2;
1455 }
1456
1457 ret = krping_accept(cb);
1458 if (ret) {
1459 printk(KERN_ERR PFX "connect error %d\n", ret);
1460 goto err2;
1461 }
1462
1463 if (cb->wlat)
1464 krping_wlat_test_server(cb);
1465 else if (cb->rlat)
1466 krping_rlat_test_server(cb);
1467 else if (cb->bw)
1468 krping_bw_test_server(cb);
1469 else
1470 krping_test_server(cb);
1471 rdma_disconnect(cb->child_cm_id);
1472 err2:
1473 krping_free_buffers(cb);
1474 err1:
1475 krping_free_qp(cb);
1476 err0:
1477 rdma_destroy_id(cb->child_cm_id);
1478 }
1479
krping_test_client(struct krping_cb * cb)1480 static void krping_test_client(struct krping_cb *cb)
1481 {
1482 int ping, start, cc, i, ret;
1483 struct ib_send_wr *bad_wr;
1484 unsigned char c;
1485
1486 start = 65;
1487 for (ping = 0; !cb->count || ping < cb->count; ping++) {
1488 cb->state = RDMA_READ_ADV;
1489
1490 /* Put some ascii text in the buffer. */
1491 cc = sprintf(cb->start_buf, "rdma-ping-%d: ", ping);
1492 for (i = cc, c = start; i < cb->size; i++) {
1493 cb->start_buf[i] = c;
1494 c++;
1495 if (c > 122)
1496 c = 65;
1497 }
1498 start++;
1499 if (start > 122)
1500 start = 65;
1501 cb->start_buf[cb->size - 1] = 0;
1502
1503 krping_format_send(cb, cb->start_dma_addr);
1504 if (cb->state == ERROR) {
1505 printk(KERN_ERR PFX "krping_format_send failed\n");
1506 break;
1507 }
1508 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1509 if (ret) {
1510 printk(KERN_ERR PFX "post send error %d\n", ret);
1511 break;
1512 }
1513
1514 /* Wait for server to ACK */
1515 wait_event_interruptible(cb->sem, cb->state >= RDMA_WRITE_ADV);
1516 if (cb->state != RDMA_WRITE_ADV) {
1517 printk(KERN_ERR PFX
1518 "wait for RDMA_WRITE_ADV state %d\n",
1519 cb->state);
1520 break;
1521 }
1522
1523 krping_format_send(cb, cb->rdma_dma_addr);
1524 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1525 if (ret) {
1526 printk(KERN_ERR PFX "post send error %d\n", ret);
1527 break;
1528 }
1529
1530 /* Wait for the server to say the RDMA Write is complete. */
1531 wait_event_interruptible(cb->sem,
1532 cb->state >= RDMA_WRITE_COMPLETE);
1533 if (cb->state != RDMA_WRITE_COMPLETE) {
1534 printk(KERN_ERR PFX
1535 "wait for RDMA_WRITE_COMPLETE state %d\n",
1536 cb->state);
1537 break;
1538 }
1539
1540 if (cb->validate)
1541 if (memcmp(cb->start_buf, cb->rdma_buf, cb->size)) {
1542 printk(KERN_ERR PFX "data mismatch!\n");
1543 break;
1544 }
1545
1546 if (cb->verbose)
1547 printk(KERN_INFO PFX "ping data: %s\n", cb->rdma_buf);
1548 #ifdef SLOW_KRPING
1549 wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
1550 #endif
1551 }
1552 }
1553
krping_rlat_test_client(struct krping_cb * cb)1554 static void krping_rlat_test_client(struct krping_cb *cb)
1555 {
1556 struct ib_send_wr *bad_wr;
1557 struct ib_wc wc;
1558 int ret;
1559
1560 cb->state = RDMA_READ_ADV;
1561
1562 /* Send STAG/TO/Len to client */
1563 krping_format_send(cb, cb->start_dma_addr);
1564 if (cb->state == ERROR) {
1565 printk(KERN_ERR PFX "krping_format_send failed\n");
1566 return;
1567 }
1568 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1569 if (ret) {
1570 printk(KERN_ERR PFX "post send error %d\n", ret);
1571 return;
1572 }
1573
1574 /* Spin waiting for send completion */
1575 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1576 if (ret < 0) {
1577 printk(KERN_ERR PFX "poll error %d\n", ret);
1578 return;
1579 }
1580 if (wc.status) {
1581 printk(KERN_ERR PFX "send completion error %d\n", wc.status);
1582 return;
1583 }
1584
1585 /* Spin waiting for server's Start STAG/TO/Len */
1586 while (cb->state < RDMA_WRITE_ADV) {
1587 krping_cq_event_handler(cb->cq, cb);
1588 }
1589
1590 #if 0
1591 {
1592 int i;
1593 struct timeval start, stop;
1594 time_t sec;
1595 suseconds_t usec;
1596 unsigned long long elapsed;
1597 struct ib_wc wc;
1598 struct ib_send_wr *bad_wr;
1599 int ne;
1600
1601 cb->rdma_sq_wr.wr.opcode = IB_WR_RDMA_WRITE;
1602 cb->rdma_sq_wr.rkey = cb->remote_rkey;
1603 cb->rdma_sq_wr.remote_addr = cb->remote_addr;
1604 cb->rdma_sq_wr.wr.sg_list->length = 0;
1605 cb->rdma_sq_wr.wr.num_sge = 0;
1606
1607 microtime(&start);
1608 for (i=0; i < 100000; i++) {
1609 if (ib_post_send(cb->qp, &cb->rdma_sq_wr.wr, &bad_wr)) {
1610 printk(KERN_ERR PFX "Couldn't post send\n");
1611 return;
1612 }
1613 do {
1614 ne = ib_poll_cq(cb->cq, 1, &wc);
1615 } while (ne == 0);
1616 if (ne < 0) {
1617 printk(KERN_ERR PFX "poll CQ failed %d\n", ne);
1618 return;
1619 }
1620 if (wc.status != IB_WC_SUCCESS) {
1621 printk(KERN_ERR PFX "Completion wth error at %s:\n",
1622 cb->server ? "server" : "client");
1623 printk(KERN_ERR PFX "Failed status %d: wr_id %d\n",
1624 wc.status, (int) wc.wr_id);
1625 return;
1626 }
1627 }
1628 microtime(&stop);
1629
1630 if (stop.tv_usec < start.tv_usec) {
1631 stop.tv_usec += 1000000;
1632 stop.tv_sec -= 1;
1633 }
1634 sec = stop.tv_sec - start.tv_sec;
1635 usec = stop.tv_usec - start.tv_usec;
1636 elapsed = sec * 1000000 + usec;
1637 printk(KERN_ERR PFX "0B-write-lat iters 100000 usec %llu\n", elapsed);
1638 }
1639 #endif
1640
1641 rlat_test(cb);
1642 }
1643
krping_wlat_test_client(struct krping_cb * cb)1644 static void krping_wlat_test_client(struct krping_cb *cb)
1645 {
1646 struct ib_send_wr *bad_wr;
1647 struct ib_wc wc;
1648 int ret;
1649
1650 cb->state = RDMA_READ_ADV;
1651
1652 /* Send STAG/TO/Len to client */
1653 krping_format_send(cb, cb->start_dma_addr);
1654 if (cb->state == ERROR) {
1655 printk(KERN_ERR PFX "krping_format_send failed\n");
1656 return;
1657 }
1658 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1659 if (ret) {
1660 printk(KERN_ERR PFX "post send error %d\n", ret);
1661 return;
1662 }
1663
1664 /* Spin waiting for send completion */
1665 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1666 if (ret < 0) {
1667 printk(KERN_ERR PFX "poll error %d\n", ret);
1668 return;
1669 }
1670 if (wc.status) {
1671 printk(KERN_ERR PFX "send completion error %d\n", wc.status);
1672 return;
1673 }
1674
1675 /* Spin waiting for server's Start STAG/TO/Len */
1676 while (cb->state < RDMA_WRITE_ADV) {
1677 krping_cq_event_handler(cb->cq, cb);
1678 }
1679
1680 wlat_test(cb);
1681 }
1682
krping_bw_test_client(struct krping_cb * cb)1683 static void krping_bw_test_client(struct krping_cb *cb)
1684 {
1685 struct ib_send_wr *bad_wr;
1686 struct ib_wc wc;
1687 int ret;
1688
1689 cb->state = RDMA_READ_ADV;
1690
1691 /* Send STAG/TO/Len to client */
1692 krping_format_send(cb, cb->start_dma_addr);
1693 if (cb->state == ERROR) {
1694 printk(KERN_ERR PFX "krping_format_send failed\n");
1695 return;
1696 }
1697 ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1698 if (ret) {
1699 printk(KERN_ERR PFX "post send error %d\n", ret);
1700 return;
1701 }
1702
1703 /* Spin waiting for send completion */
1704 while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1705 if (ret < 0) {
1706 printk(KERN_ERR PFX "poll error %d\n", ret);
1707 return;
1708 }
1709 if (wc.status) {
1710 printk(KERN_ERR PFX "send completion error %d\n", wc.status);
1711 return;
1712 }
1713
1714 /* Spin waiting for server's Start STAG/TO/Len */
1715 while (cb->state < RDMA_WRITE_ADV) {
1716 krping_cq_event_handler(cb->cq, cb);
1717 }
1718
1719 bw_test(cb);
1720 }
1721
1722 /*
1723 * Manual qp flush test
1724 */
flush_qp(struct krping_cb * cb)1725 static void flush_qp(struct krping_cb *cb)
1726 {
1727 struct ib_send_wr wr = { 0 }, *bad;
1728 struct ib_recv_wr recv_wr = { 0 }, *recv_bad;
1729 struct ib_wc wc;
1730 int ret;
1731 int flushed = 0;
1732 int ccnt = 0;
1733
1734 rdma_disconnect(cb->cm_id);
1735 DEBUG_LOG("disconnected!\n");
1736
1737 wr.opcode = IB_WR_SEND;
1738 wr.wr_id = 0xdeadbeefcafebabe;
1739 ret = ib_post_send(cb->qp, &wr, &bad);
1740 if (ret) {
1741 printk(KERN_ERR PFX "%s post_send failed ret %d\n", __func__, ret);
1742 return;
1743 }
1744
1745 recv_wr.wr_id = 0xcafebabedeadbeef;
1746 ret = ib_post_recv(cb->qp, &recv_wr, &recv_bad);
1747 if (ret) {
1748 printk(KERN_ERR PFX "%s post_recv failed ret %d\n", __func__, ret);
1749 return;
1750 }
1751
1752 /* poll until the flush WRs complete */
1753 do {
1754 ret = ib_poll_cq(cb->cq, 1, &wc);
1755 if (ret < 0) {
1756 printk(KERN_ERR PFX "ib_poll_cq failed %d\n", ret);
1757 return;
1758 }
1759 if (ret == 0)
1760 continue;
1761 ccnt++;
1762 if (wc.wr_id == 0xdeadbeefcafebabe ||
1763 wc.wr_id == 0xcafebabedeadbeef)
1764 flushed++;
1765 } while (flushed != 2);
1766 DEBUG_LOG("qp_flushed! ccnt %u\n", ccnt);
1767 }
1768
krping_fr_test(struct krping_cb * cb)1769 static void krping_fr_test(struct krping_cb *cb)
1770 {
1771 struct ib_send_wr inv, *bad;
1772 struct ib_reg_wr fr;
1773 struct ib_wc wc;
1774 u8 key = 0;
1775 struct ib_mr *mr;
1776 int ret;
1777 int size = cb->size;
1778 int plen = (((size - 1) & PAGE_MASK) + PAGE_SIZE) >> PAGE_SHIFT;
1779 unsigned long start;
1780 int count = 0;
1781 int scnt = 0;
1782 struct scatterlist sg = {0};
1783
1784 mr = ib_alloc_mr(cb->pd, IB_MR_TYPE_MEM_REG, plen);
1785 if (IS_ERR(mr)) {
1786 printk(KERN_ERR PFX "ib_alloc_mr failed %ld\n", PTR_ERR(mr));
1787 return;
1788 }
1789
1790 sg_dma_address(&sg) = (dma_addr_t)0xcafebabe0000ULL;
1791 sg_dma_len(&sg) = size;
1792 ret = ib_map_mr_sg(mr, &sg, 1, NULL, PAGE_SIZE);
1793 if (ret <= 0) {
1794 printk(KERN_ERR PFX "ib_map_mr_sge err %d\n", ret);
1795 goto err2;
1796 }
1797
1798 memset(&fr, 0, sizeof fr);
1799 fr.wr.opcode = IB_WR_REG_MR;
1800 fr.access = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE;
1801 fr.mr = mr;
1802 fr.wr.next = &inv;
1803
1804 memset(&inv, 0, sizeof inv);
1805 inv.opcode = IB_WR_LOCAL_INV;
1806 inv.send_flags = IB_SEND_SIGNALED;
1807
1808 DEBUG_LOG("fr_test: stag index 0x%x plen %u size %u depth %u\n", mr->rkey >> 8, plen, cb->size, cb->txdepth);
1809 start = time_uptime;
1810 while (!cb->count || count <= cb->count) {
1811 if (SIGPENDING(curthread)) {
1812 printk(KERN_ERR PFX "signal!\n");
1813 break;
1814 }
1815 if ((time_uptime - start) >= 9) {
1816 DEBUG_LOG("fr_test: pausing 1 second! count %u latest size %u plen %u\n", count, size, plen);
1817 wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
1818 if (cb->state == ERROR)
1819 break;
1820 start = time_uptime;
1821 }
1822 while (scnt < (cb->txdepth>>1)) {
1823 ib_update_fast_reg_key(mr, ++key);
1824 fr.key = mr->rkey;
1825 inv.ex.invalidate_rkey = mr->rkey;
1826
1827 size = arc4random() % cb->size;
1828 if (size == 0)
1829 size = cb->size;
1830 sg_dma_len(&sg) = size;
1831 ret = ib_map_mr_sg(mr, &sg, 1, NULL, PAGE_SIZE);
1832 if (ret <= 0) {
1833 printk(KERN_ERR PFX "ib_map_mr_sge err %d\n", ret);
1834 goto err2;
1835 }
1836 ret = ib_post_send(cb->qp, &fr.wr, &bad);
1837 if (ret) {
1838 printk(KERN_ERR PFX "ib_post_send failed %d\n", ret);
1839 goto err2;
1840 }
1841 scnt++;
1842 }
1843
1844 ret = ib_poll_cq(cb->cq, 1, &wc);
1845 if (ret < 0) {
1846 printk(KERN_ERR PFX "ib_poll_cq failed %d\n", ret);
1847 goto err2;
1848 }
1849 if (ret == 1) {
1850 if (wc.status) {
1851 printk(KERN_ERR PFX "completion error %u\n", wc.status);
1852 goto err2;
1853 }
1854 count++;
1855 scnt--;
1856 }
1857 }
1858 err2:
1859 flush_qp(cb);
1860 DEBUG_LOG("fr_test: done!\n");
1861 ib_dereg_mr(mr);
1862 }
1863
krping_connect_client(struct krping_cb * cb)1864 static int krping_connect_client(struct krping_cb *cb)
1865 {
1866 struct rdma_conn_param conn_param;
1867 int ret;
1868
1869 memset(&conn_param, 0, sizeof conn_param);
1870 conn_param.responder_resources = 1;
1871 conn_param.initiator_depth = 1;
1872 conn_param.retry_count = 10;
1873
1874 ret = rdma_connect(cb->cm_id, &conn_param);
1875 if (ret) {
1876 printk(KERN_ERR PFX "rdma_connect error %d\n", ret);
1877 return ret;
1878 }
1879
1880 wait_event_interruptible(cb->sem, cb->state >= CONNECTED);
1881 if (cb->state == ERROR) {
1882 printk(KERN_ERR PFX "wait for CONNECTED state %d\n", cb->state);
1883 return -1;
1884 }
1885
1886 DEBUG_LOG("rdma_connect successful\n");
1887 return 0;
1888 }
1889
krping_bind_client(struct krping_cb * cb)1890 static int krping_bind_client(struct krping_cb *cb)
1891 {
1892 struct sockaddr_storage sin;
1893 int ret;
1894
1895 fill_sockaddr(&sin, cb);
1896
1897 ret = rdma_resolve_addr(cb->cm_id, NULL, (struct sockaddr *)&sin, 2000);
1898 if (ret) {
1899 printk(KERN_ERR PFX "rdma_resolve_addr error %d\n", ret);
1900 return ret;
1901 }
1902
1903 wait_event_interruptible(cb->sem, cb->state >= ROUTE_RESOLVED);
1904 if (cb->state != ROUTE_RESOLVED) {
1905 printk(KERN_ERR PFX
1906 "addr/route resolution did not resolve: state %d\n",
1907 cb->state);
1908 return -EINTR;
1909 }
1910
1911 if (!reg_supported(cb->cm_id->device))
1912 return -EINVAL;
1913
1914 DEBUG_LOG("rdma_resolve_addr - rdma_resolve_route successful\n");
1915 return 0;
1916 }
1917
krping_run_client(struct krping_cb * cb)1918 static void krping_run_client(struct krping_cb *cb)
1919 {
1920 struct ib_recv_wr *bad_wr;
1921 int ret;
1922
1923 /* set type of service, if any */
1924 if (cb->tos != 0)
1925 rdma_set_service_type(cb->cm_id, cb->tos);
1926
1927 ret = krping_bind_client(cb);
1928 if (ret)
1929 return;
1930
1931 ret = krping_setup_qp(cb, cb->cm_id);
1932 if (ret) {
1933 printk(KERN_ERR PFX "setup_qp failed: %d\n", ret);
1934 return;
1935 }
1936
1937 ret = krping_setup_buffers(cb);
1938 if (ret) {
1939 printk(KERN_ERR PFX "krping_setup_buffers failed: %d\n", ret);
1940 goto err1;
1941 }
1942
1943 ret = ib_post_recv(cb->qp, &cb->rq_wr, &bad_wr);
1944 if (ret) {
1945 printk(KERN_ERR PFX "ib_post_recv failed: %d\n", ret);
1946 goto err2;
1947 }
1948
1949 ret = krping_connect_client(cb);
1950 if (ret) {
1951 printk(KERN_ERR PFX "connect error %d\n", ret);
1952 goto err2;
1953 }
1954
1955 if (cb->wlat)
1956 krping_wlat_test_client(cb);
1957 else if (cb->rlat)
1958 krping_rlat_test_client(cb);
1959 else if (cb->bw)
1960 krping_bw_test_client(cb);
1961 else if (cb->frtest)
1962 krping_fr_test(cb);
1963 else
1964 krping_test_client(cb);
1965 rdma_disconnect(cb->cm_id);
1966 err2:
1967 krping_free_buffers(cb);
1968 err1:
1969 krping_free_qp(cb);
1970 }
1971
1972 static uint16_t
krping_get_ipv6_scope_id(char * name)1973 krping_get_ipv6_scope_id(char *name)
1974 {
1975 struct ifnet *ifp;
1976 uint16_t retval;
1977
1978 if (name == NULL)
1979 return (0);
1980 CURVNET_SET_QUIET(TD_TO_VNET(curthread));
1981 ifp = ifunit_ref(name);
1982 CURVNET_RESTORE();
1983 if (ifp == NULL)
1984 return (0);
1985 retval = ifp->if_index;
1986 if_rele(ifp);
1987 return (retval);
1988 }
1989
krping_doit(char * cmd)1990 int krping_doit(char *cmd)
1991 {
1992 struct krping_cb *cb;
1993 int op;
1994 int ret = 0;
1995 char *optarg;
1996 char *scope;
1997 unsigned long optint;
1998
1999 cb = kzalloc(sizeof(*cb), GFP_KERNEL);
2000 if (!cb)
2001 return -ENOMEM;
2002
2003 mutex_lock(&krping_mutex);
2004 list_add_tail(&cb->list, &krping_cbs);
2005 mutex_unlock(&krping_mutex);
2006
2007 cb->server = -1;
2008 cb->state = IDLE;
2009 cb->size = 64;
2010 cb->txdepth = RPING_SQ_DEPTH;
2011 init_waitqueue_head(&cb->sem);
2012
2013 while ((op = krping_getopt("krping", &cmd, krping_opts, NULL, &optarg,
2014 &optint)) != 0) {
2015 switch (op) {
2016 case 'a':
2017 cb->addr_str = optarg;
2018 cb->addr_type = AF_INET;
2019 DEBUG_LOG("ipaddr (%s)\n", optarg);
2020 if (inet_pton(AF_INET, optarg, cb->addr) != 1) {
2021 printk(KERN_ERR PFX "bad addr string %s\n",
2022 optarg);
2023 ret = EINVAL;
2024 }
2025 break;
2026 case 'A':
2027 cb->addr_str = optarg;
2028 cb->addr_type = AF_INET6;
2029 DEBUG_LOG("ipv6addr (%s)\n", optarg);
2030 scope = strstr(optarg, "%");
2031 /* extract scope ID, if any */
2032 if (scope != NULL)
2033 *scope++ = 0;
2034 /* extract IPv6 network address */
2035 if (inet_pton(AF_INET6, optarg, cb->addr) != 1) {
2036 printk(KERN_ERR PFX "bad addr string %s\n",
2037 optarg);
2038 ret = EINVAL;
2039 } else if (IN6_IS_SCOPE_LINKLOCAL((struct in6_addr *)cb->addr) ||
2040 IN6_IS_ADDR_MC_INTFACELOCAL((struct in6_addr *)cb->addr)) {
2041 uint16_t scope_id = krping_get_ipv6_scope_id(scope);
2042 DEBUG_LOG("ipv6 scope ID = %d\n", scope_id);
2043 cb->addr[2] = scope_id >> 8;
2044 cb->addr[3] = scope_id & 0xFF;
2045 }
2046 break;
2047 case 'p':
2048 cb->port = htons(optint);
2049 DEBUG_LOG("port %d\n", (int)optint);
2050 break;
2051 case 'P':
2052 cb->poll = 1;
2053 DEBUG_LOG("server\n");
2054 break;
2055 case 's':
2056 cb->server = 1;
2057 DEBUG_LOG("server\n");
2058 break;
2059 case 'c':
2060 cb->server = 0;
2061 DEBUG_LOG("client\n");
2062 break;
2063 case 'S':
2064 cb->size = optint;
2065 if ((cb->size < 1) ||
2066 (cb->size > RPING_BUFSIZE)) {
2067 printk(KERN_ERR PFX "Invalid size %d "
2068 "(valid range is 1 to %d)\n",
2069 cb->size, RPING_BUFSIZE);
2070 ret = EINVAL;
2071 } else
2072 DEBUG_LOG("size %d\n", (int)optint);
2073 break;
2074 case 'C':
2075 cb->count = optint;
2076 if (cb->count < 0) {
2077 printk(KERN_ERR PFX "Invalid count %d\n",
2078 cb->count);
2079 ret = EINVAL;
2080 } else
2081 DEBUG_LOG("count %d\n", (int) cb->count);
2082 break;
2083 case 'v':
2084 cb->verbose++;
2085 DEBUG_LOG("verbose\n");
2086 break;
2087 case 'V':
2088 cb->validate++;
2089 DEBUG_LOG("validate data\n");
2090 break;
2091 case 'l':
2092 cb->wlat++;
2093 break;
2094 case 'L':
2095 cb->rlat++;
2096 break;
2097 case 'B':
2098 cb->bw++;
2099 break;
2100 case 'd':
2101 cb->duplex++;
2102 break;
2103 case 'I':
2104 cb->server_invalidate = 1;
2105 break;
2106 case 't':
2107 cb->tos = optint;
2108 DEBUG_LOG("type of service, tos=%d\n", (int) cb->tos);
2109 break;
2110 case 'T':
2111 cb->txdepth = optint;
2112 DEBUG_LOG("txdepth %d\n", (int) cb->txdepth);
2113 break;
2114 case 'Z':
2115 cb->local_dma_lkey = 1;
2116 DEBUG_LOG("using local dma lkey\n");
2117 break;
2118 case 'R':
2119 cb->read_inv = 1;
2120 DEBUG_LOG("using read-with-inv\n");
2121 break;
2122 case 'f':
2123 cb->frtest = 1;
2124 DEBUG_LOG("fast-reg test!\n");
2125 break;
2126 default:
2127 printk(KERN_ERR PFX "unknown opt %s\n", optarg);
2128 ret = -EINVAL;
2129 break;
2130 }
2131 }
2132 if (ret)
2133 goto out;
2134
2135 if (cb->server == -1) {
2136 printk(KERN_ERR PFX "must be either client or server\n");
2137 ret = -EINVAL;
2138 goto out;
2139 }
2140
2141 if (cb->server && cb->frtest) {
2142 printk(KERN_ERR PFX "must be client to run frtest\n");
2143 ret = -EINVAL;
2144 goto out;
2145 }
2146
2147 if ((cb->frtest + cb->bw + cb->rlat + cb->wlat) > 1) {
2148 printk(KERN_ERR PFX "Pick only one test: fr, bw, rlat, wlat\n");
2149 ret = -EINVAL;
2150 goto out;
2151 }
2152
2153 if (cb->wlat || cb->rlat || cb->bw) {
2154 printk(KERN_ERR PFX "wlat, rlat, and bw tests only support mem_mode MR - which is no longer supported\n");
2155 ret = -EINVAL;
2156 goto out;
2157 }
2158
2159 cb->cm_id = rdma_create_id(TD_TO_VNET(curthread), krping_cma_event_handler, cb, RDMA_PS_TCP, IB_QPT_RC);
2160 if (IS_ERR(cb->cm_id)) {
2161 ret = PTR_ERR(cb->cm_id);
2162 printk(KERN_ERR PFX "rdma_create_id error %d\n", ret);
2163 goto out;
2164 }
2165 DEBUG_LOG("created cm_id %p\n", cb->cm_id);
2166
2167 if (cb->server)
2168 krping_run_server(cb);
2169 else
2170 krping_run_client(cb);
2171
2172 DEBUG_LOG("destroy cm_id %p\n", cb->cm_id);
2173 rdma_destroy_id(cb->cm_id);
2174 out:
2175 mutex_lock(&krping_mutex);
2176 list_del(&cb->list);
2177 mutex_unlock(&krping_mutex);
2178 kfree(cb);
2179 return ret;
2180 }
2181
2182 void
krping_walk_cb_list(void (* f)(struct krping_stats *,void *),void * arg)2183 krping_walk_cb_list(void (*f)(struct krping_stats *, void *), void *arg)
2184 {
2185 struct krping_cb *cb;
2186
2187 mutex_lock(&krping_mutex);
2188 list_for_each_entry(cb, &krping_cbs, list)
2189 (*f)(cb->pd ? &cb->stats : NULL, arg);
2190 mutex_unlock(&krping_mutex);
2191 }
2192
2193 void
krping_cancel_all(void)2194 krping_cancel_all(void)
2195 {
2196 struct krping_cb *cb;
2197
2198 mutex_lock(&krping_mutex);
2199 list_for_each_entry(cb, &krping_cbs, list) {
2200 cb->state = ERROR;
2201 wake_up_interruptible(&cb->sem);
2202 }
2203 mutex_unlock(&krping_mutex);
2204 }
2205
2206