1 /* SPDX-License-Identifier: BSD-3-Clause
2 * Copyright(c) 2016-2017 Intel Corporation
3 */
4
5 #include <stdio.h>
6 #include <string.h>
7 #include <stdint.h>
8 #include <errno.h>
9 #include <unistd.h>
10 #include <sys/queue.h>
11
12 #include <rte_memory.h>
13 #include <rte_launch.h>
14 #include <rte_eal.h>
15 #include <rte_per_lcore.h>
16 #include <rte_lcore.h>
17 #include <rte_debug.h>
18 #include <rte_ethdev.h>
19 #include <rte_cycles.h>
20 #include <rte_eventdev.h>
21 #include <rte_pause.h>
22 #include <rte_service.h>
23 #include <rte_service_component.h>
24 #include <rte_bus_vdev.h>
25
26 #include "sw_evdev.h"
27
28 #define MAX_PORTS 16
29 #define MAX_QIDS 16
30 #define NUM_PACKETS (1<<18)
31 #define DEQUEUE_DEPTH 128
32
33 static int evdev;
34
35 struct test {
36 struct rte_mempool *mbuf_pool;
37 uint8_t port[MAX_PORTS];
38 uint8_t qid[MAX_QIDS];
39 int nb_qids;
40 uint32_t service_id;
41 };
42
43 typedef uint8_t counter_dynfield_t;
44 static int counter_dynfield_offset = -1;
45
46 static inline counter_dynfield_t *
counter_field(struct rte_mbuf * mbuf)47 counter_field(struct rte_mbuf *mbuf)
48 {
49 return RTE_MBUF_DYNFIELD(mbuf, \
50 counter_dynfield_offset, counter_dynfield_t *);
51 }
52
53 static struct rte_event release_ev;
54
55 static inline struct rte_mbuf *
rte_gen_arp(int portid,struct rte_mempool * mp)56 rte_gen_arp(int portid, struct rte_mempool *mp)
57 {
58 /*
59 * len = 14 + 46
60 * ARP, Request who-has 10.0.0.1 tell 10.0.0.2, length 46
61 */
62 static const uint8_t arp_request[] = {
63 /*0x0000:*/ 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xec, 0xa8,
64 0x6b, 0xfd, 0x02, 0x29, 0x08, 0x06, 0x00, 0x01,
65 /*0x0010:*/ 0x08, 0x00, 0x06, 0x04, 0x00, 0x01, 0xec, 0xa8,
66 0x6b, 0xfd, 0x02, 0x29, 0x0a, 0x00, 0x00, 0x01,
67 /*0x0020:*/ 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x0a, 0x00,
68 0x00, 0x02, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
69 /*0x0030:*/ 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
70 0x00, 0x00, 0x00, 0x00
71 };
72 struct rte_mbuf *m;
73 int pkt_len = sizeof(arp_request) - 1;
74
75 m = rte_pktmbuf_alloc(mp);
76 if (!m)
77 return 0;
78
79 memcpy((void *)((uintptr_t)m->buf_addr + m->data_off),
80 arp_request, pkt_len);
81 rte_pktmbuf_pkt_len(m) = pkt_len;
82 rte_pktmbuf_data_len(m) = pkt_len;
83
84 RTE_SET_USED(portid);
85
86 return m;
87 }
88
89 static void
xstats_print(void)90 xstats_print(void)
91 {
92 const uint32_t XSTATS_MAX = 1024;
93 uint32_t i;
94 uint32_t ids[XSTATS_MAX];
95 uint64_t values[XSTATS_MAX];
96 struct rte_event_dev_xstats_name xstats_names[XSTATS_MAX];
97
98 for (i = 0; i < XSTATS_MAX; i++)
99 ids[i] = i;
100
101 /* Device names / values */
102 int ret = rte_event_dev_xstats_names_get(evdev,
103 RTE_EVENT_DEV_XSTATS_DEVICE, 0,
104 xstats_names, ids, XSTATS_MAX);
105 if (ret < 0) {
106 printf("%d: xstats names get() returned error\n",
107 __LINE__);
108 return;
109 }
110 ret = rte_event_dev_xstats_get(evdev,
111 RTE_EVENT_DEV_XSTATS_DEVICE,
112 0, ids, values, ret);
113 if (ret > (signed int)XSTATS_MAX)
114 printf("%s %d: more xstats available than space\n",
115 __func__, __LINE__);
116 for (i = 0; (signed int)i < ret; i++) {
117 printf("%d : %s : %"PRIu64"\n",
118 i, xstats_names[i].name, values[i]);
119 }
120
121 /* Port names / values */
122 ret = rte_event_dev_xstats_names_get(evdev,
123 RTE_EVENT_DEV_XSTATS_PORT, 0,
124 xstats_names, ids, XSTATS_MAX);
125 ret = rte_event_dev_xstats_get(evdev,
126 RTE_EVENT_DEV_XSTATS_PORT, 1,
127 ids, values, ret);
128 if (ret > (signed int)XSTATS_MAX)
129 printf("%s %d: more xstats available than space\n",
130 __func__, __LINE__);
131 for (i = 0; (signed int)i < ret; i++) {
132 printf("%d : %s : %"PRIu64"\n",
133 i, xstats_names[i].name, values[i]);
134 }
135
136 /* Queue names / values */
137 ret = rte_event_dev_xstats_names_get(evdev,
138 RTE_EVENT_DEV_XSTATS_QUEUE, 0,
139 xstats_names, ids, XSTATS_MAX);
140 ret = rte_event_dev_xstats_get(evdev,
141 RTE_EVENT_DEV_XSTATS_QUEUE,
142 1, ids, values, ret);
143 if (ret > (signed int)XSTATS_MAX)
144 printf("%s %d: more xstats available than space\n",
145 __func__, __LINE__);
146 for (i = 0; (signed int)i < ret; i++) {
147 printf("%d : %s : %"PRIu64"\n",
148 i, xstats_names[i].name, values[i]);
149 }
150 }
151
152 /* initialization and config */
153 static inline int
init(struct test * t,int nb_queues,int nb_ports)154 init(struct test *t, int nb_queues, int nb_ports)
155 {
156 struct rte_event_dev_config config = {
157 .nb_event_queues = nb_queues,
158 .nb_event_ports = nb_ports,
159 .nb_event_queue_flows = 1024,
160 .nb_events_limit = 4096,
161 .nb_event_port_dequeue_depth = DEQUEUE_DEPTH,
162 .nb_event_port_enqueue_depth = 128,
163 };
164 int ret;
165
166 void *temp = t->mbuf_pool; /* save and restore mbuf pool */
167
168 memset(t, 0, sizeof(*t));
169 t->mbuf_pool = temp;
170
171 ret = rte_event_dev_configure(evdev, &config);
172 if (ret < 0)
173 printf("%d: Error configuring device\n", __LINE__);
174 return ret;
175 };
176
177 static inline int
create_ports(struct test * t,int num_ports)178 create_ports(struct test *t, int num_ports)
179 {
180 int i;
181 static const struct rte_event_port_conf conf = {
182 .new_event_threshold = 1024,
183 .dequeue_depth = 32,
184 .enqueue_depth = 64,
185 };
186 if (num_ports > MAX_PORTS)
187 return -1;
188
189 for (i = 0; i < num_ports; i++) {
190 if (rte_event_port_setup(evdev, i, &conf) < 0) {
191 printf("Error setting up port %d\n", i);
192 return -1;
193 }
194 t->port[i] = i;
195 }
196
197 return 0;
198 }
199
200 static inline int
create_lb_qids(struct test * t,int num_qids,uint32_t flags)201 create_lb_qids(struct test *t, int num_qids, uint32_t flags)
202 {
203 int i;
204
205 /* Q creation */
206 const struct rte_event_queue_conf conf = {
207 .schedule_type = flags,
208 .priority = RTE_EVENT_DEV_PRIORITY_NORMAL,
209 .nb_atomic_flows = 1024,
210 .nb_atomic_order_sequences = 1024,
211 };
212
213 for (i = t->nb_qids; i < t->nb_qids + num_qids; i++) {
214 if (rte_event_queue_setup(evdev, i, &conf) < 0) {
215 printf("%d: error creating qid %d\n", __LINE__, i);
216 return -1;
217 }
218 t->qid[i] = i;
219 }
220 t->nb_qids += num_qids;
221 if (t->nb_qids > MAX_QIDS)
222 return -1;
223
224 return 0;
225 }
226
227 static inline int
create_atomic_qids(struct test * t,int num_qids)228 create_atomic_qids(struct test *t, int num_qids)
229 {
230 return create_lb_qids(t, num_qids, RTE_SCHED_TYPE_ATOMIC);
231 }
232
233 static inline int
create_ordered_qids(struct test * t,int num_qids)234 create_ordered_qids(struct test *t, int num_qids)
235 {
236 return create_lb_qids(t, num_qids, RTE_SCHED_TYPE_ORDERED);
237 }
238
239
240 static inline int
create_unordered_qids(struct test * t,int num_qids)241 create_unordered_qids(struct test *t, int num_qids)
242 {
243 return create_lb_qids(t, num_qids, RTE_SCHED_TYPE_PARALLEL);
244 }
245
246 static inline int
create_directed_qids(struct test * t,int num_qids,const uint8_t ports[])247 create_directed_qids(struct test *t, int num_qids, const uint8_t ports[])
248 {
249 int i;
250
251 /* Q creation */
252 static const struct rte_event_queue_conf conf = {
253 .priority = RTE_EVENT_DEV_PRIORITY_NORMAL,
254 .event_queue_cfg = RTE_EVENT_QUEUE_CFG_SINGLE_LINK,
255 };
256
257 for (i = t->nb_qids; i < t->nb_qids + num_qids; i++) {
258 if (rte_event_queue_setup(evdev, i, &conf) < 0) {
259 printf("%d: error creating qid %d\n", __LINE__, i);
260 return -1;
261 }
262 t->qid[i] = i;
263
264 if (rte_event_port_link(evdev, ports[i - t->nb_qids],
265 &t->qid[i], NULL, 1) != 1) {
266 printf("%d: error creating link for qid %d\n",
267 __LINE__, i);
268 return -1;
269 }
270 }
271 t->nb_qids += num_qids;
272 if (t->nb_qids > MAX_QIDS)
273 return -1;
274
275 return 0;
276 }
277
278 /* destruction */
279 static inline int
cleanup(struct test * t __rte_unused)280 cleanup(struct test *t __rte_unused)
281 {
282 rte_event_dev_stop(evdev);
283 rte_event_dev_close(evdev);
284 return 0;
285 };
286
287 struct test_event_dev_stats {
288 uint64_t rx_pkts; /**< Total packets received */
289 uint64_t rx_dropped; /**< Total packets dropped (Eg Invalid QID) */
290 uint64_t tx_pkts; /**< Total packets transmitted */
291
292 /** Packets received on this port */
293 uint64_t port_rx_pkts[MAX_PORTS];
294 /** Packets dropped on this port */
295 uint64_t port_rx_dropped[MAX_PORTS];
296 /** Packets inflight on this port */
297 uint64_t port_inflight[MAX_PORTS];
298 /** Packets transmitted on this port */
299 uint64_t port_tx_pkts[MAX_PORTS];
300 /** Packets received on this qid */
301 uint64_t qid_rx_pkts[MAX_QIDS];
302 /** Packets dropped on this qid */
303 uint64_t qid_rx_dropped[MAX_QIDS];
304 /** Packets transmitted on this qid */
305 uint64_t qid_tx_pkts[MAX_QIDS];
306 };
307
308 static inline int
test_event_dev_stats_get(int dev_id,struct test_event_dev_stats * stats)309 test_event_dev_stats_get(int dev_id, struct test_event_dev_stats *stats)
310 {
311 static uint32_t i;
312 static uint32_t total_ids[3]; /* rx, tx and drop */
313 static uint32_t port_rx_pkts_ids[MAX_PORTS];
314 static uint32_t port_rx_dropped_ids[MAX_PORTS];
315 static uint32_t port_inflight_ids[MAX_PORTS];
316 static uint32_t port_tx_pkts_ids[MAX_PORTS];
317 static uint32_t qid_rx_pkts_ids[MAX_QIDS];
318 static uint32_t qid_rx_dropped_ids[MAX_QIDS];
319 static uint32_t qid_tx_pkts_ids[MAX_QIDS];
320
321
322 stats->rx_pkts = rte_event_dev_xstats_by_name_get(dev_id,
323 "dev_rx", &total_ids[0]);
324 stats->rx_dropped = rte_event_dev_xstats_by_name_get(dev_id,
325 "dev_drop", &total_ids[1]);
326 stats->tx_pkts = rte_event_dev_xstats_by_name_get(dev_id,
327 "dev_tx", &total_ids[2]);
328 for (i = 0; i < MAX_PORTS; i++) {
329 char name[32];
330 snprintf(name, sizeof(name), "port_%u_rx", i);
331 stats->port_rx_pkts[i] = rte_event_dev_xstats_by_name_get(
332 dev_id, name, &port_rx_pkts_ids[i]);
333 snprintf(name, sizeof(name), "port_%u_drop", i);
334 stats->port_rx_dropped[i] = rte_event_dev_xstats_by_name_get(
335 dev_id, name, &port_rx_dropped_ids[i]);
336 snprintf(name, sizeof(name), "port_%u_inflight", i);
337 stats->port_inflight[i] = rte_event_dev_xstats_by_name_get(
338 dev_id, name, &port_inflight_ids[i]);
339 snprintf(name, sizeof(name), "port_%u_tx", i);
340 stats->port_tx_pkts[i] = rte_event_dev_xstats_by_name_get(
341 dev_id, name, &port_tx_pkts_ids[i]);
342 }
343 for (i = 0; i < MAX_QIDS; i++) {
344 char name[32];
345 snprintf(name, sizeof(name), "qid_%u_rx", i);
346 stats->qid_rx_pkts[i] = rte_event_dev_xstats_by_name_get(
347 dev_id, name, &qid_rx_pkts_ids[i]);
348 snprintf(name, sizeof(name), "qid_%u_drop", i);
349 stats->qid_rx_dropped[i] = rte_event_dev_xstats_by_name_get(
350 dev_id, name, &qid_rx_dropped_ids[i]);
351 snprintf(name, sizeof(name), "qid_%u_tx", i);
352 stats->qid_tx_pkts[i] = rte_event_dev_xstats_by_name_get(
353 dev_id, name, &qid_tx_pkts_ids[i]);
354 }
355
356 return 0;
357 }
358
359 /* run_prio_packet_test
360 * This performs a basic packet priority check on the test instance passed in.
361 * It is factored out of the main priority tests as the same tests must be
362 * performed to ensure prioritization of each type of QID.
363 *
364 * Requirements:
365 * - An initialized test structure, including mempool
366 * - t->port[0] is initialized for both Enq / Deq of packets to the QID
367 * - t->qid[0] is the QID to be tested
368 * - if LB QID, the CQ must be mapped to the QID.
369 */
370 static int
run_prio_packet_test(struct test * t)371 run_prio_packet_test(struct test *t)
372 {
373 int err;
374 const uint32_t MAGIC_SEQN[] = {4711, 1234};
375 const uint32_t PRIORITY[] = {
376 RTE_EVENT_DEV_PRIORITY_NORMAL,
377 RTE_EVENT_DEV_PRIORITY_HIGHEST
378 };
379 unsigned int i;
380 for (i = 0; i < RTE_DIM(MAGIC_SEQN); i++) {
381 /* generate pkt and enqueue */
382 struct rte_event ev;
383 struct rte_mbuf *arp = rte_gen_arp(0, t->mbuf_pool);
384 if (!arp) {
385 printf("%d: gen of pkt failed\n", __LINE__);
386 return -1;
387 }
388 *rte_event_pmd_selftest_seqn(arp) = MAGIC_SEQN[i];
389
390 ev = (struct rte_event){
391 .priority = PRIORITY[i],
392 .op = RTE_EVENT_OP_NEW,
393 .queue_id = t->qid[0],
394 .mbuf = arp
395 };
396 err = rte_event_enqueue_burst(evdev, t->port[0], &ev, 1);
397 if (err != 1) {
398 printf("%d: error failed to enqueue\n", __LINE__);
399 return -1;
400 }
401 }
402
403 rte_service_run_iter_on_app_lcore(t->service_id, 1);
404
405 struct test_event_dev_stats stats;
406 err = test_event_dev_stats_get(evdev, &stats);
407 if (err) {
408 printf("%d: error failed to get stats\n", __LINE__);
409 return -1;
410 }
411
412 if (stats.port_rx_pkts[t->port[0]] != 2) {
413 printf("%d: error stats incorrect for directed port\n",
414 __LINE__);
415 rte_event_dev_dump(evdev, stdout);
416 return -1;
417 }
418
419 struct rte_event ev, ev2;
420 uint32_t deq_pkts;
421 deq_pkts = rte_event_dequeue_burst(evdev, t->port[0], &ev, 1, 0);
422 if (deq_pkts != 1) {
423 printf("%d: error failed to deq\n", __LINE__);
424 rte_event_dev_dump(evdev, stdout);
425 return -1;
426 }
427 if (*rte_event_pmd_selftest_seqn(ev.mbuf) != MAGIC_SEQN[1]) {
428 printf("%d: first packet out not highest priority\n",
429 __LINE__);
430 rte_event_dev_dump(evdev, stdout);
431 return -1;
432 }
433 rte_pktmbuf_free(ev.mbuf);
434
435 deq_pkts = rte_event_dequeue_burst(evdev, t->port[0], &ev2, 1, 0);
436 if (deq_pkts != 1) {
437 printf("%d: error failed to deq\n", __LINE__);
438 rte_event_dev_dump(evdev, stdout);
439 return -1;
440 }
441 if (*rte_event_pmd_selftest_seqn(ev2.mbuf) != MAGIC_SEQN[0]) {
442 printf("%d: second packet out not lower priority\n",
443 __LINE__);
444 rte_event_dev_dump(evdev, stdout);
445 return -1;
446 }
447 rte_pktmbuf_free(ev2.mbuf);
448
449 cleanup(t);
450 return 0;
451 }
452
453 static int
test_single_directed_packet(struct test * t)454 test_single_directed_packet(struct test *t)
455 {
456 const int rx_enq = 0;
457 const int wrk_enq = 2;
458 int err;
459
460 /* Create instance with 3 directed QIDs going to 3 ports */
461 if (init(t, 3, 3) < 0 ||
462 create_ports(t, 3) < 0 ||
463 create_directed_qids(t, 3, t->port) < 0)
464 return -1;
465
466 if (rte_event_dev_start(evdev) < 0) {
467 printf("%d: Error with start call\n", __LINE__);
468 return -1;
469 }
470
471 /************** FORWARD ****************/
472 struct rte_mbuf *arp = rte_gen_arp(0, t->mbuf_pool);
473 struct rte_event ev = {
474 .op = RTE_EVENT_OP_NEW,
475 .queue_id = wrk_enq,
476 .mbuf = arp,
477 };
478
479 if (!arp) {
480 printf("%d: gen of pkt failed\n", __LINE__);
481 return -1;
482 }
483
484 const uint32_t MAGIC_SEQN = 4711;
485 *rte_event_pmd_selftest_seqn(arp) = MAGIC_SEQN;
486
487 /* generate pkt and enqueue */
488 err = rte_event_enqueue_burst(evdev, rx_enq, &ev, 1);
489 if (err != 1) {
490 printf("%d: error failed to enqueue\n", __LINE__);
491 return -1;
492 }
493
494 /* Run schedule() as dir packets may need to be re-ordered */
495 rte_service_run_iter_on_app_lcore(t->service_id, 1);
496
497 struct test_event_dev_stats stats;
498 err = test_event_dev_stats_get(evdev, &stats);
499 if (err) {
500 printf("%d: error failed to get stats\n", __LINE__);
501 return -1;
502 }
503
504 if (stats.port_rx_pkts[rx_enq] != 1) {
505 printf("%d: error stats incorrect for directed port\n",
506 __LINE__);
507 return -1;
508 }
509
510 uint32_t deq_pkts;
511 deq_pkts = rte_event_dequeue_burst(evdev, wrk_enq, &ev, 1, 0);
512 if (deq_pkts != 1) {
513 printf("%d: error failed to deq\n", __LINE__);
514 return -1;
515 }
516
517 err = test_event_dev_stats_get(evdev, &stats);
518 if (stats.port_rx_pkts[wrk_enq] != 0 &&
519 stats.port_rx_pkts[wrk_enq] != 1) {
520 printf("%d: error directed stats post-dequeue\n", __LINE__);
521 return -1;
522 }
523
524 if (*rte_event_pmd_selftest_seqn(ev.mbuf) != MAGIC_SEQN) {
525 printf("%d: error magic sequence number not dequeued\n",
526 __LINE__);
527 return -1;
528 }
529
530 rte_pktmbuf_free(ev.mbuf);
531 cleanup(t);
532 return 0;
533 }
534
535 static int
test_directed_forward_credits(struct test * t)536 test_directed_forward_credits(struct test *t)
537 {
538 uint32_t i;
539 int32_t err;
540
541 if (init(t, 1, 1) < 0 ||
542 create_ports(t, 1) < 0 ||
543 create_directed_qids(t, 1, t->port) < 0)
544 return -1;
545
546 if (rte_event_dev_start(evdev) < 0) {
547 printf("%d: Error with start call\n", __LINE__);
548 return -1;
549 }
550
551 struct rte_event ev = {
552 .op = RTE_EVENT_OP_NEW,
553 .queue_id = 0,
554 };
555
556 for (i = 0; i < 1000; i++) {
557 err = rte_event_enqueue_burst(evdev, 0, &ev, 1);
558 if (err != 1) {
559 printf("%d: error failed to enqueue\n", __LINE__);
560 return -1;
561 }
562 rte_service_run_iter_on_app_lcore(t->service_id, 1);
563
564 uint32_t deq_pkts;
565 deq_pkts = rte_event_dequeue_burst(evdev, 0, &ev, 1, 0);
566 if (deq_pkts != 1) {
567 printf("%d: error failed to deq\n", __LINE__);
568 return -1;
569 }
570
571 /* re-write event to be a forward, and continue looping it */
572 ev.op = RTE_EVENT_OP_FORWARD;
573 }
574
575 cleanup(t);
576 return 0;
577 }
578
579
580 static int
test_priority_directed(struct test * t)581 test_priority_directed(struct test *t)
582 {
583 if (init(t, 1, 1) < 0 ||
584 create_ports(t, 1) < 0 ||
585 create_directed_qids(t, 1, t->port) < 0) {
586 printf("%d: Error initializing device\n", __LINE__);
587 return -1;
588 }
589
590 if (rte_event_dev_start(evdev) < 0) {
591 printf("%d: Error with start call\n", __LINE__);
592 return -1;
593 }
594
595 return run_prio_packet_test(t);
596 }
597
598 static int
test_priority_atomic(struct test * t)599 test_priority_atomic(struct test *t)
600 {
601 if (init(t, 1, 1) < 0 ||
602 create_ports(t, 1) < 0 ||
603 create_atomic_qids(t, 1) < 0) {
604 printf("%d: Error initializing device\n", __LINE__);
605 return -1;
606 }
607
608 /* map the QID */
609 if (rte_event_port_link(evdev, t->port[0], &t->qid[0], NULL, 1) != 1) {
610 printf("%d: error mapping qid to port\n", __LINE__);
611 return -1;
612 }
613 if (rte_event_dev_start(evdev) < 0) {
614 printf("%d: Error with start call\n", __LINE__);
615 return -1;
616 }
617
618 return run_prio_packet_test(t);
619 }
620
621 static int
test_priority_ordered(struct test * t)622 test_priority_ordered(struct test *t)
623 {
624 if (init(t, 1, 1) < 0 ||
625 create_ports(t, 1) < 0 ||
626 create_ordered_qids(t, 1) < 0) {
627 printf("%d: Error initializing device\n", __LINE__);
628 return -1;
629 }
630
631 /* map the QID */
632 if (rte_event_port_link(evdev, t->port[0], &t->qid[0], NULL, 1) != 1) {
633 printf("%d: error mapping qid to port\n", __LINE__);
634 return -1;
635 }
636 if (rte_event_dev_start(evdev) < 0) {
637 printf("%d: Error with start call\n", __LINE__);
638 return -1;
639 }
640
641 return run_prio_packet_test(t);
642 }
643
644 static int
test_priority_unordered(struct test * t)645 test_priority_unordered(struct test *t)
646 {
647 if (init(t, 1, 1) < 0 ||
648 create_ports(t, 1) < 0 ||
649 create_unordered_qids(t, 1) < 0) {
650 printf("%d: Error initializing device\n", __LINE__);
651 return -1;
652 }
653
654 /* map the QID */
655 if (rte_event_port_link(evdev, t->port[0], &t->qid[0], NULL, 1) != 1) {
656 printf("%d: error mapping qid to port\n", __LINE__);
657 return -1;
658 }
659 if (rte_event_dev_start(evdev) < 0) {
660 printf("%d: Error with start call\n", __LINE__);
661 return -1;
662 }
663
664 return run_prio_packet_test(t);
665 }
666
667 static int
burst_packets(struct test * t)668 burst_packets(struct test *t)
669 {
670 /************** CONFIG ****************/
671 uint32_t i;
672 int err;
673 int ret;
674
675 /* Create instance with 2 ports and 2 queues */
676 if (init(t, 2, 2) < 0 ||
677 create_ports(t, 2) < 0 ||
678 create_atomic_qids(t, 2) < 0) {
679 printf("%d: Error initializing device\n", __LINE__);
680 return -1;
681 }
682
683 /* CQ mapping to QID */
684 ret = rte_event_port_link(evdev, t->port[0], &t->qid[0], NULL, 1);
685 if (ret != 1) {
686 printf("%d: error mapping lb qid0\n", __LINE__);
687 return -1;
688 }
689 ret = rte_event_port_link(evdev, t->port[1], &t->qid[1], NULL, 1);
690 if (ret != 1) {
691 printf("%d: error mapping lb qid1\n", __LINE__);
692 return -1;
693 }
694
695 if (rte_event_dev_start(evdev) < 0) {
696 printf("%d: Error with start call\n", __LINE__);
697 return -1;
698 }
699
700 /************** FORWARD ****************/
701 const uint32_t rx_port = 0;
702 const uint32_t NUM_PKTS = 2;
703
704 for (i = 0; i < NUM_PKTS; i++) {
705 struct rte_mbuf *arp = rte_gen_arp(0, t->mbuf_pool);
706 if (!arp) {
707 printf("%d: error generating pkt\n", __LINE__);
708 return -1;
709 }
710
711 struct rte_event ev = {
712 .op = RTE_EVENT_OP_NEW,
713 .queue_id = i % 2,
714 .flow_id = i % 3,
715 .mbuf = arp,
716 };
717 /* generate pkt and enqueue */
718 err = rte_event_enqueue_burst(evdev, t->port[rx_port], &ev, 1);
719 if (err != 1) {
720 printf("%d: Failed to enqueue\n", __LINE__);
721 return -1;
722 }
723 }
724 rte_service_run_iter_on_app_lcore(t->service_id, 1);
725
726 /* Check stats for all NUM_PKTS arrived to sched core */
727 struct test_event_dev_stats stats;
728
729 err = test_event_dev_stats_get(evdev, &stats);
730 if (err) {
731 printf("%d: failed to get stats\n", __LINE__);
732 return -1;
733 }
734 if (stats.rx_pkts != NUM_PKTS || stats.tx_pkts != NUM_PKTS) {
735 printf("%d: Sched core didn't receive all %d pkts\n",
736 __LINE__, NUM_PKTS);
737 rte_event_dev_dump(evdev, stdout);
738 return -1;
739 }
740
741 uint32_t deq_pkts;
742 int p;
743
744 deq_pkts = 0;
745 /******** DEQ QID 1 *******/
746 do {
747 struct rte_event ev;
748 p = rte_event_dequeue_burst(evdev, t->port[0], &ev, 1, 0);
749 deq_pkts += p;
750 rte_pktmbuf_free(ev.mbuf);
751 } while (p);
752
753 if (deq_pkts != NUM_PKTS/2) {
754 printf("%d: Half of NUM_PKTS didn't arrive at port 1\n",
755 __LINE__);
756 return -1;
757 }
758
759 /******** DEQ QID 2 *******/
760 deq_pkts = 0;
761 do {
762 struct rte_event ev;
763 p = rte_event_dequeue_burst(evdev, t->port[1], &ev, 1, 0);
764 deq_pkts += p;
765 rte_pktmbuf_free(ev.mbuf);
766 } while (p);
767 if (deq_pkts != NUM_PKTS/2) {
768 printf("%d: Half of NUM_PKTS didn't arrive at port 2\n",
769 __LINE__);
770 return -1;
771 }
772
773 cleanup(t);
774 return 0;
775 }
776
777 static int
abuse_inflights(struct test * t)778 abuse_inflights(struct test *t)
779 {
780 const int rx_enq = 0;
781 const int wrk_enq = 2;
782 int err;
783
784 /* Create instance with 4 ports */
785 if (init(t, 1, 4) < 0 ||
786 create_ports(t, 4) < 0 ||
787 create_atomic_qids(t, 1) < 0) {
788 printf("%d: Error initializing device\n", __LINE__);
789 return -1;
790 }
791
792 /* CQ mapping to QID */
793 err = rte_event_port_link(evdev, t->port[wrk_enq], NULL, NULL, 0);
794 if (err != 1) {
795 printf("%d: error mapping lb qid\n", __LINE__);
796 cleanup(t);
797 return -1;
798 }
799
800 if (rte_event_dev_start(evdev) < 0) {
801 printf("%d: Error with start call\n", __LINE__);
802 return -1;
803 }
804
805 /* Enqueue op only */
806 err = rte_event_enqueue_burst(evdev, t->port[rx_enq], &release_ev, 1);
807 if (err != 1) {
808 printf("%d: Failed to enqueue\n", __LINE__);
809 return -1;
810 }
811
812 /* schedule */
813 rte_service_run_iter_on_app_lcore(t->service_id, 1);
814
815 struct test_event_dev_stats stats;
816
817 err = test_event_dev_stats_get(evdev, &stats);
818 if (err) {
819 printf("%d: failed to get stats\n", __LINE__);
820 return -1;
821 }
822
823 if (stats.rx_pkts != 0 ||
824 stats.tx_pkts != 0 ||
825 stats.port_inflight[wrk_enq] != 0) {
826 printf("%d: Sched core didn't handle pkt as expected\n",
827 __LINE__);
828 return -1;
829 }
830
831 cleanup(t);
832 return 0;
833 }
834
835 static int
xstats_tests(struct test * t)836 xstats_tests(struct test *t)
837 {
838 const int wrk_enq = 2;
839 int err;
840
841 /* Create instance with 4 ports */
842 if (init(t, 1, 4) < 0 ||
843 create_ports(t, 4) < 0 ||
844 create_atomic_qids(t, 1) < 0) {
845 printf("%d: Error initializing device\n", __LINE__);
846 return -1;
847 }
848
849 /* CQ mapping to QID */
850 err = rte_event_port_link(evdev, t->port[wrk_enq], NULL, NULL, 0);
851 if (err != 1) {
852 printf("%d: error mapping lb qid\n", __LINE__);
853 cleanup(t);
854 return -1;
855 }
856
857 if (rte_event_dev_start(evdev) < 0) {
858 printf("%d: Error with start call\n", __LINE__);
859 return -1;
860 }
861
862 const uint32_t XSTATS_MAX = 1024;
863
864 uint32_t i;
865 uint32_t ids[XSTATS_MAX];
866 uint64_t values[XSTATS_MAX];
867 struct rte_event_dev_xstats_name xstats_names[XSTATS_MAX];
868
869 for (i = 0; i < XSTATS_MAX; i++)
870 ids[i] = i;
871
872 /* Device names / values */
873 int ret = rte_event_dev_xstats_names_get(evdev,
874 RTE_EVENT_DEV_XSTATS_DEVICE,
875 0, xstats_names, ids, XSTATS_MAX);
876 if (ret != 8) {
877 printf("%d: expected 8 stats, got return %d\n", __LINE__, ret);
878 return -1;
879 }
880 ret = rte_event_dev_xstats_get(evdev,
881 RTE_EVENT_DEV_XSTATS_DEVICE,
882 0, ids, values, ret);
883 if (ret != 8) {
884 printf("%d: expected 8 stats, got return %d\n", __LINE__, ret);
885 return -1;
886 }
887
888 /* Port names / values */
889 ret = rte_event_dev_xstats_names_get(evdev,
890 RTE_EVENT_DEV_XSTATS_PORT, 0,
891 xstats_names, ids, XSTATS_MAX);
892 if (ret != 21) {
893 printf("%d: expected 21 stats, got return %d\n", __LINE__, ret);
894 return -1;
895 }
896 ret = rte_event_dev_xstats_get(evdev,
897 RTE_EVENT_DEV_XSTATS_PORT, 0,
898 ids, values, ret);
899 if (ret != 21) {
900 printf("%d: expected 21 stats, got return %d\n", __LINE__, ret);
901 return -1;
902 }
903
904 /* Queue names / values */
905 ret = rte_event_dev_xstats_names_get(evdev,
906 RTE_EVENT_DEV_XSTATS_QUEUE,
907 0, xstats_names, ids, XSTATS_MAX);
908 if (ret != 16) {
909 printf("%d: expected 16 stats, got return %d\n", __LINE__, ret);
910 return -1;
911 }
912
913 /* NEGATIVE TEST: with wrong queue passed, 0 stats should be returned */
914 ret = rte_event_dev_xstats_get(evdev,
915 RTE_EVENT_DEV_XSTATS_QUEUE,
916 1, ids, values, ret);
917 if (ret != -EINVAL) {
918 printf("%d: expected 0 stats, got return %d\n", __LINE__, ret);
919 return -1;
920 }
921
922 ret = rte_event_dev_xstats_get(evdev,
923 RTE_EVENT_DEV_XSTATS_QUEUE,
924 0, ids, values, ret);
925 if (ret != 16) {
926 printf("%d: expected 16 stats, got return %d\n", __LINE__, ret);
927 return -1;
928 }
929
930 /* enqueue packets to check values */
931 for (i = 0; i < 3; i++) {
932 struct rte_event ev;
933 struct rte_mbuf *arp = rte_gen_arp(0, t->mbuf_pool);
934 if (!arp) {
935 printf("%d: gen of pkt failed\n", __LINE__);
936 return -1;
937 }
938 ev.queue_id = t->qid[i];
939 ev.op = RTE_EVENT_OP_NEW;
940 ev.mbuf = arp;
941 ev.flow_id = 7;
942 *rte_event_pmd_selftest_seqn(arp) = i;
943
944 int err = rte_event_enqueue_burst(evdev, t->port[0], &ev, 1);
945 if (err != 1) {
946 printf("%d: Failed to enqueue\n", __LINE__);
947 return -1;
948 }
949 }
950
951 rte_service_run_iter_on_app_lcore(t->service_id, 1);
952
953 /* Device names / values */
954 int num_stats = rte_event_dev_xstats_names_get(evdev,
955 RTE_EVENT_DEV_XSTATS_DEVICE, 0,
956 xstats_names, ids, XSTATS_MAX);
957 if (num_stats < 0)
958 goto fail;
959 ret = rte_event_dev_xstats_get(evdev,
960 RTE_EVENT_DEV_XSTATS_DEVICE,
961 0, ids, values, num_stats);
962 static const uint64_t expected[] = {3, 3, 0, 1, 0, 0, 4, 1};
963 for (i = 0; (signed int)i < ret; i++) {
964 if (expected[i] != values[i]) {
965 printf(
966 "%d Error xstat %d (id %d) %s : %"PRIu64
967 ", expect %"PRIu64"\n",
968 __LINE__, i, ids[i], xstats_names[i].name,
969 values[i], expected[i]);
970 goto fail;
971 }
972 }
973
974 ret = rte_event_dev_xstats_reset(evdev, RTE_EVENT_DEV_XSTATS_DEVICE,
975 0, NULL, 0);
976
977 /* ensure reset statistics are zero-ed */
978 static const uint64_t expected_zero[] = {0, 0, 0, 0, 0, 0, 0, 0};
979 ret = rte_event_dev_xstats_get(evdev,
980 RTE_EVENT_DEV_XSTATS_DEVICE,
981 0, ids, values, num_stats);
982 for (i = 0; (signed int)i < ret; i++) {
983 if (expected_zero[i] != values[i]) {
984 printf(
985 "%d Error, xstat %d (id %d) %s : %"PRIu64
986 ", expect %"PRIu64"\n",
987 __LINE__, i, ids[i], xstats_names[i].name,
988 values[i], expected_zero[i]);
989 goto fail;
990 }
991 }
992
993 /* port reset checks */
994 num_stats = rte_event_dev_xstats_names_get(evdev,
995 RTE_EVENT_DEV_XSTATS_PORT, 0,
996 xstats_names, ids, XSTATS_MAX);
997 if (num_stats < 0)
998 goto fail;
999 ret = rte_event_dev_xstats_get(evdev, RTE_EVENT_DEV_XSTATS_PORT,
1000 0, ids, values, num_stats);
1001
1002 static const uint64_t port_expected[] = {
1003 3 /* rx */,
1004 0 /* tx */,
1005 0 /* drop */,
1006 0 /* inflights */,
1007 0 /* avg pkt cycles */,
1008 29 /* credits */,
1009 0 /* rx ring used */,
1010 4096 /* rx ring free */,
1011 0 /* cq ring used */,
1012 32 /* cq ring free */,
1013 0 /* dequeue calls */,
1014 /* 10 dequeue burst buckets */
1015 0, 0, 0, 0, 0,
1016 0, 0, 0, 0, 0,
1017 };
1018 if (ret != RTE_DIM(port_expected)) {
1019 printf(
1020 "%s %d: wrong number of port stats (%d), expected %zu\n",
1021 __func__, __LINE__, ret, RTE_DIM(port_expected));
1022 }
1023
1024 for (i = 0; (signed int)i < ret; i++) {
1025 if (port_expected[i] != values[i]) {
1026 printf(
1027 "%s : %d: Error stat %s is %"PRIu64
1028 ", expected %"PRIu64"\n",
1029 __func__, __LINE__, xstats_names[i].name,
1030 values[i], port_expected[i]);
1031 goto fail;
1032 }
1033 }
1034
1035 ret = rte_event_dev_xstats_reset(evdev, RTE_EVENT_DEV_XSTATS_PORT,
1036 0, NULL, 0);
1037
1038 /* ensure reset statistics are zero-ed */
1039 static const uint64_t port_expected_zero[] = {
1040 0 /* rx */,
1041 0 /* tx */,
1042 0 /* drop */,
1043 0 /* inflights */,
1044 0 /* avg pkt cycles */,
1045 29 /* credits */,
1046 0 /* rx ring used */,
1047 4096 /* rx ring free */,
1048 0 /* cq ring used */,
1049 32 /* cq ring free */,
1050 0 /* dequeue calls */,
1051 /* 10 dequeue burst buckets */
1052 0, 0, 0, 0, 0,
1053 0, 0, 0, 0, 0,
1054 };
1055 ret = rte_event_dev_xstats_get(evdev,
1056 RTE_EVENT_DEV_XSTATS_PORT,
1057 0, ids, values, num_stats);
1058 for (i = 0; (signed int)i < ret; i++) {
1059 if (port_expected_zero[i] != values[i]) {
1060 printf(
1061 "%d, Error, xstat %d (id %d) %s : %"PRIu64
1062 ", expect %"PRIu64"\n",
1063 __LINE__, i, ids[i], xstats_names[i].name,
1064 values[i], port_expected_zero[i]);
1065 goto fail;
1066 }
1067 }
1068
1069 /* QUEUE STATS TESTS */
1070 num_stats = rte_event_dev_xstats_names_get(evdev,
1071 RTE_EVENT_DEV_XSTATS_QUEUE, 0,
1072 xstats_names, ids, XSTATS_MAX);
1073 ret = rte_event_dev_xstats_get(evdev, RTE_EVENT_DEV_XSTATS_QUEUE,
1074 0, ids, values, num_stats);
1075 if (ret < 0) {
1076 printf("xstats get returned %d\n", ret);
1077 goto fail;
1078 }
1079 if ((unsigned int)ret > XSTATS_MAX)
1080 printf("%s %d: more xstats available than space\n",
1081 __func__, __LINE__);
1082
1083 static const uint64_t queue_expected[] = {
1084 3 /* rx */,
1085 3 /* tx */,
1086 0 /* drop */,
1087 3 /* inflights */,
1088 0, 0, 0, 0, /* iq 0, 1, 2, 3 used */
1089 /* QID-to-Port: pinned_flows, packets */
1090 0, 0,
1091 0, 0,
1092 1, 3,
1093 0, 0,
1094 };
1095 for (i = 0; (signed int)i < ret; i++) {
1096 if (queue_expected[i] != values[i]) {
1097 printf(
1098 "%d, Error, xstat %d (id %d) %s : %"PRIu64
1099 ", expect %"PRIu64"\n",
1100 __LINE__, i, ids[i], xstats_names[i].name,
1101 values[i], queue_expected[i]);
1102 goto fail;
1103 }
1104 }
1105
1106 /* Reset the queue stats here */
1107 ret = rte_event_dev_xstats_reset(evdev,
1108 RTE_EVENT_DEV_XSTATS_QUEUE, 0,
1109 NULL,
1110 0);
1111
1112 /* Verify that the resettable stats are reset, and others are not */
1113 static const uint64_t queue_expected_zero[] = {
1114 0 /* rx */,
1115 0 /* tx */,
1116 0 /* drop */,
1117 3 /* inflight */,
1118 0, 0, 0, 0, /* 4 iq used */
1119 /* QID-to-Port: pinned_flows, packets */
1120 0, 0,
1121 0, 0,
1122 1, 0,
1123 0, 0,
1124 };
1125
1126 ret = rte_event_dev_xstats_get(evdev, RTE_EVENT_DEV_XSTATS_QUEUE, 0,
1127 ids, values, num_stats);
1128 int fails = 0;
1129 for (i = 0; (signed int)i < ret; i++) {
1130 if (queue_expected_zero[i] != values[i]) {
1131 printf(
1132 "%d, Error, xstat %d (id %d) %s : %"PRIu64
1133 ", expect %"PRIu64"\n",
1134 __LINE__, i, ids[i], xstats_names[i].name,
1135 values[i], queue_expected_zero[i]);
1136 fails++;
1137 }
1138 }
1139 if (fails) {
1140 printf("%d : %d of values were not as expected above\n",
1141 __LINE__, fails);
1142 goto fail;
1143 }
1144
1145 cleanup(t);
1146 return 0;
1147
1148 fail:
1149 rte_event_dev_dump(0, stdout);
1150 cleanup(t);
1151 return -1;
1152 }
1153
1154
1155 static int
xstats_id_abuse_tests(struct test * t)1156 xstats_id_abuse_tests(struct test *t)
1157 {
1158 int err;
1159 const uint32_t XSTATS_MAX = 1024;
1160 const uint32_t link_port = 2;
1161
1162 uint32_t ids[XSTATS_MAX];
1163 struct rte_event_dev_xstats_name xstats_names[XSTATS_MAX];
1164
1165 /* Create instance with 4 ports */
1166 if (init(t, 1, 4) < 0 ||
1167 create_ports(t, 4) < 0 ||
1168 create_atomic_qids(t, 1) < 0) {
1169 printf("%d: Error initializing device\n", __LINE__);
1170 goto fail;
1171 }
1172
1173 err = rte_event_port_link(evdev, t->port[link_port], NULL, NULL, 0);
1174 if (err != 1) {
1175 printf("%d: error mapping lb qid\n", __LINE__);
1176 goto fail;
1177 }
1178
1179 if (rte_event_dev_start(evdev) < 0) {
1180 printf("%d: Error with start call\n", __LINE__);
1181 goto fail;
1182 }
1183
1184 /* no test for device, as it ignores the port/q number */
1185 int num_stats = rte_event_dev_xstats_names_get(evdev,
1186 RTE_EVENT_DEV_XSTATS_PORT,
1187 UINT8_MAX-1, xstats_names, ids,
1188 XSTATS_MAX);
1189 if (num_stats != 0) {
1190 printf("%d: expected %d stats, got return %d\n", __LINE__,
1191 0, num_stats);
1192 goto fail;
1193 }
1194
1195 num_stats = rte_event_dev_xstats_names_get(evdev,
1196 RTE_EVENT_DEV_XSTATS_QUEUE,
1197 UINT8_MAX-1, xstats_names, ids,
1198 XSTATS_MAX);
1199 if (num_stats != 0) {
1200 printf("%d: expected %d stats, got return %d\n", __LINE__,
1201 0, num_stats);
1202 goto fail;
1203 }
1204
1205 cleanup(t);
1206 return 0;
1207 fail:
1208 cleanup(t);
1209 return -1;
1210 }
1211
1212 static int
port_reconfig_credits(struct test * t)1213 port_reconfig_credits(struct test *t)
1214 {
1215 if (init(t, 1, 1) < 0) {
1216 printf("%d: Error initializing device\n", __LINE__);
1217 return -1;
1218 }
1219
1220 uint32_t i;
1221 const uint32_t NUM_ITERS = 32;
1222 for (i = 0; i < NUM_ITERS; i++) {
1223 const struct rte_event_queue_conf conf = {
1224 .schedule_type = RTE_SCHED_TYPE_ATOMIC,
1225 .priority = RTE_EVENT_DEV_PRIORITY_NORMAL,
1226 .nb_atomic_flows = 1024,
1227 .nb_atomic_order_sequences = 1024,
1228 };
1229 if (rte_event_queue_setup(evdev, 0, &conf) < 0) {
1230 printf("%d: error creating qid\n", __LINE__);
1231 return -1;
1232 }
1233 t->qid[0] = 0;
1234
1235 static const struct rte_event_port_conf port_conf = {
1236 .new_event_threshold = 128,
1237 .dequeue_depth = 32,
1238 .enqueue_depth = 64,
1239 };
1240 if (rte_event_port_setup(evdev, 0, &port_conf) < 0) {
1241 printf("%d Error setting up port\n", __LINE__);
1242 return -1;
1243 }
1244
1245 int links = rte_event_port_link(evdev, 0, NULL, NULL, 0);
1246 if (links != 1) {
1247 printf("%d: error mapping lb qid\n", __LINE__);
1248 goto fail;
1249 }
1250
1251 if (rte_event_dev_start(evdev) < 0) {
1252 printf("%d: Error with start call\n", __LINE__);
1253 goto fail;
1254 }
1255
1256 const uint32_t NPKTS = 1;
1257 uint32_t j;
1258 for (j = 0; j < NPKTS; j++) {
1259 struct rte_event ev;
1260 struct rte_mbuf *arp = rte_gen_arp(0, t->mbuf_pool);
1261 if (!arp) {
1262 printf("%d: gen of pkt failed\n", __LINE__);
1263 goto fail;
1264 }
1265 ev.queue_id = t->qid[0];
1266 ev.op = RTE_EVENT_OP_NEW;
1267 ev.mbuf = arp;
1268 int err = rte_event_enqueue_burst(evdev, 0, &ev, 1);
1269 if (err != 1) {
1270 printf("%d: Failed to enqueue\n", __LINE__);
1271 rte_event_dev_dump(0, stdout);
1272 goto fail;
1273 }
1274 }
1275
1276 rte_service_run_iter_on_app_lcore(t->service_id, 1);
1277
1278 struct rte_event ev[NPKTS];
1279 int deq = rte_event_dequeue_burst(evdev, t->port[0], ev,
1280 NPKTS, 0);
1281 if (deq != 1)
1282 printf("%d error; no packet dequeued\n", __LINE__);
1283
1284 /* let cleanup below stop the device on last iter */
1285 if (i != NUM_ITERS-1)
1286 rte_event_dev_stop(evdev);
1287 }
1288
1289 cleanup(t);
1290 return 0;
1291 fail:
1292 cleanup(t);
1293 return -1;
1294 }
1295
1296 static int
port_single_lb_reconfig(struct test * t)1297 port_single_lb_reconfig(struct test *t)
1298 {
1299 if (init(t, 2, 2) < 0) {
1300 printf("%d: Error initializing device\n", __LINE__);
1301 goto fail;
1302 }
1303
1304 static const struct rte_event_queue_conf conf_lb_atomic = {
1305 .priority = RTE_EVENT_DEV_PRIORITY_NORMAL,
1306 .schedule_type = RTE_SCHED_TYPE_ATOMIC,
1307 .nb_atomic_flows = 1024,
1308 .nb_atomic_order_sequences = 1024,
1309 };
1310 if (rte_event_queue_setup(evdev, 0, &conf_lb_atomic) < 0) {
1311 printf("%d: error creating qid\n", __LINE__);
1312 goto fail;
1313 }
1314
1315 static const struct rte_event_queue_conf conf_single_link = {
1316 .priority = RTE_EVENT_DEV_PRIORITY_NORMAL,
1317 .event_queue_cfg = RTE_EVENT_QUEUE_CFG_SINGLE_LINK,
1318 };
1319 if (rte_event_queue_setup(evdev, 1, &conf_single_link) < 0) {
1320 printf("%d: error creating qid\n", __LINE__);
1321 goto fail;
1322 }
1323
1324 struct rte_event_port_conf port_conf = {
1325 .new_event_threshold = 128,
1326 .dequeue_depth = 32,
1327 .enqueue_depth = 64,
1328 };
1329 if (rte_event_port_setup(evdev, 0, &port_conf) < 0) {
1330 printf("%d Error setting up port\n", __LINE__);
1331 goto fail;
1332 }
1333 if (rte_event_port_setup(evdev, 1, &port_conf) < 0) {
1334 printf("%d Error setting up port\n", __LINE__);
1335 goto fail;
1336 }
1337
1338 /* link port to lb queue */
1339 uint8_t queue_id = 0;
1340 if (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) != 1) {
1341 printf("%d: error creating link for qid\n", __LINE__);
1342 goto fail;
1343 }
1344
1345 int ret = rte_event_port_unlink(evdev, 0, &queue_id, 1);
1346 if (ret != 1) {
1347 printf("%d: Error unlinking lb port\n", __LINE__);
1348 goto fail;
1349 }
1350
1351 queue_id = 1;
1352 if (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) != 1) {
1353 printf("%d: error creating link for qid\n", __LINE__);
1354 goto fail;
1355 }
1356
1357 queue_id = 0;
1358 int err = rte_event_port_link(evdev, 1, &queue_id, NULL, 1);
1359 if (err != 1) {
1360 printf("%d: error mapping lb qid\n", __LINE__);
1361 goto fail;
1362 }
1363
1364 if (rte_event_dev_start(evdev) < 0) {
1365 printf("%d: Error with start call\n", __LINE__);
1366 goto fail;
1367 }
1368
1369 cleanup(t);
1370 return 0;
1371 fail:
1372 cleanup(t);
1373 return -1;
1374 }
1375
1376 static int
xstats_brute_force(struct test * t)1377 xstats_brute_force(struct test *t)
1378 {
1379 uint32_t i;
1380 const uint32_t XSTATS_MAX = 1024;
1381 uint32_t ids[XSTATS_MAX];
1382 uint64_t values[XSTATS_MAX];
1383 struct rte_event_dev_xstats_name xstats_names[XSTATS_MAX];
1384
1385
1386 /* Create instance with 4 ports */
1387 if (init(t, 1, 4) < 0 ||
1388 create_ports(t, 4) < 0 ||
1389 create_atomic_qids(t, 1) < 0) {
1390 printf("%d: Error initializing device\n", __LINE__);
1391 return -1;
1392 }
1393
1394 int err = rte_event_port_link(evdev, t->port[0], NULL, NULL, 0);
1395 if (err != 1) {
1396 printf("%d: error mapping lb qid\n", __LINE__);
1397 goto fail;
1398 }
1399
1400 if (rte_event_dev_start(evdev) < 0) {
1401 printf("%d: Error with start call\n", __LINE__);
1402 goto fail;
1403 }
1404
1405 for (i = 0; i < XSTATS_MAX; i++)
1406 ids[i] = i;
1407
1408 for (i = 0; i < 3; i++) {
1409 uint32_t mode = RTE_EVENT_DEV_XSTATS_DEVICE + i;
1410 uint32_t j;
1411 for (j = 0; j < UINT8_MAX; j++) {
1412 rte_event_dev_xstats_names_get(evdev, mode,
1413 j, xstats_names, ids, XSTATS_MAX);
1414
1415 rte_event_dev_xstats_get(evdev, mode, j, ids,
1416 values, XSTATS_MAX);
1417 }
1418 }
1419
1420 cleanup(t);
1421 return 0;
1422 fail:
1423 cleanup(t);
1424 return -1;
1425 }
1426
1427 static int
xstats_id_reset_tests(struct test * t)1428 xstats_id_reset_tests(struct test *t)
1429 {
1430 const int wrk_enq = 2;
1431 int err;
1432
1433 /* Create instance with 4 ports */
1434 if (init(t, 1, 4) < 0 ||
1435 create_ports(t, 4) < 0 ||
1436 create_atomic_qids(t, 1) < 0) {
1437 printf("%d: Error initializing device\n", __LINE__);
1438 return -1;
1439 }
1440
1441 /* CQ mapping to QID */
1442 err = rte_event_port_link(evdev, t->port[wrk_enq], NULL, NULL, 0);
1443 if (err != 1) {
1444 printf("%d: error mapping lb qid\n", __LINE__);
1445 goto fail;
1446 }
1447
1448 if (rte_event_dev_start(evdev) < 0) {
1449 printf("%d: Error with start call\n", __LINE__);
1450 goto fail;
1451 }
1452
1453 #define XSTATS_MAX 1024
1454 int ret;
1455 uint32_t i;
1456 uint32_t ids[XSTATS_MAX];
1457 uint64_t values[XSTATS_MAX];
1458 struct rte_event_dev_xstats_name xstats_names[XSTATS_MAX];
1459
1460 for (i = 0; i < XSTATS_MAX; i++)
1461 ids[i] = i;
1462
1463 #define NUM_DEV_STATS 8
1464 /* Device names / values */
1465 int num_stats = rte_event_dev_xstats_names_get(evdev,
1466 RTE_EVENT_DEV_XSTATS_DEVICE,
1467 0, xstats_names, ids, XSTATS_MAX);
1468 if (num_stats != NUM_DEV_STATS) {
1469 printf("%d: expected %d stats, got return %d\n", __LINE__,
1470 NUM_DEV_STATS, num_stats);
1471 goto fail;
1472 }
1473 ret = rte_event_dev_xstats_get(evdev,
1474 RTE_EVENT_DEV_XSTATS_DEVICE,
1475 0, ids, values, num_stats);
1476 if (ret != NUM_DEV_STATS) {
1477 printf("%d: expected %d stats, got return %d\n", __LINE__,
1478 NUM_DEV_STATS, ret);
1479 goto fail;
1480 }
1481
1482 #define NPKTS 7
1483 for (i = 0; i < NPKTS; i++) {
1484 struct rte_event ev;
1485 struct rte_mbuf *arp = rte_gen_arp(0, t->mbuf_pool);
1486 if (!arp) {
1487 printf("%d: gen of pkt failed\n", __LINE__);
1488 goto fail;
1489 }
1490 ev.queue_id = t->qid[i];
1491 ev.op = RTE_EVENT_OP_NEW;
1492 ev.mbuf = arp;
1493 *rte_event_pmd_selftest_seqn(arp) = i;
1494
1495 int err = rte_event_enqueue_burst(evdev, t->port[0], &ev, 1);
1496 if (err != 1) {
1497 printf("%d: Failed to enqueue\n", __LINE__);
1498 goto fail;
1499 }
1500 }
1501
1502 rte_service_run_iter_on_app_lcore(t->service_id, 1);
1503
1504 static const char * const dev_names[] = {
1505 "dev_rx", "dev_tx", "dev_drop", "dev_sched_calls",
1506 "dev_sched_no_iq_enq", "dev_sched_no_cq_enq",
1507 "dev_sched_last_iter_bitmask",
1508 "dev_sched_progress_last_iter"
1509 };
1510 uint64_t dev_expected[] = {NPKTS, NPKTS, 0, 1, 0, 0, 4, 1};
1511 for (i = 0; (int)i < ret; i++) {
1512 unsigned int id;
1513 uint64_t val = rte_event_dev_xstats_by_name_get(evdev,
1514 dev_names[i],
1515 &id);
1516 if (id != i) {
1517 printf("%d: %s id incorrect, expected %d got %d\n",
1518 __LINE__, dev_names[i], i, id);
1519 goto fail;
1520 }
1521 if (val != dev_expected[i]) {
1522 printf("%d: %s value incorrect, expected %"
1523 PRIu64" got %"PRIu64"\n", __LINE__,
1524 dev_names[i], dev_expected[i], val);
1525 goto fail;
1526 }
1527 /* reset to zero */
1528 int reset_ret = rte_event_dev_xstats_reset(evdev,
1529 RTE_EVENT_DEV_XSTATS_DEVICE, 0,
1530 &id,
1531 1);
1532 if (reset_ret) {
1533 printf("%d: failed to reset successfully\n", __LINE__);
1534 goto fail;
1535 }
1536 dev_expected[i] = 0;
1537 /* check value again */
1538 val = rte_event_dev_xstats_by_name_get(evdev, dev_names[i], 0);
1539 if (val != dev_expected[i]) {
1540 printf("%d: %s value incorrect, expected %"PRIu64
1541 " got %"PRIu64"\n", __LINE__, dev_names[i],
1542 dev_expected[i], val);
1543 goto fail;
1544 }
1545 };
1546
1547 /* 49 is stat offset from start of the devices whole xstats.
1548 * This WILL break every time we add a statistic to a port
1549 * or the device, but there is no other way to test
1550 */
1551 #define PORT_OFF 50
1552 /* num stats for the tested port. CQ size adds more stats to a port */
1553 #define NUM_PORT_STATS 21
1554 /* the port to test. */
1555 #define PORT 2
1556 num_stats = rte_event_dev_xstats_names_get(evdev,
1557 RTE_EVENT_DEV_XSTATS_PORT, PORT,
1558 xstats_names, ids, XSTATS_MAX);
1559 if (num_stats != NUM_PORT_STATS) {
1560 printf("%d: expected %d stats, got return %d\n",
1561 __LINE__, NUM_PORT_STATS, num_stats);
1562 goto fail;
1563 }
1564 ret = rte_event_dev_xstats_get(evdev, RTE_EVENT_DEV_XSTATS_PORT, PORT,
1565 ids, values, num_stats);
1566
1567 if (ret != NUM_PORT_STATS) {
1568 printf("%d: expected %d stats, got return %d\n",
1569 __LINE__, NUM_PORT_STATS, ret);
1570 goto fail;
1571 }
1572 static const char * const port_names[] = {
1573 "port_2_rx",
1574 "port_2_tx",
1575 "port_2_drop",
1576 "port_2_inflight",
1577 "port_2_avg_pkt_cycles",
1578 "port_2_credits",
1579 "port_2_rx_ring_used",
1580 "port_2_rx_ring_free",
1581 "port_2_cq_ring_used",
1582 "port_2_cq_ring_free",
1583 "port_2_dequeue_calls",
1584 "port_2_dequeues_returning_0",
1585 "port_2_dequeues_returning_1-4",
1586 "port_2_dequeues_returning_5-8",
1587 "port_2_dequeues_returning_9-12",
1588 "port_2_dequeues_returning_13-16",
1589 "port_2_dequeues_returning_17-20",
1590 "port_2_dequeues_returning_21-24",
1591 "port_2_dequeues_returning_25-28",
1592 "port_2_dequeues_returning_29-32",
1593 "port_2_dequeues_returning_33-36",
1594 };
1595 uint64_t port_expected[] = {
1596 0, /* rx */
1597 NPKTS, /* tx */
1598 0, /* drop */
1599 NPKTS, /* inflight */
1600 0, /* avg pkt cycles */
1601 0, /* credits */
1602 0, /* rx ring used */
1603 4096, /* rx ring free */
1604 NPKTS, /* cq ring used */
1605 25, /* cq ring free */
1606 0, /* dequeue zero calls */
1607 0, 0, 0, 0, 0, /* 10 dequeue buckets */
1608 0, 0, 0, 0, 0,
1609 };
1610 uint64_t port_expected_zero[] = {
1611 0, /* rx */
1612 0, /* tx */
1613 0, /* drop */
1614 NPKTS, /* inflight */
1615 0, /* avg pkt cycles */
1616 0, /* credits */
1617 0, /* rx ring used */
1618 4096, /* rx ring free */
1619 NPKTS, /* cq ring used */
1620 25, /* cq ring free */
1621 0, /* dequeue zero calls */
1622 0, 0, 0, 0, 0, /* 10 dequeue buckets */
1623 0, 0, 0, 0, 0,
1624 };
1625 if (RTE_DIM(port_expected) != NUM_PORT_STATS ||
1626 RTE_DIM(port_names) != NUM_PORT_STATS) {
1627 printf("%d: port array of wrong size\n", __LINE__);
1628 goto fail;
1629 }
1630
1631 int failed = 0;
1632 for (i = 0; (int)i < ret; i++) {
1633 unsigned int id;
1634 uint64_t val = rte_event_dev_xstats_by_name_get(evdev,
1635 port_names[i],
1636 &id);
1637 if (id != i + PORT_OFF) {
1638 printf("%d: %s id incorrect, expected %d got %d\n",
1639 __LINE__, port_names[i], i+PORT_OFF,
1640 id);
1641 failed = 1;
1642 }
1643 if (val != port_expected[i]) {
1644 printf("%d: %s value incorrect, expected %"PRIu64
1645 " got %d\n", __LINE__, port_names[i],
1646 port_expected[i], id);
1647 failed = 1;
1648 }
1649 /* reset to zero */
1650 int reset_ret = rte_event_dev_xstats_reset(evdev,
1651 RTE_EVENT_DEV_XSTATS_PORT, PORT,
1652 &id,
1653 1);
1654 if (reset_ret) {
1655 printf("%d: failed to reset successfully\n", __LINE__);
1656 failed = 1;
1657 }
1658 /* check value again */
1659 val = rte_event_dev_xstats_by_name_get(evdev, port_names[i], 0);
1660 if (val != port_expected_zero[i]) {
1661 printf("%d: %s value incorrect, expected %"PRIu64
1662 " got %"PRIu64"\n", __LINE__, port_names[i],
1663 port_expected_zero[i], val);
1664 failed = 1;
1665 }
1666 };
1667 if (failed)
1668 goto fail;
1669
1670 /* num queue stats */
1671 #define NUM_Q_STATS 16
1672 /* queue offset from start of the devices whole xstats.
1673 * This will break every time we add a statistic to a device/port/queue
1674 */
1675 #define QUEUE_OFF 92
1676 const uint32_t queue = 0;
1677 num_stats = rte_event_dev_xstats_names_get(evdev,
1678 RTE_EVENT_DEV_XSTATS_QUEUE, queue,
1679 xstats_names, ids, XSTATS_MAX);
1680 if (num_stats != NUM_Q_STATS) {
1681 printf("%d: expected %d stats, got return %d\n",
1682 __LINE__, NUM_Q_STATS, num_stats);
1683 goto fail;
1684 }
1685 ret = rte_event_dev_xstats_get(evdev, RTE_EVENT_DEV_XSTATS_QUEUE,
1686 queue, ids, values, num_stats);
1687 if (ret != NUM_Q_STATS) {
1688 printf("%d: expected 21 stats, got return %d\n", __LINE__, ret);
1689 goto fail;
1690 }
1691 static const char * const queue_names[] = {
1692 "qid_0_rx",
1693 "qid_0_tx",
1694 "qid_0_drop",
1695 "qid_0_inflight",
1696 "qid_0_iq_0_used",
1697 "qid_0_iq_1_used",
1698 "qid_0_iq_2_used",
1699 "qid_0_iq_3_used",
1700 "qid_0_port_0_pinned_flows",
1701 "qid_0_port_0_packets",
1702 "qid_0_port_1_pinned_flows",
1703 "qid_0_port_1_packets",
1704 "qid_0_port_2_pinned_flows",
1705 "qid_0_port_2_packets",
1706 "qid_0_port_3_pinned_flows",
1707 "qid_0_port_3_packets",
1708 };
1709 uint64_t queue_expected[] = {
1710 7, /* rx */
1711 7, /* tx */
1712 0, /* drop */
1713 7, /* inflight */
1714 0, /* iq 0 used */
1715 0, /* iq 1 used */
1716 0, /* iq 2 used */
1717 0, /* iq 3 used */
1718 /* QID-to-Port: pinned_flows, packets */
1719 0, 0,
1720 0, 0,
1721 1, 7,
1722 0, 0,
1723 };
1724 uint64_t queue_expected_zero[] = {
1725 0, /* rx */
1726 0, /* tx */
1727 0, /* drop */
1728 7, /* inflight */
1729 0, /* iq 0 used */
1730 0, /* iq 1 used */
1731 0, /* iq 2 used */
1732 0, /* iq 3 used */
1733 /* QID-to-Port: pinned_flows, packets */
1734 0, 0,
1735 0, 0,
1736 1, 0,
1737 0, 0,
1738 };
1739 if (RTE_DIM(queue_expected) != NUM_Q_STATS ||
1740 RTE_DIM(queue_expected_zero) != NUM_Q_STATS ||
1741 RTE_DIM(queue_names) != NUM_Q_STATS) {
1742 printf("%d : queue array of wrong size\n", __LINE__);
1743 goto fail;
1744 }
1745
1746 failed = 0;
1747 for (i = 0; (int)i < ret; i++) {
1748 unsigned int id;
1749 uint64_t val = rte_event_dev_xstats_by_name_get(evdev,
1750 queue_names[i],
1751 &id);
1752 if (id != i + QUEUE_OFF) {
1753 printf("%d: %s id incorrect, expected %d got %d\n",
1754 __LINE__, queue_names[i], i+QUEUE_OFF,
1755 id);
1756 failed = 1;
1757 }
1758 if (val != queue_expected[i]) {
1759 printf("%d: %d: %s value , expected %"PRIu64
1760 " got %"PRIu64"\n", i, __LINE__,
1761 queue_names[i], queue_expected[i], val);
1762 failed = 1;
1763 }
1764 /* reset to zero */
1765 int reset_ret = rte_event_dev_xstats_reset(evdev,
1766 RTE_EVENT_DEV_XSTATS_QUEUE,
1767 queue, &id, 1);
1768 if (reset_ret) {
1769 printf("%d: failed to reset successfully\n", __LINE__);
1770 failed = 1;
1771 }
1772 /* check value again */
1773 val = rte_event_dev_xstats_by_name_get(evdev, queue_names[i],
1774 0);
1775 if (val != queue_expected_zero[i]) {
1776 printf("%d: %s value incorrect, expected %"PRIu64
1777 " got %"PRIu64"\n", __LINE__, queue_names[i],
1778 queue_expected_zero[i], val);
1779 failed = 1;
1780 }
1781 };
1782
1783 if (failed)
1784 goto fail;
1785
1786 cleanup(t);
1787 return 0;
1788 fail:
1789 cleanup(t);
1790 return -1;
1791 }
1792
1793 static int
ordered_reconfigure(struct test * t)1794 ordered_reconfigure(struct test *t)
1795 {
1796 if (init(t, 1, 1) < 0 ||
1797 create_ports(t, 1) < 0) {
1798 printf("%d: Error initializing device\n", __LINE__);
1799 return -1;
1800 }
1801
1802 const struct rte_event_queue_conf conf = {
1803 .schedule_type = RTE_SCHED_TYPE_ORDERED,
1804 .priority = RTE_EVENT_DEV_PRIORITY_NORMAL,
1805 .nb_atomic_flows = 1024,
1806 .nb_atomic_order_sequences = 1024,
1807 };
1808
1809 if (rte_event_queue_setup(evdev, 0, &conf) < 0) {
1810 printf("%d: error creating qid\n", __LINE__);
1811 goto failed;
1812 }
1813
1814 if (rte_event_queue_setup(evdev, 0, &conf) < 0) {
1815 printf("%d: error creating qid, for 2nd time\n", __LINE__);
1816 goto failed;
1817 }
1818
1819 rte_event_port_link(evdev, t->port[0], NULL, NULL, 0);
1820 if (rte_event_dev_start(evdev) < 0) {
1821 printf("%d: Error with start call\n", __LINE__);
1822 return -1;
1823 }
1824
1825 cleanup(t);
1826 return 0;
1827 failed:
1828 cleanup(t);
1829 return -1;
1830 }
1831
1832 static int
qid_priorities(struct test * t)1833 qid_priorities(struct test *t)
1834 {
1835 /* Test works by having a CQ with enough empty space for all packets,
1836 * and enqueueing 3 packets to 3 QIDs. They must return based on the
1837 * priority of the QID, not the ingress order, to pass the test
1838 */
1839 unsigned int i;
1840 /* Create instance with 1 ports, and 3 qids */
1841 if (init(t, 3, 1) < 0 ||
1842 create_ports(t, 1) < 0) {
1843 printf("%d: Error initializing device\n", __LINE__);
1844 return -1;
1845 }
1846
1847 for (i = 0; i < 3; i++) {
1848 /* Create QID */
1849 const struct rte_event_queue_conf conf = {
1850 .schedule_type = RTE_SCHED_TYPE_ATOMIC,
1851 /* increase priority (0 == highest), as we go */
1852 .priority = RTE_EVENT_DEV_PRIORITY_NORMAL - i,
1853 .nb_atomic_flows = 1024,
1854 .nb_atomic_order_sequences = 1024,
1855 };
1856
1857 if (rte_event_queue_setup(evdev, i, &conf) < 0) {
1858 printf("%d: error creating qid %d\n", __LINE__, i);
1859 return -1;
1860 }
1861 t->qid[i] = i;
1862 }
1863 t->nb_qids = i;
1864 /* map all QIDs to port */
1865 rte_event_port_link(evdev, t->port[0], NULL, NULL, 0);
1866
1867 if (rte_event_dev_start(evdev) < 0) {
1868 printf("%d: Error with start call\n", __LINE__);
1869 return -1;
1870 }
1871
1872 /* enqueue 3 packets, setting seqn and QID to check priority */
1873 for (i = 0; i < 3; i++) {
1874 struct rte_event ev;
1875 struct rte_mbuf *arp = rte_gen_arp(0, t->mbuf_pool);
1876 if (!arp) {
1877 printf("%d: gen of pkt failed\n", __LINE__);
1878 return -1;
1879 }
1880 ev.queue_id = t->qid[i];
1881 ev.op = RTE_EVENT_OP_NEW;
1882 ev.mbuf = arp;
1883 *rte_event_pmd_selftest_seqn(arp) = i;
1884
1885 int err = rte_event_enqueue_burst(evdev, t->port[0], &ev, 1);
1886 if (err != 1) {
1887 printf("%d: Failed to enqueue\n", __LINE__);
1888 return -1;
1889 }
1890 }
1891
1892 rte_service_run_iter_on_app_lcore(t->service_id, 1);
1893
1894 /* dequeue packets, verify priority was upheld */
1895 struct rte_event ev[32];
1896 uint32_t deq_pkts =
1897 rte_event_dequeue_burst(evdev, t->port[0], ev, 32, 0);
1898 if (deq_pkts != 3) {
1899 printf("%d: failed to deq packets\n", __LINE__);
1900 rte_event_dev_dump(evdev, stdout);
1901 return -1;
1902 }
1903 for (i = 0; i < 3; i++) {
1904 if (*rte_event_pmd_selftest_seqn(ev[i].mbuf) != 2-i) {
1905 printf(
1906 "%d: qid priority test: seqn %d incorrectly prioritized\n",
1907 __LINE__, i);
1908 }
1909 }
1910
1911 cleanup(t);
1912 return 0;
1913 }
1914
1915 static int
unlink_in_progress(struct test * t)1916 unlink_in_progress(struct test *t)
1917 {
1918 /* Test unlinking API, in particular that when an unlink request has
1919 * not yet been seen by the scheduler thread, that the
1920 * unlink_in_progress() function returns the number of unlinks.
1921 */
1922 unsigned int i;
1923 /* Create instance with 1 ports, and 3 qids */
1924 if (init(t, 3, 1) < 0 ||
1925 create_ports(t, 1) < 0) {
1926 printf("%d: Error initializing device\n", __LINE__);
1927 return -1;
1928 }
1929
1930 for (i = 0; i < 3; i++) {
1931 /* Create QID */
1932 const struct rte_event_queue_conf conf = {
1933 .schedule_type = RTE_SCHED_TYPE_ATOMIC,
1934 /* increase priority (0 == highest), as we go */
1935 .priority = RTE_EVENT_DEV_PRIORITY_NORMAL - i,
1936 .nb_atomic_flows = 1024,
1937 .nb_atomic_order_sequences = 1024,
1938 };
1939
1940 if (rte_event_queue_setup(evdev, i, &conf) < 0) {
1941 printf("%d: error creating qid %d\n", __LINE__, i);
1942 return -1;
1943 }
1944 t->qid[i] = i;
1945 }
1946 t->nb_qids = i;
1947 /* map all QIDs to port */
1948 rte_event_port_link(evdev, t->port[0], NULL, NULL, 0);
1949
1950 if (rte_event_dev_start(evdev) < 0) {
1951 printf("%d: Error with start call\n", __LINE__);
1952 return -1;
1953 }
1954
1955 /* unlink all ports to have outstanding unlink requests */
1956 int ret = rte_event_port_unlink(evdev, t->port[0], NULL, 0);
1957 if (ret < 0) {
1958 printf("%d: Failed to unlink queues\n", __LINE__);
1959 return -1;
1960 }
1961
1962 /* get active unlinks here, expect 3 */
1963 int unlinks_in_progress =
1964 rte_event_port_unlinks_in_progress(evdev, t->port[0]);
1965 if (unlinks_in_progress != 3) {
1966 printf("%d: Expected num unlinks in progress == 3, got %d\n",
1967 __LINE__, unlinks_in_progress);
1968 return -1;
1969 }
1970
1971 /* run scheduler service on this thread to ack the unlinks */
1972 rte_service_run_iter_on_app_lcore(t->service_id, 1);
1973
1974 /* active unlinks expected as 0 as scheduler thread has acked */
1975 unlinks_in_progress =
1976 rte_event_port_unlinks_in_progress(evdev, t->port[0]);
1977 if (unlinks_in_progress != 0) {
1978 printf("%d: Expected num unlinks in progress == 0, got %d\n",
1979 __LINE__, unlinks_in_progress);
1980 }
1981
1982 cleanup(t);
1983 return 0;
1984 }
1985
1986 static int
load_balancing(struct test * t)1987 load_balancing(struct test *t)
1988 {
1989 const int rx_enq = 0;
1990 int err;
1991 uint32_t i;
1992
1993 if (init(t, 1, 4) < 0 ||
1994 create_ports(t, 4) < 0 ||
1995 create_atomic_qids(t, 1) < 0) {
1996 printf("%d: Error initializing device\n", __LINE__);
1997 return -1;
1998 }
1999
2000 for (i = 0; i < 3; i++) {
2001 /* map port 1 - 3 inclusive */
2002 if (rte_event_port_link(evdev, t->port[i+1], &t->qid[0],
2003 NULL, 1) != 1) {
2004 printf("%d: error mapping qid to port %d\n",
2005 __LINE__, i);
2006 return -1;
2007 }
2008 }
2009
2010 if (rte_event_dev_start(evdev) < 0) {
2011 printf("%d: Error with start call\n", __LINE__);
2012 return -1;
2013 }
2014
2015 /************** FORWARD ****************/
2016 /*
2017 * Create a set of flows that test the load-balancing operation of the
2018 * implementation. Fill CQ 0 and 1 with flows 0 and 1, and test
2019 * with a new flow, which should be sent to the 3rd mapped CQ
2020 */
2021 static uint32_t flows[] = {0, 1, 1, 0, 0, 2, 2, 0, 2};
2022
2023 for (i = 0; i < RTE_DIM(flows); i++) {
2024 struct rte_mbuf *arp = rte_gen_arp(0, t->mbuf_pool);
2025 if (!arp) {
2026 printf("%d: gen of pkt failed\n", __LINE__);
2027 return -1;
2028 }
2029
2030 struct rte_event ev = {
2031 .op = RTE_EVENT_OP_NEW,
2032 .queue_id = t->qid[0],
2033 .flow_id = flows[i],
2034 .mbuf = arp,
2035 };
2036 /* generate pkt and enqueue */
2037 err = rte_event_enqueue_burst(evdev, t->port[rx_enq], &ev, 1);
2038 if (err != 1) {
2039 printf("%d: Failed to enqueue\n", __LINE__);
2040 return -1;
2041 }
2042 }
2043
2044 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2045
2046 struct test_event_dev_stats stats;
2047 err = test_event_dev_stats_get(evdev, &stats);
2048 if (err) {
2049 printf("%d: failed to get stats\n", __LINE__);
2050 return -1;
2051 }
2052
2053 if (stats.port_inflight[1] != 4) {
2054 printf("%d:%s: port 1 inflight not correct\n", __LINE__,
2055 __func__);
2056 return -1;
2057 }
2058 if (stats.port_inflight[2] != 2) {
2059 printf("%d:%s: port 2 inflight not correct\n", __LINE__,
2060 __func__);
2061 return -1;
2062 }
2063 if (stats.port_inflight[3] != 3) {
2064 printf("%d:%s: port 3 inflight not correct\n", __LINE__,
2065 __func__);
2066 return -1;
2067 }
2068
2069 cleanup(t);
2070 return 0;
2071 }
2072
2073 static int
load_balancing_history(struct test * t)2074 load_balancing_history(struct test *t)
2075 {
2076 struct test_event_dev_stats stats = {0};
2077 const int rx_enq = 0;
2078 int err;
2079 uint32_t i;
2080
2081 /* Create instance with 1 atomic QID going to 3 ports + 1 prod port */
2082 if (init(t, 1, 4) < 0 ||
2083 create_ports(t, 4) < 0 ||
2084 create_atomic_qids(t, 1) < 0)
2085 return -1;
2086
2087 /* CQ mapping to QID */
2088 if (rte_event_port_link(evdev, t->port[1], &t->qid[0], NULL, 1) != 1) {
2089 printf("%d: error mapping port 1 qid\n", __LINE__);
2090 return -1;
2091 }
2092 if (rte_event_port_link(evdev, t->port[2], &t->qid[0], NULL, 1) != 1) {
2093 printf("%d: error mapping port 2 qid\n", __LINE__);
2094 return -1;
2095 }
2096 if (rte_event_port_link(evdev, t->port[3], &t->qid[0], NULL, 1) != 1) {
2097 printf("%d: error mapping port 3 qid\n", __LINE__);
2098 return -1;
2099 }
2100 if (rte_event_dev_start(evdev) < 0) {
2101 printf("%d: Error with start call\n", __LINE__);
2102 return -1;
2103 }
2104
2105 /*
2106 * Create a set of flows that test the load-balancing operation of the
2107 * implementation. Fill CQ 0, 1 and 2 with flows 0, 1 and 2, drop
2108 * the packet from CQ 0, send in a new set of flows. Ensure that:
2109 * 1. The new flow 3 gets into the empty CQ0
2110 * 2. packets for existing flow gets added into CQ1
2111 * 3. Next flow 0 pkt is now onto CQ2, since CQ0 and CQ1 now contain
2112 * more outstanding pkts
2113 *
2114 * This test makes sure that when a flow ends (i.e. all packets
2115 * have been completed for that flow), that the flow can be moved
2116 * to a different CQ when new packets come in for that flow.
2117 */
2118 static uint32_t flows1[] = {0, 1, 1, 2};
2119
2120 for (i = 0; i < RTE_DIM(flows1); i++) {
2121 struct rte_mbuf *arp = rte_gen_arp(0, t->mbuf_pool);
2122 struct rte_event ev = {
2123 .flow_id = flows1[i],
2124 .op = RTE_EVENT_OP_NEW,
2125 .queue_id = t->qid[0],
2126 .event_type = RTE_EVENT_TYPE_CPU,
2127 .priority = RTE_EVENT_DEV_PRIORITY_NORMAL,
2128 .mbuf = arp
2129 };
2130
2131 if (!arp) {
2132 printf("%d: gen of pkt failed\n", __LINE__);
2133 return -1;
2134 }
2135 arp->hash.rss = flows1[i];
2136 err = rte_event_enqueue_burst(evdev, t->port[rx_enq], &ev, 1);
2137 if (err != 1) {
2138 printf("%d: Failed to enqueue\n", __LINE__);
2139 return -1;
2140 }
2141 }
2142
2143 /* call the scheduler */
2144 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2145
2146 /* Dequeue the flow 0 packet from port 1, so that we can then drop */
2147 struct rte_event ev;
2148 if (!rte_event_dequeue_burst(evdev, t->port[1], &ev, 1, 0)) {
2149 printf("%d: failed to dequeue\n", __LINE__);
2150 return -1;
2151 }
2152 if (ev.mbuf->hash.rss != flows1[0]) {
2153 printf("%d: unexpected flow received\n", __LINE__);
2154 return -1;
2155 }
2156
2157 /* drop the flow 0 packet from port 1 */
2158 rte_event_enqueue_burst(evdev, t->port[1], &release_ev, 1);
2159
2160 /* call the scheduler */
2161 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2162
2163 /*
2164 * Set up the next set of flows, first a new flow to fill up
2165 * CQ 0, so that the next flow 0 packet should go to CQ2
2166 */
2167 static uint32_t flows2[] = { 3, 3, 3, 1, 1, 0 };
2168
2169 for (i = 0; i < RTE_DIM(flows2); i++) {
2170 struct rte_mbuf *arp = rte_gen_arp(0, t->mbuf_pool);
2171 struct rte_event ev = {
2172 .flow_id = flows2[i],
2173 .op = RTE_EVENT_OP_NEW,
2174 .queue_id = t->qid[0],
2175 .event_type = RTE_EVENT_TYPE_CPU,
2176 .priority = RTE_EVENT_DEV_PRIORITY_NORMAL,
2177 .mbuf = arp
2178 };
2179
2180 if (!arp) {
2181 printf("%d: gen of pkt failed\n", __LINE__);
2182 return -1;
2183 }
2184 arp->hash.rss = flows2[i];
2185
2186 err = rte_event_enqueue_burst(evdev, t->port[rx_enq], &ev, 1);
2187 if (err != 1) {
2188 printf("%d: Failed to enqueue\n", __LINE__);
2189 return -1;
2190 }
2191 }
2192
2193 /* schedule */
2194 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2195
2196 err = test_event_dev_stats_get(evdev, &stats);
2197 if (err) {
2198 printf("%d:failed to get stats\n", __LINE__);
2199 return -1;
2200 }
2201
2202 /*
2203 * Now check the resulting inflights on each port.
2204 */
2205 if (stats.port_inflight[1] != 3) {
2206 printf("%d:%s: port 1 inflight not correct\n", __LINE__,
2207 __func__);
2208 printf("Inflights, ports 1, 2, 3: %u, %u, %u\n",
2209 (unsigned int)stats.port_inflight[1],
2210 (unsigned int)stats.port_inflight[2],
2211 (unsigned int)stats.port_inflight[3]);
2212 return -1;
2213 }
2214 if (stats.port_inflight[2] != 4) {
2215 printf("%d:%s: port 2 inflight not correct\n", __LINE__,
2216 __func__);
2217 printf("Inflights, ports 1, 2, 3: %u, %u, %u\n",
2218 (unsigned int)stats.port_inflight[1],
2219 (unsigned int)stats.port_inflight[2],
2220 (unsigned int)stats.port_inflight[3]);
2221 return -1;
2222 }
2223 if (stats.port_inflight[3] != 2) {
2224 printf("%d:%s: port 3 inflight not correct\n", __LINE__,
2225 __func__);
2226 printf("Inflights, ports 1, 2, 3: %u, %u, %u\n",
2227 (unsigned int)stats.port_inflight[1],
2228 (unsigned int)stats.port_inflight[2],
2229 (unsigned int)stats.port_inflight[3]);
2230 return -1;
2231 }
2232
2233 for (i = 1; i <= 3; i++) {
2234 struct rte_event ev;
2235 while (rte_event_dequeue_burst(evdev, i, &ev, 1, 0))
2236 rte_event_enqueue_burst(evdev, i, &release_ev, 1);
2237 }
2238 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2239
2240 cleanup(t);
2241 return 0;
2242 }
2243
2244 static int
invalid_qid(struct test * t)2245 invalid_qid(struct test *t)
2246 {
2247 struct test_event_dev_stats stats;
2248 const int rx_enq = 0;
2249 int err;
2250 uint32_t i;
2251
2252 if (init(t, 1, 4) < 0 ||
2253 create_ports(t, 4) < 0 ||
2254 create_atomic_qids(t, 1) < 0) {
2255 printf("%d: Error initializing device\n", __LINE__);
2256 return -1;
2257 }
2258
2259 /* CQ mapping to QID */
2260 for (i = 0; i < 4; i++) {
2261 err = rte_event_port_link(evdev, t->port[i], &t->qid[0],
2262 NULL, 1);
2263 if (err != 1) {
2264 printf("%d: error mapping port 1 qid\n", __LINE__);
2265 return -1;
2266 }
2267 }
2268
2269 if (rte_event_dev_start(evdev) < 0) {
2270 printf("%d: Error with start call\n", __LINE__);
2271 return -1;
2272 }
2273
2274 /*
2275 * Send in a packet with an invalid qid to the scheduler.
2276 * We should see the packed enqueued OK, but the inflights for
2277 * that packet should not be incremented, and the rx_dropped
2278 * should be incremented.
2279 */
2280 static uint32_t flows1[] = {20};
2281
2282 for (i = 0; i < RTE_DIM(flows1); i++) {
2283 struct rte_mbuf *arp = rte_gen_arp(0, t->mbuf_pool);
2284 if (!arp) {
2285 printf("%d: gen of pkt failed\n", __LINE__);
2286 return -1;
2287 }
2288
2289 struct rte_event ev = {
2290 .op = RTE_EVENT_OP_NEW,
2291 .queue_id = t->qid[0] + flows1[i],
2292 .flow_id = i,
2293 .mbuf = arp,
2294 };
2295 /* generate pkt and enqueue */
2296 err = rte_event_enqueue_burst(evdev, t->port[rx_enq], &ev, 1);
2297 if (err != 1) {
2298 printf("%d: Failed to enqueue\n", __LINE__);
2299 return -1;
2300 }
2301 }
2302
2303 /* call the scheduler */
2304 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2305
2306 err = test_event_dev_stats_get(evdev, &stats);
2307 if (err) {
2308 printf("%d: failed to get stats\n", __LINE__);
2309 return -1;
2310 }
2311
2312 /*
2313 * Now check the resulting inflights on the port, and the rx_dropped.
2314 */
2315 if (stats.port_inflight[0] != 0) {
2316 printf("%d:%s: port 1 inflight count not correct\n", __LINE__,
2317 __func__);
2318 rte_event_dev_dump(evdev, stdout);
2319 return -1;
2320 }
2321 if (stats.port_rx_dropped[0] != 1) {
2322 printf("%d:%s: port 1 drops\n", __LINE__, __func__);
2323 rte_event_dev_dump(evdev, stdout);
2324 return -1;
2325 }
2326 /* each packet drop should only be counted in one place - port or dev */
2327 if (stats.rx_dropped != 0) {
2328 printf("%d:%s: port 1 dropped count not correct\n", __LINE__,
2329 __func__);
2330 rte_event_dev_dump(evdev, stdout);
2331 return -1;
2332 }
2333
2334 cleanup(t);
2335 return 0;
2336 }
2337
2338 static int
single_packet(struct test * t)2339 single_packet(struct test *t)
2340 {
2341 const uint32_t MAGIC_SEQN = 7321;
2342 struct rte_event ev;
2343 struct test_event_dev_stats stats;
2344 const int rx_enq = 0;
2345 const int wrk_enq = 2;
2346 int err;
2347
2348 /* Create instance with 4 ports */
2349 if (init(t, 1, 4) < 0 ||
2350 create_ports(t, 4) < 0 ||
2351 create_atomic_qids(t, 1) < 0) {
2352 printf("%d: Error initializing device\n", __LINE__);
2353 return -1;
2354 }
2355
2356 /* CQ mapping to QID */
2357 err = rte_event_port_link(evdev, t->port[wrk_enq], NULL, NULL, 0);
2358 if (err != 1) {
2359 printf("%d: error mapping lb qid\n", __LINE__);
2360 cleanup(t);
2361 return -1;
2362 }
2363
2364 if (rte_event_dev_start(evdev) < 0) {
2365 printf("%d: Error with start call\n", __LINE__);
2366 return -1;
2367 }
2368
2369 /************** Gen pkt and enqueue ****************/
2370 struct rte_mbuf *arp = rte_gen_arp(0, t->mbuf_pool);
2371 if (!arp) {
2372 printf("%d: gen of pkt failed\n", __LINE__);
2373 return -1;
2374 }
2375
2376 ev.op = RTE_EVENT_OP_NEW;
2377 ev.priority = RTE_EVENT_DEV_PRIORITY_NORMAL;
2378 ev.mbuf = arp;
2379 ev.queue_id = 0;
2380 ev.flow_id = 3;
2381 *rte_event_pmd_selftest_seqn(arp) = MAGIC_SEQN;
2382
2383 err = rte_event_enqueue_burst(evdev, t->port[rx_enq], &ev, 1);
2384 if (err != 1) {
2385 printf("%d: Failed to enqueue\n", __LINE__);
2386 return -1;
2387 }
2388
2389 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2390
2391 err = test_event_dev_stats_get(evdev, &stats);
2392 if (err) {
2393 printf("%d: failed to get stats\n", __LINE__);
2394 return -1;
2395 }
2396
2397 if (stats.rx_pkts != 1 ||
2398 stats.tx_pkts != 1 ||
2399 stats.port_inflight[wrk_enq] != 1) {
2400 printf("%d: Sched core didn't handle pkt as expected\n",
2401 __LINE__);
2402 rte_event_dev_dump(evdev, stdout);
2403 return -1;
2404 }
2405
2406 uint32_t deq_pkts;
2407
2408 deq_pkts = rte_event_dequeue_burst(evdev, t->port[wrk_enq], &ev, 1, 0);
2409 if (deq_pkts < 1) {
2410 printf("%d: Failed to deq\n", __LINE__);
2411 return -1;
2412 }
2413
2414 err = test_event_dev_stats_get(evdev, &stats);
2415 if (err) {
2416 printf("%d: failed to get stats\n", __LINE__);
2417 return -1;
2418 }
2419
2420 err = test_event_dev_stats_get(evdev, &stats);
2421 if (*rte_event_pmd_selftest_seqn(ev.mbuf) != MAGIC_SEQN) {
2422 printf("%d: magic sequence number not dequeued\n", __LINE__);
2423 return -1;
2424 }
2425
2426 rte_pktmbuf_free(ev.mbuf);
2427 err = rte_event_enqueue_burst(evdev, t->port[wrk_enq], &release_ev, 1);
2428 if (err != 1) {
2429 printf("%d: Failed to enqueue\n", __LINE__);
2430 return -1;
2431 }
2432 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2433
2434 err = test_event_dev_stats_get(evdev, &stats);
2435 if (stats.port_inflight[wrk_enq] != 0) {
2436 printf("%d: port inflight not correct\n", __LINE__);
2437 return -1;
2438 }
2439
2440 cleanup(t);
2441 return 0;
2442 }
2443
2444 static int
inflight_counts(struct test * t)2445 inflight_counts(struct test *t)
2446 {
2447 struct rte_event ev;
2448 struct test_event_dev_stats stats;
2449 const int rx_enq = 0;
2450 const int p1 = 1;
2451 const int p2 = 2;
2452 int err;
2453 int i;
2454
2455 /* Create instance with 4 ports */
2456 if (init(t, 2, 3) < 0 ||
2457 create_ports(t, 3) < 0 ||
2458 create_atomic_qids(t, 2) < 0) {
2459 printf("%d: Error initializing device\n", __LINE__);
2460 return -1;
2461 }
2462
2463 /* CQ mapping to QID */
2464 err = rte_event_port_link(evdev, t->port[p1], &t->qid[0], NULL, 1);
2465 if (err != 1) {
2466 printf("%d: error mapping lb qid\n", __LINE__);
2467 cleanup(t);
2468 return -1;
2469 }
2470 err = rte_event_port_link(evdev, t->port[p2], &t->qid[1], NULL, 1);
2471 if (err != 1) {
2472 printf("%d: error mapping lb qid\n", __LINE__);
2473 cleanup(t);
2474 return -1;
2475 }
2476
2477 if (rte_event_dev_start(evdev) < 0) {
2478 printf("%d: Error with start call\n", __LINE__);
2479 return -1;
2480 }
2481
2482 /************** FORWARD ****************/
2483 #define QID1_NUM 5
2484 for (i = 0; i < QID1_NUM; i++) {
2485 struct rte_mbuf *arp = rte_gen_arp(0, t->mbuf_pool);
2486
2487 if (!arp) {
2488 printf("%d: gen of pkt failed\n", __LINE__);
2489 goto err;
2490 }
2491
2492 ev.queue_id = t->qid[0];
2493 ev.op = RTE_EVENT_OP_NEW;
2494 ev.mbuf = arp;
2495 err = rte_event_enqueue_burst(evdev, t->port[rx_enq], &ev, 1);
2496 if (err != 1) {
2497 printf("%d: Failed to enqueue\n", __LINE__);
2498 goto err;
2499 }
2500 }
2501 #define QID2_NUM 3
2502 for (i = 0; i < QID2_NUM; i++) {
2503 struct rte_mbuf *arp = rte_gen_arp(0, t->mbuf_pool);
2504
2505 if (!arp) {
2506 printf("%d: gen of pkt failed\n", __LINE__);
2507 goto err;
2508 }
2509 ev.queue_id = t->qid[1];
2510 ev.op = RTE_EVENT_OP_NEW;
2511 ev.mbuf = arp;
2512 err = rte_event_enqueue_burst(evdev, t->port[rx_enq], &ev, 1);
2513 if (err != 1) {
2514 printf("%d: Failed to enqueue\n", __LINE__);
2515 goto err;
2516 }
2517 }
2518
2519 /* schedule */
2520 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2521
2522 err = test_event_dev_stats_get(evdev, &stats);
2523 if (err) {
2524 printf("%d: failed to get stats\n", __LINE__);
2525 goto err;
2526 }
2527
2528 if (stats.rx_pkts != QID1_NUM + QID2_NUM ||
2529 stats.tx_pkts != QID1_NUM + QID2_NUM) {
2530 printf("%d: Sched core didn't handle pkt as expected\n",
2531 __LINE__);
2532 goto err;
2533 }
2534
2535 if (stats.port_inflight[p1] != QID1_NUM) {
2536 printf("%d: %s port 1 inflight not correct\n", __LINE__,
2537 __func__);
2538 goto err;
2539 }
2540 if (stats.port_inflight[p2] != QID2_NUM) {
2541 printf("%d: %s port 2 inflight not correct\n", __LINE__,
2542 __func__);
2543 goto err;
2544 }
2545
2546 /************** DEQUEUE INFLIGHT COUNT CHECKS ****************/
2547 /* port 1 */
2548 struct rte_event events[QID1_NUM + QID2_NUM];
2549 uint32_t deq_pkts = rte_event_dequeue_burst(evdev, t->port[p1], events,
2550 RTE_DIM(events), 0);
2551
2552 if (deq_pkts != QID1_NUM) {
2553 printf("%d: Port 1: DEQUEUE inflight failed\n", __LINE__);
2554 goto err;
2555 }
2556 err = test_event_dev_stats_get(evdev, &stats);
2557 if (stats.port_inflight[p1] != QID1_NUM) {
2558 printf("%d: port 1 inflight decrement after DEQ != 0\n",
2559 __LINE__);
2560 goto err;
2561 }
2562 for (i = 0; i < QID1_NUM; i++) {
2563 err = rte_event_enqueue_burst(evdev, t->port[p1], &release_ev,
2564 1);
2565 if (err != 1) {
2566 printf("%d: %s rte enqueue of inf release failed\n",
2567 __LINE__, __func__);
2568 goto err;
2569 }
2570 }
2571
2572 /*
2573 * As the scheduler core decrements inflights, it needs to run to
2574 * process packets to act on the drop messages
2575 */
2576 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2577
2578 err = test_event_dev_stats_get(evdev, &stats);
2579 if (stats.port_inflight[p1] != 0) {
2580 printf("%d: port 1 inflight NON NULL after DROP\n", __LINE__);
2581 goto err;
2582 }
2583
2584 /* port2 */
2585 deq_pkts = rte_event_dequeue_burst(evdev, t->port[p2], events,
2586 RTE_DIM(events), 0);
2587 if (deq_pkts != QID2_NUM) {
2588 printf("%d: Port 2: DEQUEUE inflight failed\n", __LINE__);
2589 goto err;
2590 }
2591 err = test_event_dev_stats_get(evdev, &stats);
2592 if (stats.port_inflight[p2] != QID2_NUM) {
2593 printf("%d: port 1 inflight decrement after DEQ != 0\n",
2594 __LINE__);
2595 goto err;
2596 }
2597 for (i = 0; i < QID2_NUM; i++) {
2598 err = rte_event_enqueue_burst(evdev, t->port[p2], &release_ev,
2599 1);
2600 if (err != 1) {
2601 printf("%d: %s rte enqueue of inf release failed\n",
2602 __LINE__, __func__);
2603 goto err;
2604 }
2605 }
2606
2607 /*
2608 * As the scheduler core decrements inflights, it needs to run to
2609 * process packets to act on the drop messages
2610 */
2611 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2612
2613 err = test_event_dev_stats_get(evdev, &stats);
2614 if (stats.port_inflight[p2] != 0) {
2615 printf("%d: port 2 inflight NON NULL after DROP\n", __LINE__);
2616 goto err;
2617 }
2618 cleanup(t);
2619 return 0;
2620
2621 err:
2622 rte_event_dev_dump(evdev, stdout);
2623 cleanup(t);
2624 return -1;
2625 }
2626
2627 static int
parallel_basic(struct test * t,int check_order)2628 parallel_basic(struct test *t, int check_order)
2629 {
2630 const uint8_t rx_port = 0;
2631 const uint8_t w1_port = 1;
2632 const uint8_t w3_port = 3;
2633 const uint8_t tx_port = 4;
2634 int err;
2635 int i;
2636 uint32_t deq_pkts, j;
2637 struct rte_mbuf *mbufs[3];
2638 struct rte_mbuf *mbufs_out[3] = { 0 };
2639 const uint32_t MAGIC_SEQN = 1234;
2640
2641 /* Create instance with 4 ports */
2642 if (init(t, 2, tx_port + 1) < 0 ||
2643 create_ports(t, tx_port + 1) < 0 ||
2644 (check_order ? create_ordered_qids(t, 1) :
2645 create_unordered_qids(t, 1)) < 0 ||
2646 create_directed_qids(t, 1, &tx_port)) {
2647 printf("%d: Error initializing device\n", __LINE__);
2648 return -1;
2649 }
2650
2651 /*
2652 * CQ mapping to QID
2653 * We need three ports, all mapped to the same ordered qid0. Then we'll
2654 * take a packet out to each port, re-enqueue in reverse order,
2655 * then make sure the reordering has taken place properly when we
2656 * dequeue from the tx_port.
2657 *
2658 * Simplified test setup diagram:
2659 *
2660 * rx_port w1_port
2661 * \ / \
2662 * qid0 - w2_port - qid1
2663 * \ / \
2664 * w3_port tx_port
2665 */
2666 /* CQ mapping to QID for LB ports (directed mapped on create) */
2667 for (i = w1_port; i <= w3_port; i++) {
2668 err = rte_event_port_link(evdev, t->port[i], &t->qid[0], NULL,
2669 1);
2670 if (err != 1) {
2671 printf("%d: error mapping lb qid\n", __LINE__);
2672 cleanup(t);
2673 return -1;
2674 }
2675 }
2676
2677 if (rte_event_dev_start(evdev) < 0) {
2678 printf("%d: Error with start call\n", __LINE__);
2679 return -1;
2680 }
2681
2682 /* Enqueue 3 packets to the rx port */
2683 for (i = 0; i < 3; i++) {
2684 struct rte_event ev;
2685 mbufs[i] = rte_gen_arp(0, t->mbuf_pool);
2686 if (!mbufs[i]) {
2687 printf("%d: gen of pkt failed\n", __LINE__);
2688 return -1;
2689 }
2690
2691 ev.queue_id = t->qid[0];
2692 ev.op = RTE_EVENT_OP_NEW;
2693 ev.mbuf = mbufs[i];
2694 *rte_event_pmd_selftest_seqn(mbufs[i]) = MAGIC_SEQN + i;
2695
2696 /* generate pkt and enqueue */
2697 err = rte_event_enqueue_burst(evdev, t->port[rx_port], &ev, 1);
2698 if (err != 1) {
2699 printf("%d: Failed to enqueue pkt %u, retval = %u\n",
2700 __LINE__, i, err);
2701 return -1;
2702 }
2703 }
2704
2705 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2706
2707 /* use extra slot to make logic in loops easier */
2708 struct rte_event deq_ev[w3_port + 1];
2709
2710 /* Dequeue the 3 packets, one from each worker port */
2711 for (i = w1_port; i <= w3_port; i++) {
2712 deq_pkts = rte_event_dequeue_burst(evdev, t->port[i],
2713 &deq_ev[i], 1, 0);
2714 if (deq_pkts != 1) {
2715 printf("%d: Failed to deq\n", __LINE__);
2716 rte_event_dev_dump(evdev, stdout);
2717 return -1;
2718 }
2719 }
2720
2721 /* Enqueue each packet in reverse order, flushing after each one */
2722 for (i = w3_port; i >= w1_port; i--) {
2723
2724 deq_ev[i].op = RTE_EVENT_OP_FORWARD;
2725 deq_ev[i].queue_id = t->qid[1];
2726 err = rte_event_enqueue_burst(evdev, t->port[i], &deq_ev[i], 1);
2727 if (err != 1) {
2728 printf("%d: Failed to enqueue\n", __LINE__);
2729 return -1;
2730 }
2731 }
2732 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2733
2734 /* dequeue from the tx ports, we should get 3 packets */
2735 deq_pkts = rte_event_dequeue_burst(evdev, t->port[tx_port], deq_ev,
2736 3, 0);
2737
2738 /* Check to see if we've got all 3 packets */
2739 if (deq_pkts != 3) {
2740 printf("%d: expected 3 pkts at tx port got %d from port %d\n",
2741 __LINE__, deq_pkts, tx_port);
2742 rte_event_dev_dump(evdev, stdout);
2743 return 1;
2744 }
2745
2746 /* Check to see if the sequence numbers are in expected order */
2747 if (check_order) {
2748 for (j = 0 ; j < deq_pkts ; j++) {
2749 if (*rte_event_pmd_selftest_seqn(deq_ev[j].mbuf) !=
2750 MAGIC_SEQN + j) {
2751 printf("%d: Incorrect sequence number(%d) from port %d\n",
2752 __LINE__,
2753 *rte_event_pmd_selftest_seqn(mbufs_out[j]),
2754 tx_port);
2755 return -1;
2756 }
2757 }
2758 }
2759
2760 /* Destroy the instance */
2761 cleanup(t);
2762 return 0;
2763 }
2764
2765 static int
ordered_basic(struct test * t)2766 ordered_basic(struct test *t)
2767 {
2768 return parallel_basic(t, 1);
2769 }
2770
2771 static int
unordered_basic(struct test * t)2772 unordered_basic(struct test *t)
2773 {
2774 return parallel_basic(t, 0);
2775 }
2776
2777 static int
holb(struct test * t)2778 holb(struct test *t) /* test to check we avoid basic head-of-line blocking */
2779 {
2780 const struct rte_event new_ev = {
2781 .op = RTE_EVENT_OP_NEW
2782 /* all other fields zero */
2783 };
2784 struct rte_event ev = new_ev;
2785 unsigned int rx_port = 0; /* port we get the first flow on */
2786 char rx_port_used_stat[64];
2787 char rx_port_free_stat[64];
2788 char other_port_used_stat[64];
2789
2790 if (init(t, 1, 2) < 0 ||
2791 create_ports(t, 2) < 0 ||
2792 create_atomic_qids(t, 1) < 0) {
2793 printf("%d: Error initializing device\n", __LINE__);
2794 return -1;
2795 }
2796 int nb_links = rte_event_port_link(evdev, t->port[1], NULL, NULL, 0);
2797 if (rte_event_port_link(evdev, t->port[0], NULL, NULL, 0) != 1 ||
2798 nb_links != 1) {
2799 printf("%d: Error links queue to ports\n", __LINE__);
2800 goto err;
2801 }
2802 if (rte_event_dev_start(evdev) < 0) {
2803 printf("%d: Error with start call\n", __LINE__);
2804 goto err;
2805 }
2806
2807 /* send one packet and see where it goes, port 0 or 1 */
2808 if (rte_event_enqueue_burst(evdev, t->port[0], &ev, 1) != 1) {
2809 printf("%d: Error doing first enqueue\n", __LINE__);
2810 goto err;
2811 }
2812 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2813
2814 if (rte_event_dev_xstats_by_name_get(evdev, "port_0_cq_ring_used", NULL)
2815 != 1)
2816 rx_port = 1;
2817
2818 snprintf(rx_port_used_stat, sizeof(rx_port_used_stat),
2819 "port_%u_cq_ring_used", rx_port);
2820 snprintf(rx_port_free_stat, sizeof(rx_port_free_stat),
2821 "port_%u_cq_ring_free", rx_port);
2822 snprintf(other_port_used_stat, sizeof(other_port_used_stat),
2823 "port_%u_cq_ring_used", rx_port ^ 1);
2824 if (rte_event_dev_xstats_by_name_get(evdev, rx_port_used_stat, NULL)
2825 != 1) {
2826 printf("%d: Error, first event not scheduled\n", __LINE__);
2827 goto err;
2828 }
2829
2830 /* now fill up the rx port's queue with one flow to cause HOLB */
2831 do {
2832 ev = new_ev;
2833 if (rte_event_enqueue_burst(evdev, t->port[0], &ev, 1) != 1) {
2834 printf("%d: Error with enqueue\n", __LINE__);
2835 goto err;
2836 }
2837 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2838 } while (rte_event_dev_xstats_by_name_get(evdev,
2839 rx_port_free_stat, NULL) != 0);
2840
2841 /* one more packet, which needs to stay in IQ - i.e. HOLB */
2842 ev = new_ev;
2843 if (rte_event_enqueue_burst(evdev, t->port[0], &ev, 1) != 1) {
2844 printf("%d: Error with enqueue\n", __LINE__);
2845 goto err;
2846 }
2847 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2848
2849 /* check that the other port still has an empty CQ */
2850 if (rte_event_dev_xstats_by_name_get(evdev, other_port_used_stat, NULL)
2851 != 0) {
2852 printf("%d: Error, second port CQ is not empty\n", __LINE__);
2853 goto err;
2854 }
2855 /* check IQ now has one packet */
2856 if (rte_event_dev_xstats_by_name_get(evdev, "qid_0_iq_0_used", NULL)
2857 != 1) {
2858 printf("%d: Error, QID does not have exactly 1 packet\n",
2859 __LINE__);
2860 goto err;
2861 }
2862
2863 /* send another flow, which should pass the other IQ entry */
2864 ev = new_ev;
2865 ev.flow_id = 1;
2866 if (rte_event_enqueue_burst(evdev, t->port[0], &ev, 1) != 1) {
2867 printf("%d: Error with enqueue\n", __LINE__);
2868 goto err;
2869 }
2870 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2871
2872 if (rte_event_dev_xstats_by_name_get(evdev, other_port_used_stat, NULL)
2873 != 1) {
2874 printf("%d: Error, second flow did not pass out first\n",
2875 __LINE__);
2876 goto err;
2877 }
2878
2879 if (rte_event_dev_xstats_by_name_get(evdev, "qid_0_iq_0_used", NULL)
2880 != 1) {
2881 printf("%d: Error, QID does not have exactly 1 packet\n",
2882 __LINE__);
2883 goto err;
2884 }
2885 cleanup(t);
2886 return 0;
2887 err:
2888 rte_event_dev_dump(evdev, stdout);
2889 cleanup(t);
2890 return -1;
2891 }
2892
2893 static void
flush(uint8_t dev_id __rte_unused,struct rte_event event,void * arg)2894 flush(uint8_t dev_id __rte_unused, struct rte_event event, void *arg)
2895 {
2896 *((uint8_t *) arg) += (event.u64 == 0xCA11BACC) ? 1 : 0;
2897 }
2898
2899 static int
dev_stop_flush(struct test * t)2900 dev_stop_flush(struct test *t) /* test to check we can properly flush events */
2901 {
2902 const struct rte_event new_ev = {
2903 .op = RTE_EVENT_OP_NEW,
2904 .u64 = 0xCA11BACC,
2905 .queue_id = 0
2906 };
2907 struct rte_event ev = new_ev;
2908 uint8_t count = 0;
2909 int i;
2910
2911 if (init(t, 1, 1) < 0 ||
2912 create_ports(t, 1) < 0 ||
2913 create_atomic_qids(t, 1) < 0) {
2914 printf("%d: Error initializing device\n", __LINE__);
2915 return -1;
2916 }
2917
2918 /* Link the queue so *_start() doesn't error out */
2919 if (rte_event_port_link(evdev, t->port[0], NULL, NULL, 0) != 1) {
2920 printf("%d: Error linking queue to port\n", __LINE__);
2921 goto err;
2922 }
2923
2924 if (rte_event_dev_start(evdev) < 0) {
2925 printf("%d: Error with start call\n", __LINE__);
2926 goto err;
2927 }
2928
2929 for (i = 0; i < DEQUEUE_DEPTH + 1; i++) {
2930 if (rte_event_enqueue_burst(evdev, t->port[0], &ev, 1) != 1) {
2931 printf("%d: Error enqueuing events\n", __LINE__);
2932 goto err;
2933 }
2934 }
2935
2936 /* Schedule the events from the port to the IQ. At least one event
2937 * should be remaining in the queue.
2938 */
2939 rte_service_run_iter_on_app_lcore(t->service_id, 1);
2940
2941 if (rte_event_dev_stop_flush_callback_register(evdev, flush, &count)) {
2942 printf("%d: Error installing the flush callback\n", __LINE__);
2943 goto err;
2944 }
2945
2946 cleanup(t);
2947
2948 if (count == 0) {
2949 printf("%d: Error executing the flush callback\n", __LINE__);
2950 goto err;
2951 }
2952
2953 if (rte_event_dev_stop_flush_callback_register(evdev, NULL, NULL)) {
2954 printf("%d: Error uninstalling the flush callback\n", __LINE__);
2955 goto err;
2956 }
2957
2958 return 0;
2959 err:
2960 rte_event_dev_dump(evdev, stdout);
2961 cleanup(t);
2962 return -1;
2963 }
2964
2965 static int
worker_loopback_worker_fn(void * arg)2966 worker_loopback_worker_fn(void *arg)
2967 {
2968 struct test *t = arg;
2969 uint8_t port = t->port[1];
2970 int count = 0;
2971 int enqd;
2972
2973 /*
2974 * Takes packets from the input port and then loops them back through
2975 * the Eventdev. Each packet gets looped through QIDs 0-8, 16 times
2976 * so each packet goes through 8*16 = 128 times.
2977 */
2978 printf("%d: \tWorker function started\n", __LINE__);
2979 while (count < NUM_PACKETS) {
2980 #define BURST_SIZE 32
2981 struct rte_event ev[BURST_SIZE];
2982 uint16_t i, nb_rx = rte_event_dequeue_burst(evdev, port, ev,
2983 BURST_SIZE, 0);
2984 if (nb_rx == 0) {
2985 rte_pause();
2986 continue;
2987 }
2988
2989 for (i = 0; i < nb_rx; i++) {
2990 ev[i].queue_id++;
2991 if (ev[i].queue_id != 8) {
2992 ev[i].op = RTE_EVENT_OP_FORWARD;
2993 enqd = rte_event_enqueue_burst(evdev, port,
2994 &ev[i], 1);
2995 if (enqd != 1) {
2996 printf("%d: Can't enqueue FWD!!\n",
2997 __LINE__);
2998 return -1;
2999 }
3000 continue;
3001 }
3002
3003 ev[i].queue_id = 0;
3004 (*counter_field(ev[i].mbuf))++;
3005 if (*counter_field(ev[i].mbuf) != 16) {
3006 ev[i].op = RTE_EVENT_OP_FORWARD;
3007 enqd = rte_event_enqueue_burst(evdev, port,
3008 &ev[i], 1);
3009 if (enqd != 1) {
3010 printf("%d: Can't enqueue FWD!!\n",
3011 __LINE__);
3012 return -1;
3013 }
3014 continue;
3015 }
3016 /* we have hit 16 iterations through system - drop */
3017 rte_pktmbuf_free(ev[i].mbuf);
3018 count++;
3019 ev[i].op = RTE_EVENT_OP_RELEASE;
3020 enqd = rte_event_enqueue_burst(evdev, port, &ev[i], 1);
3021 if (enqd != 1) {
3022 printf("%d drop enqueue failed\n", __LINE__);
3023 return -1;
3024 }
3025 }
3026 }
3027
3028 return 0;
3029 }
3030
3031 static int
worker_loopback_producer_fn(void * arg)3032 worker_loopback_producer_fn(void *arg)
3033 {
3034 struct test *t = arg;
3035 uint8_t port = t->port[0];
3036 uint64_t count = 0;
3037
3038 printf("%d: \tProducer function started\n", __LINE__);
3039 while (count < NUM_PACKETS) {
3040 struct rte_mbuf *m = 0;
3041 do {
3042 m = rte_pktmbuf_alloc(t->mbuf_pool);
3043 } while (m == NULL);
3044
3045 *counter_field(m) = 0;
3046
3047 struct rte_event ev = {
3048 .op = RTE_EVENT_OP_NEW,
3049 .queue_id = t->qid[0],
3050 .flow_id = (uintptr_t)m & 0xFFFF,
3051 .mbuf = m,
3052 };
3053
3054 if (rte_event_enqueue_burst(evdev, port, &ev, 1) != 1) {
3055 while (rte_event_enqueue_burst(evdev, port, &ev, 1) !=
3056 1)
3057 rte_pause();
3058 }
3059
3060 count++;
3061 }
3062
3063 return 0;
3064 }
3065
3066 static int
worker_loopback(struct test * t,uint8_t disable_implicit_release)3067 worker_loopback(struct test *t, uint8_t disable_implicit_release)
3068 {
3069 /* use a single producer core, and a worker core to see what happens
3070 * if the worker loops packets back multiple times
3071 */
3072 struct test_event_dev_stats stats;
3073 uint64_t print_cycles = 0, cycles = 0;
3074 uint64_t tx_pkts = 0;
3075 int err;
3076 int w_lcore, p_lcore;
3077
3078 static const struct rte_mbuf_dynfield counter_dynfield_desc = {
3079 .name = "rte_event_sw_dynfield_selftest_counter",
3080 .size = sizeof(counter_dynfield_t),
3081 .align = __alignof__(counter_dynfield_t),
3082 };
3083 counter_dynfield_offset =
3084 rte_mbuf_dynfield_register(&counter_dynfield_desc);
3085 if (counter_dynfield_offset < 0) {
3086 printf("Error registering mbuf field\n");
3087 return -rte_errno;
3088 }
3089
3090 if (init(t, 8, 2) < 0 ||
3091 create_atomic_qids(t, 8) < 0) {
3092 printf("%d: Error initializing device\n", __LINE__);
3093 return -1;
3094 }
3095
3096 /* RX with low max events */
3097 static struct rte_event_port_conf conf = {
3098 .dequeue_depth = 32,
3099 .enqueue_depth = 64,
3100 };
3101 /* beware: this cannot be initialized in the static above as it would
3102 * only be initialized once - and this needs to be set for multiple runs
3103 */
3104 conf.new_event_threshold = 512;
3105 conf.event_port_cfg = disable_implicit_release ?
3106 RTE_EVENT_PORT_CFG_DISABLE_IMPL_REL : 0;
3107
3108 if (rte_event_port_setup(evdev, 0, &conf) < 0) {
3109 printf("Error setting up RX port\n");
3110 return -1;
3111 }
3112 t->port[0] = 0;
3113 /* TX with higher max events */
3114 conf.new_event_threshold = 4096;
3115 if (rte_event_port_setup(evdev, 1, &conf) < 0) {
3116 printf("Error setting up TX port\n");
3117 return -1;
3118 }
3119 t->port[1] = 1;
3120
3121 /* CQ mapping to QID */
3122 err = rte_event_port_link(evdev, t->port[1], NULL, NULL, 0);
3123 if (err != 8) { /* should have mapped all queues*/
3124 printf("%d: error mapping port 2 to all qids\n", __LINE__);
3125 return -1;
3126 }
3127
3128 if (rte_event_dev_start(evdev) < 0) {
3129 printf("%d: Error with start call\n", __LINE__);
3130 return -1;
3131 }
3132
3133 p_lcore = rte_get_next_lcore(
3134 /* start core */ -1,
3135 /* skip main */ 1,
3136 /* wrap */ 0);
3137 w_lcore = rte_get_next_lcore(p_lcore, 1, 0);
3138
3139 rte_eal_remote_launch(worker_loopback_producer_fn, t, p_lcore);
3140 rte_eal_remote_launch(worker_loopback_worker_fn, t, w_lcore);
3141
3142 print_cycles = cycles = rte_get_timer_cycles();
3143 while (rte_eal_get_lcore_state(p_lcore) != WAIT ||
3144 rte_eal_get_lcore_state(w_lcore) != WAIT) {
3145
3146 rte_service_run_iter_on_app_lcore(t->service_id, 1);
3147
3148 uint64_t new_cycles = rte_get_timer_cycles();
3149
3150 if (new_cycles - print_cycles > rte_get_timer_hz()) {
3151 test_event_dev_stats_get(evdev, &stats);
3152 printf(
3153 "%d: \tSched Rx = %"PRIu64", Tx = %"PRIu64"\n",
3154 __LINE__, stats.rx_pkts, stats.tx_pkts);
3155
3156 print_cycles = new_cycles;
3157 }
3158 if (new_cycles - cycles > rte_get_timer_hz() * 3) {
3159 test_event_dev_stats_get(evdev, &stats);
3160 if (stats.tx_pkts == tx_pkts) {
3161 rte_event_dev_dump(evdev, stdout);
3162 printf("Dumping xstats:\n");
3163 xstats_print();
3164 printf(
3165 "%d: No schedules for seconds, deadlock\n",
3166 __LINE__);
3167 return -1;
3168 }
3169 tx_pkts = stats.tx_pkts;
3170 cycles = new_cycles;
3171 }
3172 }
3173 rte_service_run_iter_on_app_lcore(t->service_id, 1);
3174 /* ensure all completions are flushed */
3175
3176 rte_eal_mp_wait_lcore();
3177
3178 cleanup(t);
3179 return 0;
3180 }
3181
3182 static struct rte_mempool *eventdev_func_mempool;
3183
3184 int
test_sw_eventdev(void)3185 test_sw_eventdev(void)
3186 {
3187 struct test *t;
3188 int ret;
3189
3190 t = malloc(sizeof(struct test));
3191 if (t == NULL)
3192 return -1;
3193 /* manually initialize the op, older gcc's complain on static
3194 * initialization of struct elements that are a bitfield.
3195 */
3196 release_ev.op = RTE_EVENT_OP_RELEASE;
3197
3198 const char *eventdev_name = "event_sw";
3199 evdev = rte_event_dev_get_dev_id(eventdev_name);
3200 if (evdev < 0) {
3201 printf("%d: Eventdev %s not found - creating.\n",
3202 __LINE__, eventdev_name);
3203 if (rte_vdev_init(eventdev_name, NULL) < 0) {
3204 printf("Error creating eventdev\n");
3205 goto test_fail;
3206 }
3207 evdev = rte_event_dev_get_dev_id(eventdev_name);
3208 if (evdev < 0) {
3209 printf("Error finding newly created eventdev\n");
3210 goto test_fail;
3211 }
3212 }
3213
3214 if (rte_event_dev_service_id_get(evdev, &t->service_id) < 0) {
3215 printf("Failed to get service ID for software event dev\n");
3216 goto test_fail;
3217 }
3218
3219 rte_service_runstate_set(t->service_id, 1);
3220 rte_service_set_runstate_mapped_check(t->service_id, 0);
3221
3222 /* Only create mbuf pool once, reuse for each test run */
3223 if (!eventdev_func_mempool) {
3224 eventdev_func_mempool = rte_pktmbuf_pool_create(
3225 "EVENTDEV_SW_SA_MBUF_POOL",
3226 (1<<12), /* 4k buffers */
3227 32 /*MBUF_CACHE_SIZE*/,
3228 0,
3229 512, /* use very small mbufs */
3230 rte_socket_id());
3231 if (!eventdev_func_mempool) {
3232 printf("ERROR creating mempool\n");
3233 goto test_fail;
3234 }
3235 }
3236 t->mbuf_pool = eventdev_func_mempool;
3237 printf("*** Running Single Directed Packet test...\n");
3238 ret = test_single_directed_packet(t);
3239 if (ret != 0) {
3240 printf("ERROR - Single Directed Packet test FAILED.\n");
3241 goto test_fail;
3242 }
3243 printf("*** Running Directed Forward Credit test...\n");
3244 ret = test_directed_forward_credits(t);
3245 if (ret != 0) {
3246 printf("ERROR - Directed Forward Credit test FAILED.\n");
3247 goto test_fail;
3248 }
3249 printf("*** Running Single Load Balanced Packet test...\n");
3250 ret = single_packet(t);
3251 if (ret != 0) {
3252 printf("ERROR - Single Packet test FAILED.\n");
3253 goto test_fail;
3254 }
3255 printf("*** Running Unordered Basic test...\n");
3256 ret = unordered_basic(t);
3257 if (ret != 0) {
3258 printf("ERROR - Unordered Basic test FAILED.\n");
3259 goto test_fail;
3260 }
3261 printf("*** Running Ordered Basic test...\n");
3262 ret = ordered_basic(t);
3263 if (ret != 0) {
3264 printf("ERROR - Ordered Basic test FAILED.\n");
3265 goto test_fail;
3266 }
3267 printf("*** Running Burst Packets test...\n");
3268 ret = burst_packets(t);
3269 if (ret != 0) {
3270 printf("ERROR - Burst Packets test FAILED.\n");
3271 goto test_fail;
3272 }
3273 printf("*** Running Load Balancing test...\n");
3274 ret = load_balancing(t);
3275 if (ret != 0) {
3276 printf("ERROR - Load Balancing test FAILED.\n");
3277 goto test_fail;
3278 }
3279 printf("*** Running Prioritized Directed test...\n");
3280 ret = test_priority_directed(t);
3281 if (ret != 0) {
3282 printf("ERROR - Prioritized Directed test FAILED.\n");
3283 goto test_fail;
3284 }
3285 printf("*** Running Prioritized Atomic test...\n");
3286 ret = test_priority_atomic(t);
3287 if (ret != 0) {
3288 printf("ERROR - Prioritized Atomic test FAILED.\n");
3289 goto test_fail;
3290 }
3291
3292 printf("*** Running Prioritized Ordered test...\n");
3293 ret = test_priority_ordered(t);
3294 if (ret != 0) {
3295 printf("ERROR - Prioritized Ordered test FAILED.\n");
3296 goto test_fail;
3297 }
3298 printf("*** Running Prioritized Unordered test...\n");
3299 ret = test_priority_unordered(t);
3300 if (ret != 0) {
3301 printf("ERROR - Prioritized Unordered test FAILED.\n");
3302 goto test_fail;
3303 }
3304 printf("*** Running Invalid QID test...\n");
3305 ret = invalid_qid(t);
3306 if (ret != 0) {
3307 printf("ERROR - Invalid QID test FAILED.\n");
3308 goto test_fail;
3309 }
3310 printf("*** Running Load Balancing History test...\n");
3311 ret = load_balancing_history(t);
3312 if (ret != 0) {
3313 printf("ERROR - Load Balancing History test FAILED.\n");
3314 goto test_fail;
3315 }
3316 printf("*** Running Inflight Count test...\n");
3317 ret = inflight_counts(t);
3318 if (ret != 0) {
3319 printf("ERROR - Inflight Count test FAILED.\n");
3320 goto test_fail;
3321 }
3322 printf("*** Running Abuse Inflights test...\n");
3323 ret = abuse_inflights(t);
3324 if (ret != 0) {
3325 printf("ERROR - Abuse Inflights test FAILED.\n");
3326 goto test_fail;
3327 }
3328 printf("*** Running XStats test...\n");
3329 ret = xstats_tests(t);
3330 if (ret != 0) {
3331 printf("ERROR - XStats test FAILED.\n");
3332 goto test_fail;
3333 }
3334 printf("*** Running XStats ID Reset test...\n");
3335 ret = xstats_id_reset_tests(t);
3336 if (ret != 0) {
3337 printf("ERROR - XStats ID Reset test FAILED.\n");
3338 goto test_fail;
3339 }
3340 printf("*** Running XStats Brute Force test...\n");
3341 ret = xstats_brute_force(t);
3342 if (ret != 0) {
3343 printf("ERROR - XStats Brute Force test FAILED.\n");
3344 goto test_fail;
3345 }
3346 printf("*** Running XStats ID Abuse test...\n");
3347 ret = xstats_id_abuse_tests(t);
3348 if (ret != 0) {
3349 printf("ERROR - XStats ID Abuse test FAILED.\n");
3350 goto test_fail;
3351 }
3352 printf("*** Running QID Priority test...\n");
3353 ret = qid_priorities(t);
3354 if (ret != 0) {
3355 printf("ERROR - QID Priority test FAILED.\n");
3356 goto test_fail;
3357 }
3358 printf("*** Running Unlink-in-progress test...\n");
3359 ret = unlink_in_progress(t);
3360 if (ret != 0) {
3361 printf("ERROR - Unlink in progress test FAILED.\n");
3362 goto test_fail;
3363 }
3364 printf("*** Running Ordered Reconfigure test...\n");
3365 ret = ordered_reconfigure(t);
3366 if (ret != 0) {
3367 printf("ERROR - Ordered Reconfigure test FAILED.\n");
3368 goto test_fail;
3369 }
3370 printf("*** Running Port LB Single Reconfig test...\n");
3371 ret = port_single_lb_reconfig(t);
3372 if (ret != 0) {
3373 printf("ERROR - Port LB Single Reconfig test FAILED.\n");
3374 goto test_fail;
3375 }
3376 printf("*** Running Port Reconfig Credits test...\n");
3377 ret = port_reconfig_credits(t);
3378 if (ret != 0) {
3379 printf("ERROR - Port Reconfig Credits Reset test FAILED.\n");
3380 goto test_fail;
3381 }
3382 printf("*** Running Head-of-line-blocking test...\n");
3383 ret = holb(t);
3384 if (ret != 0) {
3385 printf("ERROR - Head-of-line-blocking test FAILED.\n");
3386 goto test_fail;
3387 }
3388 printf("*** Running Stop Flush test...\n");
3389 ret = dev_stop_flush(t);
3390 if (ret != 0) {
3391 printf("ERROR - Stop Flush test FAILED.\n");
3392 goto test_fail;
3393 }
3394 if (rte_lcore_count() >= 3) {
3395 printf("*** Running Worker loopback test...\n");
3396 ret = worker_loopback(t, 0);
3397 if (ret != 0) {
3398 printf("ERROR - Worker loopback test FAILED.\n");
3399 return ret;
3400 }
3401
3402 printf("*** Running Worker loopback test (implicit release disabled)...\n");
3403 ret = worker_loopback(t, 1);
3404 if (ret != 0) {
3405 printf("ERROR - Worker loopback test FAILED.\n");
3406 goto test_fail;
3407 }
3408 } else {
3409 printf("### Not enough cores for worker loopback tests.\n");
3410 printf("### Need at least 3 cores for the tests.\n");
3411 }
3412
3413 /*
3414 * Free test instance, leaving mempool initialized, and a pointer to it
3415 * in static eventdev_func_mempool, as it is re-used on re-runs
3416 */
3417 free(t);
3418
3419 printf("SW Eventdev Selftest Successful.\n");
3420 return 0;
3421 test_fail:
3422 free(t);
3423 printf("SW Eventdev Selftest Failed.\n");
3424 return -1;
3425 }
3426