Lines Matching refs:ev

19 mark_fwd_latency(struct rte_event *const ev,  in mark_fwd_latency()  argument
22 if (unlikely((ev->queue_id % nb_stages) == 0)) { in mark_fwd_latency()
23 struct perf_elt *const m = ev->event_ptr; in mark_fwd_latency()
30 fwd_event(struct rte_event *const ev, uint8_t *const sched_type_list, in fwd_event() argument
33 ev->queue_id++; in fwd_event()
34 ev->sched_type = sched_type_list[ev->queue_id % nb_stages]; in fwd_event()
35 ev->op = RTE_EVENT_OP_FORWARD; in fwd_event()
36 ev->event_type = RTE_EVENT_TYPE_CPU; in fwd_event()
43 struct rte_event ev; in perf_queue_worker() local
46 uint16_t event = rte_event_dequeue_burst(dev, port, &ev, 1, 0); in perf_queue_worker()
54 mark_fwd_latency(&ev, nb_stages); in perf_queue_worker()
57 if (unlikely((ev.queue_id % nb_stages) == laststage)) { in perf_queue_worker()
60 &ev, w, bufs, sz, cnt); in perf_queue_worker()
63 &ev, w, bufs, sz, cnt); in perf_queue_worker()
65 fwd_event(&ev, sched_type_list, nb_stages); in perf_queue_worker()
66 while (rte_event_enqueue_burst(dev, port, &ev, 1) != 1) in perf_queue_worker()
79 struct rte_event ev[BURST_SIZE + 1]; in perf_queue_worker_burst() local
82 uint16_t const nb_rx = rte_event_dequeue_burst(dev, port, ev, in perf_queue_worker_burst()
92 rte_prefetch0(ev[i+1].event_ptr); in perf_queue_worker_burst()
96 mark_fwd_latency(&ev[i], nb_stages); in perf_queue_worker_burst()
99 if (unlikely((ev[i].queue_id % nb_stages) == in perf_queue_worker_burst()
103 pool, &ev[i], w, bufs, sz, cnt); in perf_queue_worker_burst()
106 &ev[i], w, bufs, sz, cnt); in perf_queue_worker_burst()
108 ev[i].op = RTE_EVENT_OP_RELEASE; in perf_queue_worker_burst()
110 fwd_event(&ev[i], sched_type_list, nb_stages); in perf_queue_worker_burst()
116 enq = rte_event_enqueue_burst(dev, port, ev, nb_rx); in perf_queue_worker_burst()
119 ev + enq, nb_rx - enq); in perf_queue_worker_burst()