| /dpdk/examples/l3fwd/ |
| H A D | l3fwd_fib.c | 279 events, deq_len, 0); in fib_event_loop() 298 events[i].queue_id = tx_q_id; in fib_event_loop() 311 fib_parse_packet(events[i].mbuf, in fib_event_loop() 320 events[i].queue_id = tx_q_id; in fib_event_loop() 328 fib_parse_packet(events[i].mbuf, in fib_event_loop() 351 events[i].mbuf->port = nh; in fib_event_loop() 356 events, nb_deq); in fib_event_loop() 359 event_p_id, events + nb_enq, in fib_event_loop() 369 events + nb_enq, in fib_event_loop() 516 events[i].queue_id = tx_q_id; in fib_event_loop_vector() [all …]
|
| H A D | l3fwd_lpm.c | 317 struct rte_event events[MAX_PKT_BURST]; in lpm_event_loop_burst() local 334 events, deq_len, 0); in lpm_event_loop_burst() 342 events[i].queue_id = tx_q_id; in lpm_event_loop_burst() 343 events[i].op = RTE_EVENT_OP_FORWARD; in lpm_event_loop_burst() 355 events, nb_deq); in lpm_event_loop_burst() 358 event_p_id, events + nb_enq, in lpm_event_loop_burst() 364 event_p_id, events, nb_deq, 0); in lpm_event_loop_burst() 368 events + nb_enq, in lpm_event_loop_burst() 458 struct rte_event events[MAX_PKT_BURST]; in lpm_event_loop_vector() local 482 events[i].queue_id = tx_q_id; in lpm_event_loop_vector() [all …]
|
| H A D | l3fwd_event.c | 292 l3fwd_event_vector_array_free(struct rte_event events[], uint16_t num) in l3fwd_event_vector_array_free() argument 297 rte_pktmbuf_free_bulk(events[i].vec->mbufs, in l3fwd_event_vector_array_free() 298 events[i].vec->nb_elem); in l3fwd_event_vector_array_free() 299 rte_mempool_put(rte_mempool_from_obj(events[i].vec), in l3fwd_event_vector_array_free() 300 events[i].vec); in l3fwd_event_vector_array_free() 316 struct rte_event events[], uint16_t nb_enq, in l3fwd_event_worker_cleanup() argument 323 l3fwd_event_vector_array_free(events + nb_enq, in l3fwd_event_worker_cleanup() 327 rte_pktmbuf_free(events[i].mbuf); in l3fwd_event_worker_cleanup() 330 events[i].op = RTE_EVENT_OP_RELEASE; in l3fwd_event_worker_cleanup() 331 rte_event_enqueue_burst(event_d_id, event_p_id, events, nb_deq); in l3fwd_event_worker_cleanup()
|
| H A D | l3fwd_em.c | 717 struct rte_event events[MAX_PKT_BURST]; in em_event_loop_burst() local 734 events, deq_len, 0); in em_event_loop_burst() 750 events[i].queue_id = tx_q_id; in em_event_loop_burst() 751 events[i].op = RTE_EVENT_OP_FORWARD; in em_event_loop_burst() 761 events, nb_deq); in em_event_loop_burst() 764 event_p_id, events + nb_enq, in em_event_loop_burst() 770 event_p_id, events, nb_deq, 0); in em_event_loop_burst() 774 events + nb_enq, in em_event_loop_burst() 843 struct rte_event events[MAX_PKT_BURST]; in em_event_loop_vector() local 867 events[i].queue_id = tx_q_id; in em_event_loop_vector() [all …]
|
| H A D | l3fwd_em_sequential.h | 104 l3fwd_em_process_events(int nb_rx, struct rte_event **events, in l3fwd_em_process_events() argument 109 rte_prefetch0(rte_pktmbuf_mtod(events[0]->mbuf, in l3fwd_em_process_events() 113 struct rte_mbuf *mbuf = events[j]->mbuf; in l3fwd_em_process_events() 117 events[i]->mbuf, in l3fwd_em_process_events()
|
| H A D | l3fwd_em.h | 154 l3fwd_em_no_opt_process_events(int nb_rx, struct rte_event **events, in l3fwd_em_no_opt_process_events() argument 161 rte_prefetch0(rte_pktmbuf_mtod(events[j]->mbuf, void *)); in l3fwd_em_no_opt_process_events() 168 rte_prefetch0(rte_pktmbuf_mtod(events[ in l3fwd_em_no_opt_process_events() 170 l3fwd_em_simple_process(events[j]->mbuf, qconf); in l3fwd_em_no_opt_process_events() 175 l3fwd_em_simple_process(events[j]->mbuf, qconf); in l3fwd_em_no_opt_process_events()
|
| /dpdk/examples/l2fwd-event/ |
| H A D | l2fwd_common.c | 119 l2fwd_event_vector_array_free(struct rte_event events[], uint16_t num) in l2fwd_event_vector_array_free() argument 124 rte_pktmbuf_free_bulk(events[i].vec->mbufs, in l2fwd_event_vector_array_free() 125 events[i].vec->nb_elem); in l2fwd_event_vector_array_free() 126 rte_mempool_put(rte_mempool_from_obj(events[i].vec), in l2fwd_event_vector_array_free() 127 events[i].vec); in l2fwd_event_vector_array_free() 143 struct rte_event events[], uint16_t nb_enq, in l2fwd_event_worker_cleanup() argument 150 l2fwd_event_vector_array_free(events + nb_enq, in l2fwd_event_worker_cleanup() 154 rte_pktmbuf_free(events[i].mbuf); in l2fwd_event_worker_cleanup() 157 events[i].op = RTE_EVENT_OP_RELEASE; in l2fwd_event_worker_cleanup() 158 rte_event_enqueue_burst(event_d_id, port_id, events, nb_deq); in l2fwd_event_worker_cleanup()
|
| /dpdk/lib/eal/windows/ |
| H A D | eal_interrupts.c | 47 OVERLAPPED_ENTRY events[16]; in eal_intr_thread_main() local 52 intr_iocp, events, RTE_DIM(events), &event_count, in eal_intr_thread_main() 69 if (events[i].lpCompletionKey == IOCP_KEY_SHUTDOWN) { in eal_intr_thread_main() 73 eal_intr_process(&events[i]); in eal_intr_thread_main() 241 rte_epoll_wait(int epfd, struct rte_epoll_event *events, in rte_epoll_wait() argument 245 RTE_SET_USED(events); in rte_epoll_wait() 253 rte_epoll_wait_interruptible(int epfd, struct rte_epoll_event *events, in rte_epoll_wait_interruptible() argument 257 RTE_SET_USED(events); in rte_epoll_wait_interruptible()
|
| /dpdk/doc/guides/prog_guide/ |
| H A D | eventdev.rst | 9 with features to schedule events. This is achieved using the PMD architecture 109 device. An event queue contains events of different flows associated with 117 support events of the type that it was created with. 124 events to the eventdev. An example of a PMD of this type is the eventdev 156 events from an eventdev. Ports are linked to queues in order to retrieve events 342 enqueue some events into the system for it to schedule. The events to be 383 struct rte_event events[BATCH_SIZE]; 388 struct rte_mbuf *mbuf = events[i].mbuf; 390 events[i].queue_id++; 422 this might also include any prefetched events. [all …]
|
| H A D | trace_lib.rst | 14 code generate events that are saved on a giant tape: a trace file. 16 analyze the trace events with timestamps and multi-core views. 28 events: user accesses, exceptional conditions (errors and warnings, for 148 captured events in the trace buffer. 150 When the trace buffer is full, new trace events will be discarded. 168 View and analyze the recorded events 172 events. 193 all their events, merging them in chronological order. 196 filtering. Below example grep the events for ``ethdev`` only:: 201 recorded events. Below example count the number of ``ethdev`` events:: [all …]
|
| H A D | profile_app.rst | 14 Intel processors provide performance counters to monitor events. 56 The ARM64 architecture provide performance counters to monitor events. The 58 addition to the standard events, ``perf`` can be used to profile arm64 59 specific PMU (Performance Monitor Unit) events through raw events (``-e`` 63 `ARM64 specific PMU events enumeration <http://infocenter.arm.com/help/index.jsp?topic=/com.arm.doc…
|
| H A D | event_crypto_adapter.rst | 8 programming model with features to schedule events. 41 completions from crypto device and enqueues them as events to the event device. 44 In this mode, events dequeued from the adapter will be treated as new events. 62 events to crypto adapter. If not, application retrieves crypto adapter's event 64 queue to this port and starts enqueuing crypto operations as events to eventdev 65 using ``rte_event_enqueue_burst()``. The adapter then dequeues the events and 67 complete, the adapter enqueues events to the event device. The application can 68 use this mode when ingress packet ordering is needed. In this mode, events 69 dequeued from the adapter will be treated as forwarded events. The application 126 (``RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_OP_FWD``), events with crypto [all …]
|
| /dpdk/drivers/event/sw/ |
| H A D | iq_chunk.h | 15 struct rte_event events[SW_EVS_PER_Q_CHUNK]; member 71 iq->tail->events[iq->tail_idx++] = *ev; in iq_enqueue() 103 return &iq->head->events[iq->head_idx]; in iq_peek() 126 ev[total++] = current->events[index++]; in iq_dequeue_burst() 173 iq->head->events[idx + i] = ev[i]; in iq_put_back() 182 iq->head->events[i] = ev[remaining + i]; in iq_put_back() 190 iq->head->events[iq->head_idx + i] = ev[i]; in iq_put_back()
|
| /dpdk/doc/guides/eventdevs/ |
| H A D | dsw.rst | 8 distributes the task of scheduling events among all the eventdev ports 51 neither attempting to enqueue nor to dequeue events, it must 62 send enqueued events immediately to the destination port, but instead 65 In case no more events are enqueued on a port with buffered events, 66 these events will be sent after the application has performed a number 92 parallel events on the same queue.
|
| /dpdk/lib/eventdev/ |
| H A D | rte_event_ring.h | 92 const struct rte_event *events, in rte_event_ring_enqueue_burst() argument 98 num = rte_ring_enqueue_burst_elem(&r->r, events, in rte_event_ring_enqueue_burst() 128 struct rte_event *events, in rte_event_ring_dequeue_burst() argument 134 num = rte_ring_dequeue_burst_elem(&r->r, events, in rte_event_ring_dequeue_burst()
|
| /dpdk/lib/eal/linux/ |
| H A D | eal_interrupts.c | 977 events[n].data.fd, in eal_intr_process_interrupts() 999 "descriptor %d\n", events[n].data.fd); in eal_intr_process_interrupts() 1073 struct epoll_event events[totalfds]; in eal_intr_handle_interrupts() local 1077 nfds = epoll_wait(pfd, events, totalfds, in eal_intr_handle_interrupts() 1115 .events = EPOLLIN | EPOLLPRI, in eal_intr_thread_main() 1256 struct rte_epoll_event *events) in eal_epoll_process_event() argument 1275 events[count].fd = rev->fd; in eal_epoll_process_event() 1276 events[count].epfd = rev->epfd; in eal_epoll_process_event() 1277 events[count].epdata.event = evs[i].events; in eal_epoll_process_event() 1322 if (!events) { in eal_epoll_wait() [all …]
|
| /dpdk/examples/eventdev_pipeline/ |
| H A D | pipeline_worker_generic.c | 68 struct rte_event events[BATCH_SIZE]; in worker_generic_burst() local 86 nb_rx = rte_event_dequeue_burst(dev_id, port_id, events, in worker_generic_burst() 87 RTE_DIM(events), 0); in worker_generic_burst() 98 if (events[i].queue_id == cdata.qid[0]) in worker_generic_burst() 99 events[i].flow_id = events[i].mbuf->hash.rss in worker_generic_burst() 102 events[i].queue_id = cdata.next_qid[events[i].queue_id]; in worker_generic_burst() 103 events[i].op = RTE_EVENT_OP_FORWARD; in worker_generic_burst() 104 events[i].sched_type = cdata.queue_type; in worker_generic_burst() 108 nb_tx = rte_event_enqueue_burst(dev_id, port_id, events, nb_rx); in worker_generic_burst() 111 events + nb_tx, in worker_generic_burst() [all …]
|
| H A D | pipeline_common.h | 151 worker_cleanup(uint8_t dev_id, uint8_t port_id, struct rte_event events[], in worker_cleanup() argument 161 if (events[i].op == RTE_EVENT_OP_RELEASE) in worker_cleanup() 163 rte_pktmbuf_free(events[i].mbuf); in worker_cleanup() 167 events[i].op = RTE_EVENT_OP_RELEASE; in worker_cleanup() 168 rte_event_enqueue_burst(dev_id, port_id, events, nb_deq); in worker_cleanup()
|
| /dpdk/lib/port/ |
| H A D | rte_port_eventdev.c | 244 events[i].mbuf = pkts[i]; in rte_port_eventdev_writer_tx_bulk() 245 events[i].queue_id = p->queue_id; in rte_port_eventdev_writer_tx_bulk() 246 events[i].sched_type = p->sched_type; in rte_port_eventdev_writer_tx_bulk() 247 events[i].op = p->evt_op; in rte_port_eventdev_writer_tx_bulk() 251 events, n_pkts); in rte_port_eventdev_writer_tx_bulk() 474 struct rte_event events[RTE_PORT_IN_BURST_SIZE_MAX] = {}; in rte_port_eventdev_writer_nodrop_tx_bulk() local 477 events[i].mbuf = pkts[i]; in rte_port_eventdev_writer_nodrop_tx_bulk() 478 events[i].queue_id = p->queue_id; in rte_port_eventdev_writer_nodrop_tx_bulk() 479 events[i].sched_type = p->sched_type; in rte_port_eventdev_writer_nodrop_tx_bulk() 480 events[i].op = p->evt_op; in rte_port_eventdev_writer_nodrop_tx_bulk() [all …]
|
| /dpdk/lib/cmdline/ |
| H A D | cmdline_os_windows.c | 79 DWORD events; in cmdline_poll_char_console() local 81 if (!PeekConsoleInput(handle, &record, 1, &events)) { in cmdline_poll_char_console() 86 if ((events == 0) || !cmdline_is_key_down(&record)) in cmdline_poll_char_console() 149 DWORD events; in cmdline_read_char() local 164 if (!ReadConsoleInput(handle, &record, 1, &events)) { in cmdline_read_char()
|
| /dpdk/drivers/event/dsw/ |
| H A D | dsw_event.c | 1160 const struct rte_event events[], in dsw_event_enqueue_burst_generic() argument 1196 switch (events[i].op) { in dsw_event_enqueue_burst_generic() 1230 const struct rte_event *event = &events[i]; in dsw_event_enqueue_burst_generic() 1253 return dsw_event_enqueue_burst_generic(source_port, events, in dsw_event_enqueue_burst() 1266 return dsw_event_enqueue_burst_generic(source_port, events, in dsw_event_enqueue_new_burst() 1280 return dsw_event_enqueue_burst_generic(source_port, events, in dsw_event_enqueue_forward_burst() 1288 return dsw_event_dequeue_burst(port, events, 1, wait); in dsw_event_dequeue() 1302 struct rte_event *event = &events[i]; in dsw_port_record_seen_events() 1339 rte_memcpy(events, &port->in_buffer[port->in_buffer_start], in dsw_port_dequeue_burst() 1390 dsw_port_record_seen_events(port, events, dequeued); in dsw_event_dequeue_burst() [all …]
|
| H A D | dsw_evdev.h | 262 const struct rte_event events[], 265 const struct rte_event events[], 268 const struct rte_event events[], 272 uint16_t dsw_event_dequeue_burst(void *port, struct rte_event *events,
|
| /dpdk/drivers/event/dlb2/ |
| H A D | dlb2.c | 2976 RTE_ASSERT(events != NULL); in __dlb2_event_enqueue_burst() 3306 struct rte_event *events, in dlb2_process_dequeue_qes() argument 3334 events[num].u64 = qe->data; in dlb2_process_dequeue_qes() 3385 events[0].u64 = qes[0].data; in dlb2_process_dequeue_four_qes() 3386 events[1].u64 = qes[1].data; in dlb2_process_dequeue_four_qes() 3945 &events[num], in dlb2_hw_dequeue_sparse() 3950 &events[num], in dlb2_hw_dequeue_sparse() 3966 &events[num], in dlb2_hw_dequeue_sparse() 4047 struct rte_event *events, in dlb2_hw_dequeue() argument 4093 &events[num], in dlb2_hw_dequeue() [all …]
|
| /dpdk/lib/eal/freebsd/ |
| H A D | eal_interrupts.c | 460 eal_intr_process_interrupts(struct kevent *events, int nfds) in eal_intr_process_interrupts() argument 471 int event_fd = events[n].ident; in eal_intr_process_interrupts() 597 struct kevent events[MAX_INTR_EVENTS]; in eal_intr_thread_main() local 603 nfds = kevent(kq, NULL, 0, events, MAX_INTR_EVENTS, NULL); in eal_intr_thread_main() 618 eal_intr_process_interrupts(events, nfds); in eal_intr_thread_main() 701 rte_epoll_wait(int epfd, struct rte_epoll_event *events, in rte_epoll_wait() argument 705 RTE_SET_USED(events); in rte_epoll_wait() 713 rte_epoll_wait_interruptible(int epfd, struct rte_epoll_event *events, in rte_epoll_wait_interruptible() argument 717 RTE_SET_USED(events); in rte_epoll_wait_interruptible()
|
| /dpdk/lib/eal/include/ |
| H A D | rte_epoll.h | 68 rte_epoll_wait(int epfd, struct rte_epoll_event *events, 90 rte_epoll_wait_interruptible(int epfd, struct rte_epoll_event *events,
|