| /dpdk/drivers/event/dpaa2/ |
| H A D | dpaa2_eventdev_selftest.c | 43 uint8_t event_type; member 219 uint32_t flow_id, uint8_t event_type, in update_event_and_validation_attr() argument 228 attr->event_type = event_type; in update_event_and_validation_attr() 237 ev->event_type = event_type; in update_event_and_validation_attr() 246 inject_events(uint32_t flow_id, uint8_t event_type, uint8_t sub_event_type, in inject_events() argument 259 update_event_and_validation_attr(m, &ev, flow_id, event_type, in inject_events() 323 RTE_TEST_ASSERT_EQUAL(attr->event_type, ev->event_type, in validate_event() 325 attr->event_type, ev->event_type); in validate_event() 588 if (event.event_type == RTE_EVENT_TYPE_CPU) in flush()
|
| /dpdk/drivers/event/octeontx/ |
| H A D | ssovf_evdev_selftest.c | 41 uint8_t event_type; member 264 uint32_t flow_id, uint8_t event_type, in update_event_and_validation_attr() argument 273 attr->event_type = event_type; in update_event_and_validation_attr() 281 ev->event_type = event_type; in update_event_and_validation_attr() 368 RTE_TEST_ASSERT_EQUAL(attr->event_type, ev->event_type, in validate_event() 370 attr->event_type, ev->event_type); in validate_event() 903 ev.event_type = RTE_EVENT_TYPE_CPU; in worker_flow_based_pipeline() 1058 ev.event_type = RTE_EVENT_TYPE_CPU; in worker_group_based_pipeline() 1215 ev.event_type = RTE_EVENT_TYPE_CPU; in worker_flow_based_pipeline_max_stages_rand_sched_type() 1290 ev.event_type = RTE_EVENT_TYPE_CPU; in worker_queue_based_pipeline_max_stages_rand_sched_type() [all …]
|
| H A D | ssovf_worker.h | 182 if (ev->event_type == RTE_EVENT_TYPE_ETHDEV) in ssows_get_work() 186 else if (ev->event_type == RTE_EVENT_TYPE_CRYPTODEV) in ssows_get_work()
|
| /dpdk/drivers/event/cnxk/ |
| H A D | cnxk_eventdev_selftest.c | 40 uint8_t event_type; member 254 attr->event_type = event_type; in update_event_and_validation_attr() 262 ev->event_type = event_type; in update_event_and_validation_attr() 349 RTE_TEST_ASSERT_EQUAL(attr->event_type, ev->event_type, in validate_event() 351 attr->event_type, ev->event_type); in validate_event() 908 ev.event_type = RTE_EVENT_TYPE_CPU; in worker_flow_based_pipeline() 1064 ev.event_type = RTE_EVENT_TYPE_CPU; in worker_group_based_pipeline() 1221 ev.event_type = RTE_EVENT_TYPE_CPU; in worker_flow_based_pipeline_max_stages_rand_sched_type() 1297 ev.event_type = RTE_EVENT_TYPE_CPU; in worker_queue_based_pipeline_max_stages_rand_sched_type() 1342 ev.event_type = RTE_EVENT_TYPE_CPU; in worker_mixed_pipeline_max_stages_rand_sched_type() [all …]
|
| H A D | cnxk_eventdev_adptr.c | 10 uint32_t event_type) in cnxk_sso_updt_xae_cnt() argument 14 switch (event_type) { in cnxk_sso_updt_xae_cnt()
|
| H A D | cnxk_eventdev.h | 233 uint32_t event_type);
|
| /dpdk/app/test/ |
| H A D | test_event_timer_adapter.c | 443 .ev.event_type = RTE_EVENT_TYPE_TIMER, in test_timer_state() 500 .ev.event_type = RTE_EVENT_TYPE_TIMER, in _arm_timers() 614 .ev.event_type = RTE_EVENT_TYPE_TIMER, in _arm_timers_burst() 697 .ev.event_type = RTE_EVENT_TYPE_TIMER, in test_timer_cancel_periodic() 739 .ev.event_type = RTE_EVENT_TYPE_TIMER, in test_timer_cancel() 781 .ev.event_type = RTE_EVENT_TYPE_TIMER, in _cancel_producer() 820 .ev.event_type = RTE_EVENT_TYPE_TIMER, in _cancel_producer_burst() 1001 .ev.event_type = RTE_EVENT_TYPE_TIMER, in test_timer_cancel_random() 1187 .ev.event_type = RTE_EVENT_TYPE_TIMER, in stat_inc_reset_ev_enq() 1280 .ev.event_type = RTE_EVENT_TYPE_TIMER, in event_timer_arm() [all …]
|
| /dpdk/drivers/net/pfe/ |
| H A D | pfe_hif_lib.c | 80 hif_lib_indicate_client(struct hif_client_s *client, int event_type, in hif_lib_indicate_client() argument 83 if (!client || event_type >= HIF_EVENT_MAX || in hif_lib_indicate_client() 87 if (!test_and_set_bit(qno, &client->queue_mask[event_type])) in hif_lib_indicate_client() 88 client->event_handler(client->priv, event_type, qno); in hif_lib_indicate_client() 243 __rte_unused int event_type, __rte_unused int qno) in hif_lib_event_dummy() argument
|
| /dpdk/app/test-eventdev/ |
| H A D | test_perf_atq.c | 34 ev->event_type = RTE_EVENT_TYPE_CPU; in atq_fwd_event() 53 (ev.event_type == RTE_EVENT_TYPE_CRYPTODEV)) { in perf_atq_worker() 113 (ev[i].event_type == RTE_EVENT_TYPE_CRYPTODEV)) { in perf_atq_worker_burst()
|
| H A D | test_perf_queue.c | 36 ev->event_type = RTE_EVENT_TYPE_CPU; in fwd_event() 55 (ev.event_type == RTE_EVENT_TYPE_CRYPTODEV)) { in perf_queue_worker() 115 (ev[i].event_type == RTE_EVENT_TYPE_CRYPTODEV)) { in perf_queue_worker_burst()
|
| H A D | test_pipeline_common.h | 99 ev->event_type = RTE_EVENT_TYPE_CPU; in pipeline_fwd_event() 107 ev->event_type = RTE_EVENT_TYPE_CPU_VECTOR; in pipeline_fwd_event_vector()
|
| H A D | test_perf_common.c | 58 ev.event_type = RTE_EVENT_TYPE_CPU; in perf_producer() 117 ev[i].event_type = RTE_EVENT_TYPE_CPU; in perf_producer_burst() 173 tim.ev.event_type = RTE_EVENT_TYPE_TIMER; in perf_event_timer_producer() 239 tim.ev.event_type = RTE_EVENT_TYPE_TIMER; in perf_event_timer_producer_burst() 353 ev.event_type = RTE_EVENT_TYPE_CPU; in crypto_adapter_enq_op_fwd()
|
| H A D | test_order_atq.c | 18 ev->event_type = RTE_EVENT_TYPE_CPU; in order_atq_process_stage_0()
|
| H A D | test_order_queue.c | 18 ev->event_type = RTE_EVENT_TYPE_CPU; in order_queue_process_stage_0()
|
| H A D | test_pipeline_common.c | 525 if (ev.event_type & RTE_EVENT_TYPE_VECTOR) in pipeline_event_port_flush() 544 if (ev[i].event_type & RTE_EVENT_TYPE_VECTOR) in pipeline_worker_cleanup()
|
| /dpdk/drivers/net/mlx4/ |
| H A D | mlx4_intr.c | 198 switch (event.event_type) { in mlx4_interrupt_handler() 210 event.event_type, event.element.port_num); in mlx4_interrupt_handler()
|
| H A D | mlx4_mr.h | 79 void mlx4_mr_mem_event_cb(enum rte_mem_event event_type, const void *addr,
|
| /dpdk/drivers/event/dlb2/ |
| H A D | dlb2.c | 2764 ev[0].event_type, in dlb2_event_build_hcws() 2768 ev[1].event_type, in dlb2_event_build_hcws() 2772 ev[2].event_type, in dlb2_event_build_hcws() 2776 ev[3].event_type, in dlb2_event_build_hcws() 2819 qe[i].u.event_type.major = ev[i].event_type; in dlb2_event_build_hcws() 2820 qe[i].u.event_type.sub = ev[i].sub_event_type; in dlb2_event_build_hcws() 3337 events[num].event_type = qe->u.event_type.major; in dlb2_process_dequeue_qes() 3338 events[num].sub_event_type = qe->u.event_type.sub; in dlb2_process_dequeue_qes() 3478 qes[0].u.event_type.major << DLB_EVENT_EV_TYPE_SHIFT | in dlb2_process_dequeue_four_qes() 3483 qes[1].u.event_type.major << DLB_EVENT_EV_TYPE_SHIFT | in dlb2_process_dequeue_four_qes() [all …]
|
| /dpdk/lib/eal/include/ |
| H A D | rte_memory.h | 616 typedef void (*rte_mem_event_callback_t)(enum rte_mem_event event_type,
|
| /dpdk/examples/ipsec-secgw/ |
| H A D | ipsec_worker.c | 812 switch (ev.event_type) { in ipsec_wrkr_non_burst_int_port_drv_mode() 821 ev.event_type); in ipsec_wrkr_non_burst_int_port_drv_mode() 945 switch (ev.event_type) { in ipsec_wrkr_non_burst_int_port_app_mode() 954 ev.event_type); in ipsec_wrkr_non_burst_int_port_app_mode()
|
| /dpdk/drivers/net/mlx5/linux/ |
| H A D | mlx5_ethdev_os.c | 836 if (!tmp && event.event_type == IBV_EVENT_DEVICE_FATAL) { in mlx5_dev_interrupt_handler() 852 event.event_type); in mlx5_dev_interrupt_handler() 861 event.event_type, tmp); in mlx5_dev_interrupt_handler() 870 event.event_type, tmp); in mlx5_dev_interrupt_handler() 879 dev->data->port_id, event.event_type); in mlx5_dev_interrupt_handler()
|
| /dpdk/examples/l2fwd-event/ |
| H A D | l2fwd_common.c | 135 if (ev.event_type & RTE_EVENT_TYPE_VECTOR) in l2fwd_event_port_flush()
|
| /dpdk/doc/guides/prog_guide/ |
| H A D | event_timer_adapter.rst | 58 * ``event_type`` - Will be set to ``RTE_EVENT_TYPE_TIMER`` by the event timer 222 conn->evtim->ev.event_type = RTE_EVENT_TYPE_TIMER; 287 switch(ev.event_type) {
|
| H A D | eventdev.rst | 45 * ``event_type`` - The source of this event, e.g. RTE_EVENT_TYPE_ETHDEV or CPU. 47 the same event_type (see above) 355 ev[i].event_type = RTE_EVENT_TYPE_ETHDEV;
|
| /dpdk/examples/l3fwd/ |
| H A D | l3fwd_event.c | 308 if (ev.event_type & RTE_EVENT_TYPE_VECTOR) in l3fwd_event_port_flush()
|