Lines Matching refs:cdata
40 if (ev.queue_id == cdata.qid[0]) in worker_generic()
42 % cdata.num_fids; in worker_generic()
44 ev.queue_id = cdata.next_qid[ev.queue_id]; in worker_generic()
46 ev.sched_type = cdata.queue_type; in worker_generic()
55 if (!cdata.quiet) in worker_generic()
96 if (events[i].queue_id == cdata.qid[0]) in worker_generic_burst()
98 % cdata.num_fids; in worker_generic_burst()
100 events[i].queue_id = cdata.next_qid[events[i].queue_id]; in worker_generic_burst()
102 events[i].sched_type = cdata.queue_type; in worker_generic_burst()
115 if (!cdata.quiet) in worker_generic_burst()
127 const uint8_t nb_queues = cdata.num_stages + 1; in setup_eventdev_generic()
128 const uint8_t nb_ports = cdata.num_workers; in setup_eventdev_generic()
139 .dequeue_depth = cdata.worker_cq_depth, in setup_eventdev_generic()
144 .schedule_type = cdata.queue_type, in setup_eventdev_generic()
193 for (i = 0; i < cdata.num_stages; i++) { in setup_eventdev_generic()
198 cdata.qid[i] = i; in setup_eventdev_generic()
199 cdata.next_qid[i] = i+1; in setup_eventdev_generic()
201 if (cdata.enable_queue_priorities) { in setup_eventdev_generic()
232 cdata.tx_queue_id = i; in setup_eventdev_generic()
242 for (i = 0; i < cdata.num_workers; i++) { in setup_eventdev_generic()
251 for (s = 0; s < cdata.num_stages; s++) { in setup_eventdev_generic()
389 if (!cdata.num_mbuf) in init_ports()
390 cdata.num_mbuf = 16384 * num_ports; in init_ports()
393 /* mbufs */ cdata.num_mbuf, in init_ports()
419 .dequeue_depth = cdata.worker_cq_depth, in init_adapters()
435 ret = rte_event_eth_rx_adapter_create(cdata.rx_adapter_id, evdev_id, in init_adapters()
439 cdata.rx_adapter_id); in init_adapters()
441 ret = rte_event_eth_tx_adapter_create(cdata.tx_adapter_id, evdev_id, in init_adapters()
445 cdata.tx_adapter_id); in init_adapters()
449 queue_conf.ev.sched_type = cdata.queue_type; in init_adapters()
450 queue_conf.ev.queue_id = cdata.qid[0]; in init_adapters()
453 ret = rte_event_eth_rx_adapter_queue_add(cdata.rx_adapter_id, i, in init_adapters()
459 ret = rte_event_eth_tx_adapter_queue_add(cdata.tx_adapter_id, i, in init_adapters()
466 ret = rte_event_eth_tx_adapter_event_port_get(cdata.tx_adapter_id, in init_adapters()
471 ret = rte_event_port_link(evdev_id, tx_port_id, &cdata.tx_queue_id, in init_adapters()
477 ret = rte_event_eth_rx_adapter_service_id_get(cdata.rx_adapter_id, in init_adapters()
486 ret = rte_event_eth_tx_adapter_service_id_get(cdata.tx_adapter_id, in init_adapters()
495 ret = rte_event_eth_rx_adapter_start(cdata.rx_adapter_id); in init_adapters()
498 cdata.rx_adapter_id); in init_adapters()
500 ret = rte_event_eth_tx_adapter_start(cdata.tx_adapter_id); in init_adapters()
503 cdata.tx_adapter_id); in init_adapters()
522 if (cdata.all_type_queues && !(eventdev_info.event_dev_cap & in generic_opt_check()
538 if (cdata.worker_lcore_mask == 0 || in generic_opt_check()
539 (rx_needed && cdata.rx_lcore_mask == 0) || in generic_opt_check()
540 (cdata.tx_lcore_mask == 0) || in generic_opt_check()
541 (sched_needed && cdata.sched_lcore_mask == 0)) { in generic_opt_check()
547 cdata.rx_lcore_mask, cdata.tx_lcore_mask, in generic_opt_check()
548 cdata.sched_lcore_mask, in generic_opt_check()
549 cdata.worker_lcore_mask); in generic_opt_check()