Home
last modified time | relevance | path

Searched refs:dev_id (Results 1 – 25 of 235) sorted by relevance

12345678910

/dpdk/app/test-bbdev/
H A Dtest_bbdev.c97 for (dev_id = 0; dev_id < RTE_BBDEV_MAX_DEVS; dev_id++) { in test_bbdev_configure_invalid_dev_id()
198 dev_id); in test_bbdev_configure_stop_device()
223 dev_id); in test_bbdev_configure_stop_device()
230 dev_id); in test_bbdev_configure_stop_device()
248 dev_id); in test_bbdev_configure_stop_device()
311 dev_id); in test_bbdev_configure_stop_queue()
344 dev_id); in test_bbdev_configure_stop_queue()
664 dev_id); in test_bbdev_stats()
678 dev_id); in test_bbdev_stats()
682 dev_id); in test_bbdev_stats()
[all …]
/dpdk/lib/rawdev/
H A Drte_rawdev.c61 dev = &rte_rawdevs[dev_id]; in rte_rawdev_socket_id()
102 dev = &rte_rawdevs[dev_id]; in rte_rawdev_configure()
389 dev_id); in rte_rawdev_start()
417 dev_id); in rte_rawdev_stop()
442 dev_id); in rte_rawdev_close()
465 uint16_t dev_id; in rte_rawdev_find_free_device_index() local
467 for (dev_id = 0; dev_id < RTE_RAWDEV_MAX_DEVS; dev_id++) { in rte_rawdev_find_free_device_index()
470 return dev_id; in rte_rawdev_find_free_device_index()
480 uint16_t dev_id; in rte_rawdev_pmd_allocate() local
507 rawdev->dev_id = dev_id; in rte_rawdev_pmd_allocate()
[all …]
H A Drte_rawdev.h62 rte_rawdev_socket_id(uint16_t dev_id);
185 rte_rawdev_queue_setup(uint16_t dev_id,
235 rte_rawdev_start(uint16_t dev_id);
245 rte_rawdev_stop(uint16_t dev_id);
259 rte_rawdev_close(uint16_t dev_id);
273 rte_rawdev_reset(uint16_t dev_id);
289 uint16_t dev_id; member
365 rte_rawdev_get_attr(uint16_t dev_id,
384 rte_rawdev_set_attr(uint16_t dev_id,
505 rte_rawdev_xstats_get(uint16_t dev_id,
[all …]
/dpdk/lib/regexdev/
H A Drte_regexdev.c124 dev->data->dev_id = dev_id; in rte_regexdev_register()
218 dev_id); in rte_regexdev_configure()
228 dev_id); in rte_regexdev_configure()
235 dev_id); in rte_regexdev_configure()
242 dev_id); in rte_regexdev_configure()
247 dev_id); in rte_regexdev_configure()
257 dev_id); in rte_regexdev_configure()
268 dev_id); in rte_regexdev_configure()
280 dev_id); in rte_regexdev_configure()
308 dev_id); in rte_regexdev_queue_pair_setup()
[all …]
H A Drte_regexdev.h214 if (!rte_regexdev_is_valid_dev(dev_id)) { \
220 #define RTE_REGEXDEV_VALID_DEV_ID_OR_RET(dev_id) do { \ argument
221 if (!rte_regexdev_is_valid_dev(dev_id)) { \
241 int rte_regexdev_is_valid_dev(uint16_t dev_id);
824 rte_regexdev_start(uint8_t dev_id);
847 rte_regexdev_stop(uint8_t dev_id);
863 rte_regexdev_close(uint8_t dev_id);
1010 rte_regexdev_rule_db_update(uint8_t dev_id,
1122 rte_regexdev_xstats_names_get(uint8_t dev_id,
1218 rte_regexdev_selftest(uint8_t dev_id);
[all …]
/dpdk/lib/bbdev/
H A Drte_bbdev.c50 dev_id); \
210 bbdev->data->dev_id = dev_id; in rte_bbdev_allocate()
240 dev_id = bbdev->data->dev_id; in rte_bbdev_release()
303 dev_id++; in rte_bbdev_find_next()
304 for (; dev_id < RTE_BBDEV_MAX_DEVS; dev_id++) in rte_bbdev_find_next()
307 return dev_id; in rte_bbdev_find_next()
324 dev_id); in rte_bbdev_setup_queues()
358 dev_id); in rte_bbdev_setup_queues()
384 dev_id); in rte_bbdev_setup_queues()
412 dev_id); in rte_bbdev_intr_enable()
[all …]
H A Drte_bbdev.h63 rte_bbdev_is_valid(uint16_t dev_id);
76 rte_bbdev_find_next(uint16_t dev_id);
122 rte_bbdev_intr_enable(uint16_t dev_id);
166 rte_bbdev_start(uint16_t dev_id);
179 rte_bbdev_stop(uint16_t dev_id);
192 rte_bbdev_close(uint16_t dev_id);
268 rte_bbdev_stats_reset(uint16_t dev_id);
406 uint16_t dev_id; /**< Device ID */ member
478 struct rte_bbdev *dev = &rte_bbdev_devices[dev_id]; in rte_bbdev_enqueue_enc_ops()
508 struct rte_bbdev *dev = &rte_bbdev_devices[dev_id]; in rte_bbdev_enqueue_dec_ops()
[all …]
/dpdk/lib/compressdev/
H A Drte_compressdev.c214 uint8_t dev_id; in rte_compressdev_find_free_device_index() local
216 for (dev_id = 0; dev_id < RTE_COMPRESS_MAX_DEVS; dev_id++) { in rte_compressdev_find_free_device_index()
219 return dev_id; in rte_compressdev_find_free_device_index()
228 uint8_t dev_id; in rte_compressdev_pmd_allocate() local
258 compressdev->data->dev_id = dev_id; in rte_compressdev_pmd_allocate()
441 dev_id, diag); in rte_compressdev_configure()
520 dev_id); in rte_compressdev_close()
674 dev_id, ret); in rte_compressdev_private_xform_create()
697 dev_id, ret); in rte_compressdev_private_xform_free()
722 dev_id, ret); in rte_compressdev_stream_create()
[all …]
H A Drte_compressdev.h58 rte_compressdev_capability_get(uint8_t dev_id,
150 rte_compressdev_name_get(uint8_t dev_id);
194 rte_compressdev_socket_id(uint8_t dev_id);
225 rte_compressdev_configure(uint8_t dev_id,
244 rte_compressdev_start(uint8_t dev_id);
255 rte_compressdev_stop(uint8_t dev_id);
273 rte_compressdev_close(uint8_t dev_id);
313 rte_compressdev_queue_pair_count(uint8_t dev_id);
341 rte_compressdev_stats_reset(uint8_t dev_id);
501 rte_compressdev_stream_create(uint8_t dev_id,
[all …]
/dpdk/lib/eventdev/
H A Drte_eventdev.c345 dev_id); in rte_event_dev_configure()
368 dev_id, in rte_event_dev_configure()
401 dev_id, in rte_event_dev_configure()
422 dev_id); in rte_event_dev_configure()
437 dev_id); in rte_event_dev_configure()
895 dev_id); in rte_event_queue_attr_set()
1292 dev_id); in rte_event_dev_start()
1338 dev_id); in rte_event_dev_stop()
1410 for (dev_id = 0; dev_id < RTE_EVENT_MAX_DEVS; dev_id++) { in eventdev_find_free_device_index()
1454 eventdev->data->dev_id = dev_id; in rte_event_pmd_allocate()
[all …]
H A Drte_eventdev.h987 rte_event_dev_start(uint8_t dev_id);
1008 rte_event_dev_stop(uint8_t dev_id);
1056 rte_event_dev_close(uint8_t dev_id);
1902 fp_ops = &rte_event_fp_ops[dev_id]; in __rte_event_enqueue_burst()
1905 if (dev_id >= RTE_EVENT_MAX_DEVS || in __rte_event_enqueue_burst()
1976 fp_ops = &rte_event_fp_ops[dev_id]; in rte_event_enqueue_burst()
2028 fp_ops = &rte_event_fp_ops[dev_id]; in rte_event_enqueue_new_burst()
2080 fp_ops = &rte_event_fp_ops[dev_id]; in rte_event_enqueue_forward_burst()
2158 fp_ops = &rte_event_fp_ops[dev_id]; in rte_event_dequeue_burst()
2161 if (dev_id >= RTE_EVENT_MAX_DEVS || in rte_event_dequeue_burst()
[all …]
/dpdk/app/test/
H A Dtest_dmadev.c119 await_hw(dev_id, vchan); in do_multi_copies()
203 await_hw(dev_id, vchan); in test_enqueue_copies()
252 await_hw(dev_id, vchan); in test_enqueue_copies()
316 await_hw(dev_id, vchan); in test_failure_in_full_burst()
401 await_hw(dev_id, vchan); in test_individual_status_query_with_failure()
464 await_hw(dev_id, vchan); in test_single_item_status_query_with_failure()
488 await_hw(dev_id, vchan); in test_single_item_status_query_with_failure()
804 dev_id); in test_dmadev_instance()
816 dev_id); in test_dmadev_instance()
827 rte_dma_stop(dev_id); in test_dmadev_instance()
[all …]
/dpdk/lib/gpudev/
H A Dgpudev.c86 if (dev_id >= 0 && dev_id < gpu_max && in rte_gpu_is_valid()
104 dev_id = 0; in rte_gpu_find_next()
108 dev_id++; in rte_gpu_find_next()
120 for (dev_id = 0; dev_id < gpu_max; dev_id++) { in gpu_find_free_id()
226 dev->mpshared->info.dev_id = dev_id; in rte_gpu_allocate()
264 for (dev_id = 0; dev_id < gpu_max; dev_id++) { in rte_gpu_attach()
328 dev_id = dev->mpshared->info.dev_id; in rte_gpu_release()
495 dev_id = dev->mpshared->info.dev_id; in rte_gpu_notify()
754 devflag->dev_id = dev_id; in rte_gpu_comm_create_flag()
914 comm_list[idx_l].dev_id = dev_id; in rte_gpu_comm_create_list()
[all …]
H A Drte_gpudev.h55 int16_t dev_id; member
89 uint16_t dev_id; member
122 uint16_t dev_id; member
180 bool rte_gpu_is_valid(int16_t dev_id);
238 #define RTE_GPU_FOREACH(dev_id) \ argument
239 RTE_GPU_FOREACH_CHILD(dev_id, RTE_GPU_ID_ANY)
250 #define RTE_GPU_FOREACH_PARENT(dev_id) \ argument
266 dev_id >= 0; \
267 dev_id = rte_gpu_find_next(dev_id + 1, parent))
285 int rte_gpu_close(int16_t dev_id);
[all …]
/dpdk/lib/cryptodev/
H A Drte_cryptodev.c845 for (dev_id = 0; dev_id < RTE_CRYPTO_MAX_DEVS; dev_id++) { in rte_cryptodev_find_free_device_index()
889 cryptodev->data->dev_id = dev_id; in rte_cryptodev_pmd_allocate()
921 dev_id = cryptodev->data->dev_id; in rte_cryptodev_pmd_release_device()
1083 dev_id); in rte_cryptodev_start()
1116 dev_id); in rte_cryptodev_stop()
1144 dev_id); in rte_cryptodev_close()
1714 dev_id); in rte_cryptodev_sym_session_init()
1743 dev_id); in rte_cryptodev_sym_session_init()
1805 for (dev_id = 0; dev_id < RTE_CRYPTO_MAX_DEVS; dev_id++) in rte_cryptodev_asym_session_pool_create()
1934 dev_id); in rte_cryptodev_asym_session_create()
[all …]
H A Drte_cryptodev_trace.h24 RTE_TRACE_POINT_ARGS(uint8_t dev_id,
26 rte_trace_point_emit_u8(dev_id);
34 rte_trace_point_emit_u8(dev_id);
40 RTE_TRACE_POINT_ARGS(uint8_t dev_id),
41 rte_trace_point_emit_u8(dev_id);
47 rte_trace_point_emit_u8(dev_id);
55 rte_trace_point_emit_u8(dev_id);
101 rte_trace_point_emit_u8(dev_id);
116 rte_trace_point_emit_u8(dev_id);
125 rte_trace_point_emit_u8(dev_id);
[all …]
/dpdk/lib/dmadev/
H A Drte_dmadev.c56 while (dev_id < dma_devices_max && rte_dma_devices[dev_id].state == RTE_DMA_DEV_UNUSED) in rte_dma_next_dev()
57 dev_id++; in rte_dma_next_dev()
60 return dev_id; in rte_dma_next_dev()
238 int16_t dev_id; in dma_allocate_primary() local
261 if (dev_id < 0) { in dma_allocate_primary()
270 dev->data->dev_id = dev_id; in dma_allocate_primary()
385 return dev->data->dev_id; in rte_dma_get_dev_id_by_name()
391 return (dev_id >= 0) && (dev_id < dma_devices_max) && in rte_dma_is_valid()
449 dev_id); in rte_dma_configure()
580 dev_id); in rte_dma_vchan_setup()
[all …]
H A Drte_dmadev.h207 bool rte_dma_is_valid(int16_t dev_id);
389 int rte_dma_start(int16_t dev_id);
406 int rte_dma_stop(int16_t dev_id);
423 int rte_dma_close(int16_t dev_id);
722 int rte_dma_dump(int16_t dev_id, FILE *f);
861 if (!rte_dma_is_valid(dev_id) || length == 0) in rte_dma_copy()
958 if (!rte_dma_is_valid(dev_id) || length == 0) in rte_dma_fill()
986 rte_dma_submit(int16_t dev_id, uint16_t vchan) in rte_dma_submit() argument
991 if (!rte_dma_is_valid(dev_id)) in rte_dma_submit()
1032 if (!rte_dma_is_valid(dev_id) || nb_cpls == 0) in rte_dma_completed()
[all …]
/dpdk/drivers/raw/ioat/
H A Dioat_rawdev_test.c50 rte_ioat_perform_ops(dev_id); in do_multi_copies()
59 if (rte_ioat_enqueue_copy(dev_id, in do_multi_copies()
70 rte_ioat_perform_ops(dev_id); in do_multi_copies()
127 test_enqueue_copies(int dev_id) in test_enqueue_copies() argument
154 rte_ioat_perform_ops(dev_id); in test_enqueue_copies()
246 test_enqueue_fill(int dev_id) in test_enqueue_fill() argument
267 rte_ioat_perform_ops(dev_id); in test_enqueue_fill()
291 test_burst_capacity(int dev_id) in test_burst_capacity() argument
542 rte_ioat_fence(dev_id); in test_completion_status()
714 rte_rawdev_stop(dev_id); in ioat_rawdev_test()
[all …]
H A Drte_ioat_rawdev_fns.h115 __ioat_burst_capacity(int dev_id) in __ioat_burst_capacity() argument
187 __ioat_fence(int dev_id) in __ioat_fence() argument
206 __ioat_perform_ops(int dev_id) in __ioat_perform_ops() argument
288 rte_ioat_burst_capacity(int dev_id) in rte_ioat_burst_capacity() argument
325 rte_ioat_fence(int dev_id) in rte_ioat_fence() argument
330 return __idxd_fence(dev_id); in rte_ioat_fence()
332 return __ioat_fence(dev_id); in rte_ioat_fence()
336 rte_ioat_perform_ops(int dev_id) in rte_ioat_perform_ops() argument
341 return __idxd_perform_ops(dev_id); in rte_ioat_perform_ops()
343 return __ioat_perform_ops(dev_id); in rte_ioat_perform_ops()
[all …]
/dpdk/drivers/raw/cnxk_bphy/
H A Drte_pmd_bphy.h191 rte_pmd_bphy_intr_init(uint16_t dev_id) in rte_pmd_bphy_intr_init() argument
202 rte_pmd_bphy_intr_fini(uint16_t dev_id) in rte_pmd_bphy_intr_fini() argument
213 rte_pmd_bphy_intr_register(uint16_t dev_id, int irq_num, in rte_pmd_bphy_intr_register() argument
233 rte_pmd_bphy_intr_unregister(uint16_t dev_id, int irq_num) in rte_pmd_bphy_intr_unregister() argument
298 return __rte_pmd_bphy_enq_deq(dev_id, lmac, &msg, NULL, 0); in rte_pmd_bphy_cgx_intlbk_disable()
308 return __rte_pmd_bphy_enq_deq(dev_id, lmac, &msg, NULL, 0); in rte_pmd_bphy_cgx_intlbk_enable()
318 return __rte_pmd_bphy_enq_deq(dev_id, lmac, &msg, NULL, 0); in rte_pmd_bphy_cgx_ptp_rx_disable()
328 return __rte_pmd_bphy_enq_deq(dev_id, lmac, &msg, NULL, 0); in rte_pmd_bphy_cgx_ptp_rx_enable()
358 rte_pmd_bphy_cgx_start_rxtx(uint16_t dev_id, uint16_t lmac) in rte_pmd_bphy_cgx_start_rxtx() argument
368 rte_pmd_bphy_cgx_stop_rxtx(uint16_t dev_id, uint16_t lmac) in rte_pmd_bphy_cgx_stop_rxtx() argument
[all …]
H A Dcnxk_bphy_cgx_test.c38 cnxk_bphy_cgx_dev_selftest(uint16_t dev_id) in cnxk_bphy_cgx_dev_selftest() argument
43 queues = rte_rawdev_queue_count(dev_id); in cnxk_bphy_cgx_dev_selftest()
47 ret = rte_rawdev_start(dev_id); in cnxk_bphy_cgx_dev_selftest()
55 ret = rte_rawdev_queue_conf_get(dev_id, i, &descs, in cnxk_bphy_cgx_dev_selftest()
67 ret = rte_pmd_bphy_cgx_stop_rxtx(dev_id, i); in cnxk_bphy_cgx_dev_selftest()
73 ret = rte_pmd_bphy_cgx_start_rxtx(dev_id, i); in cnxk_bphy_cgx_dev_selftest()
85 ret = cnxk_bphy_cgx_link_cond(dev_id, i, 0); in cnxk_bphy_cgx_dev_selftest()
96 ret = cnxk_bphy_cgx_link_cond(dev_id, i, 1); in cnxk_bphy_cgx_dev_selftest()
134 ret = rte_pmd_bphy_cgx_set_fec(dev_id, i, fec); in cnxk_bphy_cgx_dev_selftest()
141 ret = rte_pmd_bphy_cgx_set_fec(dev_id, i, fec); in cnxk_bphy_cgx_dev_selftest()
[all …]
H A Dcnxk_bphy_irq.c14 cnxk_bphy_get_bphy_dev_by_dev_id(uint16_t dev_id) in cnxk_bphy_get_bphy_dev_by_dev_id() argument
18 if (!rte_rawdev_pmd_is_valid_dev(dev_id)) in cnxk_bphy_get_bphy_dev_by_dev_id()
21 rawdev = &rte_rawdevs[dev_id]; in cnxk_bphy_get_bphy_dev_by_dev_id()
27 cnxk_bphy_irq_max_get(uint16_t dev_id) in cnxk_bphy_irq_max_get() argument
32 bphy_dev = cnxk_bphy_get_bphy_dev_by_dev_id(dev_id); in cnxk_bphy_irq_max_get()
39 cnxk_bphy_intr_init(uint16_t dev_id) in cnxk_bphy_intr_init() argument
41 struct bphy_device *bphy_dev = cnxk_bphy_get_bphy_dev_by_dev_id(dev_id); in cnxk_bphy_intr_init()
51 cnxk_bphy_intr_fini(uint16_t dev_id) in cnxk_bphy_intr_fini() argument
61 cnxk_bphy_intr_register(uint16_t dev_id, int irq_num, in cnxk_bphy_intr_register() argument
83 cnxk_bphy_intr_unregister(uint16_t dev_id, int irq_num) in cnxk_bphy_intr_unregister() argument
[all …]
/dpdk/drivers/raw/cnxk_gpio/
H A Drte_pmd_cnxk_gpio.h164 return __rte_pmd_gpio_enq_deq(dev_id, gpio, &msg, NULL, 0); in rte_pmd_gpio_set_pin_value()
181 rte_pmd_gpio_set_pin_edge(uint16_t dev_id, int gpio, in rte_pmd_gpio_set_pin_edge() argument
189 return __rte_pmd_gpio_enq_deq(dev_id, gpio, &msg, NULL, 0); in rte_pmd_gpio_set_pin_edge()
213 return __rte_pmd_gpio_enq_deq(dev_id, gpio, &msg, NULL, 0); in rte_pmd_gpio_set_pin_dir()
243 return __rte_pmd_gpio_enq_deq(dev_id, gpio, &msg, NULL, 0); in rte_pmd_gpio_set_pin_active_low()
283 rte_pmd_gpio_get_pin_edge(uint16_t dev_id, int gpio, in rte_pmd_gpio_get_pin_edge() argument
371 return __rte_pmd_gpio_enq_deq(dev_id, gpio, &msg, NULL, 0); in rte_pmd_gpio_register_irq()
386 rte_pmd_gpio_unregister_irq(uint16_t dev_id, int gpio) in rte_pmd_gpio_unregister_irq() argument
410 rte_pmd_gpio_enable_interrupt(uint16_t dev_id, int gpio, in rte_pmd_gpio_enable_interrupt() argument
413 return rte_pmd_gpio_set_pin_edge(dev_id, gpio, edge); in rte_pmd_gpio_enable_interrupt()
[all …]
H A Dcnxk_gpio_selftest.c106 ret = rte_pmd_gpio_set_pin_edge(dev_id, gpio, in cnxk_gpio_test_input()
112 ret = rte_pmd_gpio_set_pin_edge(dev_id, gpio, in cnxk_gpio_test_input()
155 cnxk_gpio_test_irq(uint16_t dev_id, int gpio) in cnxk_gpio_test_irq() argument
183 rte_pmd_gpio_disable_interrupt(dev_id, gpio); in cnxk_gpio_test_irq()
184 rte_pmd_gpio_unregister_irq(dev_id, gpio); in cnxk_gpio_test_irq()
243 ret = rte_pmd_gpio_set_pin_edge(dev_id, gpio, in cnxk_gpio_test_output()
303 cnxk_gpio_selftest(uint16_t dev_id) in cnxk_gpio_selftest() argument
318 queues = rte_rawdev_queue_count(dev_id); in cnxk_gpio_selftest()
322 ret = rte_rawdev_start(dev_id); in cnxk_gpio_selftest()
372 ret = rte_rawdev_queue_release(dev_id, i); in cnxk_gpio_selftest()
[all …]

12345678910