| /dpdk/app/test-bbdev/ |
| H A D | test_bbdev.c | 97 for (dev_id = 0; dev_id < RTE_BBDEV_MAX_DEVS; dev_id++) { in test_bbdev_configure_invalid_dev_id() 198 dev_id); in test_bbdev_configure_stop_device() 223 dev_id); in test_bbdev_configure_stop_device() 230 dev_id); in test_bbdev_configure_stop_device() 248 dev_id); in test_bbdev_configure_stop_device() 311 dev_id); in test_bbdev_configure_stop_queue() 344 dev_id); in test_bbdev_configure_stop_queue() 664 dev_id); in test_bbdev_stats() 678 dev_id); in test_bbdev_stats() 682 dev_id); in test_bbdev_stats() [all …]
|
| /dpdk/lib/rawdev/ |
| H A D | rte_rawdev.c | 61 dev = &rte_rawdevs[dev_id]; in rte_rawdev_socket_id() 102 dev = &rte_rawdevs[dev_id]; in rte_rawdev_configure() 389 dev_id); in rte_rawdev_start() 417 dev_id); in rte_rawdev_stop() 442 dev_id); in rte_rawdev_close() 465 uint16_t dev_id; in rte_rawdev_find_free_device_index() local 467 for (dev_id = 0; dev_id < RTE_RAWDEV_MAX_DEVS; dev_id++) { in rte_rawdev_find_free_device_index() 470 return dev_id; in rte_rawdev_find_free_device_index() 480 uint16_t dev_id; in rte_rawdev_pmd_allocate() local 507 rawdev->dev_id = dev_id; in rte_rawdev_pmd_allocate() [all …]
|
| H A D | rte_rawdev.h | 62 rte_rawdev_socket_id(uint16_t dev_id); 185 rte_rawdev_queue_setup(uint16_t dev_id, 235 rte_rawdev_start(uint16_t dev_id); 245 rte_rawdev_stop(uint16_t dev_id); 259 rte_rawdev_close(uint16_t dev_id); 273 rte_rawdev_reset(uint16_t dev_id); 289 uint16_t dev_id; member 365 rte_rawdev_get_attr(uint16_t dev_id, 384 rte_rawdev_set_attr(uint16_t dev_id, 505 rte_rawdev_xstats_get(uint16_t dev_id, [all …]
|
| /dpdk/lib/regexdev/ |
| H A D | rte_regexdev.c | 124 dev->data->dev_id = dev_id; in rte_regexdev_register() 218 dev_id); in rte_regexdev_configure() 228 dev_id); in rte_regexdev_configure() 235 dev_id); in rte_regexdev_configure() 242 dev_id); in rte_regexdev_configure() 247 dev_id); in rte_regexdev_configure() 257 dev_id); in rte_regexdev_configure() 268 dev_id); in rte_regexdev_configure() 280 dev_id); in rte_regexdev_configure() 308 dev_id); in rte_regexdev_queue_pair_setup() [all …]
|
| H A D | rte_regexdev.h | 214 if (!rte_regexdev_is_valid_dev(dev_id)) { \ 220 #define RTE_REGEXDEV_VALID_DEV_ID_OR_RET(dev_id) do { \ argument 221 if (!rte_regexdev_is_valid_dev(dev_id)) { \ 241 int rte_regexdev_is_valid_dev(uint16_t dev_id); 824 rte_regexdev_start(uint8_t dev_id); 847 rte_regexdev_stop(uint8_t dev_id); 863 rte_regexdev_close(uint8_t dev_id); 1010 rte_regexdev_rule_db_update(uint8_t dev_id, 1122 rte_regexdev_xstats_names_get(uint8_t dev_id, 1218 rte_regexdev_selftest(uint8_t dev_id); [all …]
|
| /dpdk/lib/bbdev/ |
| H A D | rte_bbdev.c | 50 dev_id); \ 210 bbdev->data->dev_id = dev_id; in rte_bbdev_allocate() 240 dev_id = bbdev->data->dev_id; in rte_bbdev_release() 303 dev_id++; in rte_bbdev_find_next() 304 for (; dev_id < RTE_BBDEV_MAX_DEVS; dev_id++) in rte_bbdev_find_next() 307 return dev_id; in rte_bbdev_find_next() 324 dev_id); in rte_bbdev_setup_queues() 358 dev_id); in rte_bbdev_setup_queues() 384 dev_id); in rte_bbdev_setup_queues() 412 dev_id); in rte_bbdev_intr_enable() [all …]
|
| H A D | rte_bbdev.h | 63 rte_bbdev_is_valid(uint16_t dev_id); 76 rte_bbdev_find_next(uint16_t dev_id); 122 rte_bbdev_intr_enable(uint16_t dev_id); 166 rte_bbdev_start(uint16_t dev_id); 179 rte_bbdev_stop(uint16_t dev_id); 192 rte_bbdev_close(uint16_t dev_id); 268 rte_bbdev_stats_reset(uint16_t dev_id); 406 uint16_t dev_id; /**< Device ID */ member 478 struct rte_bbdev *dev = &rte_bbdev_devices[dev_id]; in rte_bbdev_enqueue_enc_ops() 508 struct rte_bbdev *dev = &rte_bbdev_devices[dev_id]; in rte_bbdev_enqueue_dec_ops() [all …]
|
| /dpdk/lib/compressdev/ |
| H A D | rte_compressdev.c | 214 uint8_t dev_id; in rte_compressdev_find_free_device_index() local 216 for (dev_id = 0; dev_id < RTE_COMPRESS_MAX_DEVS; dev_id++) { in rte_compressdev_find_free_device_index() 219 return dev_id; in rte_compressdev_find_free_device_index() 228 uint8_t dev_id; in rte_compressdev_pmd_allocate() local 258 compressdev->data->dev_id = dev_id; in rte_compressdev_pmd_allocate() 441 dev_id, diag); in rte_compressdev_configure() 520 dev_id); in rte_compressdev_close() 674 dev_id, ret); in rte_compressdev_private_xform_create() 697 dev_id, ret); in rte_compressdev_private_xform_free() 722 dev_id, ret); in rte_compressdev_stream_create() [all …]
|
| H A D | rte_compressdev.h | 58 rte_compressdev_capability_get(uint8_t dev_id, 150 rte_compressdev_name_get(uint8_t dev_id); 194 rte_compressdev_socket_id(uint8_t dev_id); 225 rte_compressdev_configure(uint8_t dev_id, 244 rte_compressdev_start(uint8_t dev_id); 255 rte_compressdev_stop(uint8_t dev_id); 273 rte_compressdev_close(uint8_t dev_id); 313 rte_compressdev_queue_pair_count(uint8_t dev_id); 341 rte_compressdev_stats_reset(uint8_t dev_id); 501 rte_compressdev_stream_create(uint8_t dev_id, [all …]
|
| /dpdk/lib/eventdev/ |
| H A D | rte_eventdev.c | 345 dev_id); in rte_event_dev_configure() 368 dev_id, in rte_event_dev_configure() 401 dev_id, in rte_event_dev_configure() 422 dev_id); in rte_event_dev_configure() 437 dev_id); in rte_event_dev_configure() 895 dev_id); in rte_event_queue_attr_set() 1292 dev_id); in rte_event_dev_start() 1338 dev_id); in rte_event_dev_stop() 1410 for (dev_id = 0; dev_id < RTE_EVENT_MAX_DEVS; dev_id++) { in eventdev_find_free_device_index() 1454 eventdev->data->dev_id = dev_id; in rte_event_pmd_allocate() [all …]
|
| H A D | rte_eventdev.h | 987 rte_event_dev_start(uint8_t dev_id); 1008 rte_event_dev_stop(uint8_t dev_id); 1056 rte_event_dev_close(uint8_t dev_id); 1902 fp_ops = &rte_event_fp_ops[dev_id]; in __rte_event_enqueue_burst() 1905 if (dev_id >= RTE_EVENT_MAX_DEVS || in __rte_event_enqueue_burst() 1976 fp_ops = &rte_event_fp_ops[dev_id]; in rte_event_enqueue_burst() 2028 fp_ops = &rte_event_fp_ops[dev_id]; in rte_event_enqueue_new_burst() 2080 fp_ops = &rte_event_fp_ops[dev_id]; in rte_event_enqueue_forward_burst() 2158 fp_ops = &rte_event_fp_ops[dev_id]; in rte_event_dequeue_burst() 2161 if (dev_id >= RTE_EVENT_MAX_DEVS || in rte_event_dequeue_burst() [all …]
|
| /dpdk/app/test/ |
| H A D | test_dmadev.c | 119 await_hw(dev_id, vchan); in do_multi_copies() 203 await_hw(dev_id, vchan); in test_enqueue_copies() 252 await_hw(dev_id, vchan); in test_enqueue_copies() 316 await_hw(dev_id, vchan); in test_failure_in_full_burst() 401 await_hw(dev_id, vchan); in test_individual_status_query_with_failure() 464 await_hw(dev_id, vchan); in test_single_item_status_query_with_failure() 488 await_hw(dev_id, vchan); in test_single_item_status_query_with_failure() 804 dev_id); in test_dmadev_instance() 816 dev_id); in test_dmadev_instance() 827 rte_dma_stop(dev_id); in test_dmadev_instance() [all …]
|
| /dpdk/lib/gpudev/ |
| H A D | gpudev.c | 86 if (dev_id >= 0 && dev_id < gpu_max && in rte_gpu_is_valid() 104 dev_id = 0; in rte_gpu_find_next() 108 dev_id++; in rte_gpu_find_next() 120 for (dev_id = 0; dev_id < gpu_max; dev_id++) { in gpu_find_free_id() 226 dev->mpshared->info.dev_id = dev_id; in rte_gpu_allocate() 264 for (dev_id = 0; dev_id < gpu_max; dev_id++) { in rte_gpu_attach() 328 dev_id = dev->mpshared->info.dev_id; in rte_gpu_release() 495 dev_id = dev->mpshared->info.dev_id; in rte_gpu_notify() 754 devflag->dev_id = dev_id; in rte_gpu_comm_create_flag() 914 comm_list[idx_l].dev_id = dev_id; in rte_gpu_comm_create_list() [all …]
|
| H A D | rte_gpudev.h | 55 int16_t dev_id; member 89 uint16_t dev_id; member 122 uint16_t dev_id; member 180 bool rte_gpu_is_valid(int16_t dev_id); 238 #define RTE_GPU_FOREACH(dev_id) \ argument 239 RTE_GPU_FOREACH_CHILD(dev_id, RTE_GPU_ID_ANY) 250 #define RTE_GPU_FOREACH_PARENT(dev_id) \ argument 266 dev_id >= 0; \ 267 dev_id = rte_gpu_find_next(dev_id + 1, parent)) 285 int rte_gpu_close(int16_t dev_id); [all …]
|
| /dpdk/lib/cryptodev/ |
| H A D | rte_cryptodev.c | 845 for (dev_id = 0; dev_id < RTE_CRYPTO_MAX_DEVS; dev_id++) { in rte_cryptodev_find_free_device_index() 889 cryptodev->data->dev_id = dev_id; in rte_cryptodev_pmd_allocate() 921 dev_id = cryptodev->data->dev_id; in rte_cryptodev_pmd_release_device() 1083 dev_id); in rte_cryptodev_start() 1116 dev_id); in rte_cryptodev_stop() 1144 dev_id); in rte_cryptodev_close() 1714 dev_id); in rte_cryptodev_sym_session_init() 1743 dev_id); in rte_cryptodev_sym_session_init() 1805 for (dev_id = 0; dev_id < RTE_CRYPTO_MAX_DEVS; dev_id++) in rte_cryptodev_asym_session_pool_create() 1934 dev_id); in rte_cryptodev_asym_session_create() [all …]
|
| H A D | rte_cryptodev_trace.h | 24 RTE_TRACE_POINT_ARGS(uint8_t dev_id, 26 rte_trace_point_emit_u8(dev_id); 34 rte_trace_point_emit_u8(dev_id); 40 RTE_TRACE_POINT_ARGS(uint8_t dev_id), 41 rte_trace_point_emit_u8(dev_id); 47 rte_trace_point_emit_u8(dev_id); 55 rte_trace_point_emit_u8(dev_id); 101 rte_trace_point_emit_u8(dev_id); 116 rte_trace_point_emit_u8(dev_id); 125 rte_trace_point_emit_u8(dev_id); [all …]
|
| /dpdk/lib/dmadev/ |
| H A D | rte_dmadev.c | 56 while (dev_id < dma_devices_max && rte_dma_devices[dev_id].state == RTE_DMA_DEV_UNUSED) in rte_dma_next_dev() 57 dev_id++; in rte_dma_next_dev() 60 return dev_id; in rte_dma_next_dev() 238 int16_t dev_id; in dma_allocate_primary() local 261 if (dev_id < 0) { in dma_allocate_primary() 270 dev->data->dev_id = dev_id; in dma_allocate_primary() 385 return dev->data->dev_id; in rte_dma_get_dev_id_by_name() 391 return (dev_id >= 0) && (dev_id < dma_devices_max) && in rte_dma_is_valid() 449 dev_id); in rte_dma_configure() 580 dev_id); in rte_dma_vchan_setup() [all …]
|
| H A D | rte_dmadev.h | 207 bool rte_dma_is_valid(int16_t dev_id); 389 int rte_dma_start(int16_t dev_id); 406 int rte_dma_stop(int16_t dev_id); 423 int rte_dma_close(int16_t dev_id); 722 int rte_dma_dump(int16_t dev_id, FILE *f); 861 if (!rte_dma_is_valid(dev_id) || length == 0) in rte_dma_copy() 958 if (!rte_dma_is_valid(dev_id) || length == 0) in rte_dma_fill() 986 rte_dma_submit(int16_t dev_id, uint16_t vchan) in rte_dma_submit() argument 991 if (!rte_dma_is_valid(dev_id)) in rte_dma_submit() 1032 if (!rte_dma_is_valid(dev_id) || nb_cpls == 0) in rte_dma_completed() [all …]
|
| /dpdk/drivers/raw/ioat/ |
| H A D | ioat_rawdev_test.c | 50 rte_ioat_perform_ops(dev_id); in do_multi_copies() 59 if (rte_ioat_enqueue_copy(dev_id, in do_multi_copies() 70 rte_ioat_perform_ops(dev_id); in do_multi_copies() 127 test_enqueue_copies(int dev_id) in test_enqueue_copies() argument 154 rte_ioat_perform_ops(dev_id); in test_enqueue_copies() 246 test_enqueue_fill(int dev_id) in test_enqueue_fill() argument 267 rte_ioat_perform_ops(dev_id); in test_enqueue_fill() 291 test_burst_capacity(int dev_id) in test_burst_capacity() argument 542 rte_ioat_fence(dev_id); in test_completion_status() 714 rte_rawdev_stop(dev_id); in ioat_rawdev_test() [all …]
|
| H A D | rte_ioat_rawdev_fns.h | 115 __ioat_burst_capacity(int dev_id) in __ioat_burst_capacity() argument 187 __ioat_fence(int dev_id) in __ioat_fence() argument 206 __ioat_perform_ops(int dev_id) in __ioat_perform_ops() argument 288 rte_ioat_burst_capacity(int dev_id) in rte_ioat_burst_capacity() argument 325 rte_ioat_fence(int dev_id) in rte_ioat_fence() argument 330 return __idxd_fence(dev_id); in rte_ioat_fence() 332 return __ioat_fence(dev_id); in rte_ioat_fence() 336 rte_ioat_perform_ops(int dev_id) in rte_ioat_perform_ops() argument 341 return __idxd_perform_ops(dev_id); in rte_ioat_perform_ops() 343 return __ioat_perform_ops(dev_id); in rte_ioat_perform_ops() [all …]
|
| /dpdk/drivers/raw/cnxk_bphy/ |
| H A D | rte_pmd_bphy.h | 191 rte_pmd_bphy_intr_init(uint16_t dev_id) in rte_pmd_bphy_intr_init() argument 202 rte_pmd_bphy_intr_fini(uint16_t dev_id) in rte_pmd_bphy_intr_fini() argument 213 rte_pmd_bphy_intr_register(uint16_t dev_id, int irq_num, in rte_pmd_bphy_intr_register() argument 233 rte_pmd_bphy_intr_unregister(uint16_t dev_id, int irq_num) in rte_pmd_bphy_intr_unregister() argument 298 return __rte_pmd_bphy_enq_deq(dev_id, lmac, &msg, NULL, 0); in rte_pmd_bphy_cgx_intlbk_disable() 308 return __rte_pmd_bphy_enq_deq(dev_id, lmac, &msg, NULL, 0); in rte_pmd_bphy_cgx_intlbk_enable() 318 return __rte_pmd_bphy_enq_deq(dev_id, lmac, &msg, NULL, 0); in rte_pmd_bphy_cgx_ptp_rx_disable() 328 return __rte_pmd_bphy_enq_deq(dev_id, lmac, &msg, NULL, 0); in rte_pmd_bphy_cgx_ptp_rx_enable() 358 rte_pmd_bphy_cgx_start_rxtx(uint16_t dev_id, uint16_t lmac) in rte_pmd_bphy_cgx_start_rxtx() argument 368 rte_pmd_bphy_cgx_stop_rxtx(uint16_t dev_id, uint16_t lmac) in rte_pmd_bphy_cgx_stop_rxtx() argument [all …]
|
| H A D | cnxk_bphy_cgx_test.c | 38 cnxk_bphy_cgx_dev_selftest(uint16_t dev_id) in cnxk_bphy_cgx_dev_selftest() argument 43 queues = rte_rawdev_queue_count(dev_id); in cnxk_bphy_cgx_dev_selftest() 47 ret = rte_rawdev_start(dev_id); in cnxk_bphy_cgx_dev_selftest() 55 ret = rte_rawdev_queue_conf_get(dev_id, i, &descs, in cnxk_bphy_cgx_dev_selftest() 67 ret = rte_pmd_bphy_cgx_stop_rxtx(dev_id, i); in cnxk_bphy_cgx_dev_selftest() 73 ret = rte_pmd_bphy_cgx_start_rxtx(dev_id, i); in cnxk_bphy_cgx_dev_selftest() 85 ret = cnxk_bphy_cgx_link_cond(dev_id, i, 0); in cnxk_bphy_cgx_dev_selftest() 96 ret = cnxk_bphy_cgx_link_cond(dev_id, i, 1); in cnxk_bphy_cgx_dev_selftest() 134 ret = rte_pmd_bphy_cgx_set_fec(dev_id, i, fec); in cnxk_bphy_cgx_dev_selftest() 141 ret = rte_pmd_bphy_cgx_set_fec(dev_id, i, fec); in cnxk_bphy_cgx_dev_selftest() [all …]
|
| H A D | cnxk_bphy_irq.c | 14 cnxk_bphy_get_bphy_dev_by_dev_id(uint16_t dev_id) in cnxk_bphy_get_bphy_dev_by_dev_id() argument 18 if (!rte_rawdev_pmd_is_valid_dev(dev_id)) in cnxk_bphy_get_bphy_dev_by_dev_id() 21 rawdev = &rte_rawdevs[dev_id]; in cnxk_bphy_get_bphy_dev_by_dev_id() 27 cnxk_bphy_irq_max_get(uint16_t dev_id) in cnxk_bphy_irq_max_get() argument 32 bphy_dev = cnxk_bphy_get_bphy_dev_by_dev_id(dev_id); in cnxk_bphy_irq_max_get() 39 cnxk_bphy_intr_init(uint16_t dev_id) in cnxk_bphy_intr_init() argument 41 struct bphy_device *bphy_dev = cnxk_bphy_get_bphy_dev_by_dev_id(dev_id); in cnxk_bphy_intr_init() 51 cnxk_bphy_intr_fini(uint16_t dev_id) in cnxk_bphy_intr_fini() argument 61 cnxk_bphy_intr_register(uint16_t dev_id, int irq_num, in cnxk_bphy_intr_register() argument 83 cnxk_bphy_intr_unregister(uint16_t dev_id, int irq_num) in cnxk_bphy_intr_unregister() argument [all …]
|
| /dpdk/drivers/raw/cnxk_gpio/ |
| H A D | rte_pmd_cnxk_gpio.h | 164 return __rte_pmd_gpio_enq_deq(dev_id, gpio, &msg, NULL, 0); in rte_pmd_gpio_set_pin_value() 181 rte_pmd_gpio_set_pin_edge(uint16_t dev_id, int gpio, in rte_pmd_gpio_set_pin_edge() argument 189 return __rte_pmd_gpio_enq_deq(dev_id, gpio, &msg, NULL, 0); in rte_pmd_gpio_set_pin_edge() 213 return __rte_pmd_gpio_enq_deq(dev_id, gpio, &msg, NULL, 0); in rte_pmd_gpio_set_pin_dir() 243 return __rte_pmd_gpio_enq_deq(dev_id, gpio, &msg, NULL, 0); in rte_pmd_gpio_set_pin_active_low() 283 rte_pmd_gpio_get_pin_edge(uint16_t dev_id, int gpio, in rte_pmd_gpio_get_pin_edge() argument 371 return __rte_pmd_gpio_enq_deq(dev_id, gpio, &msg, NULL, 0); in rte_pmd_gpio_register_irq() 386 rte_pmd_gpio_unregister_irq(uint16_t dev_id, int gpio) in rte_pmd_gpio_unregister_irq() argument 410 rte_pmd_gpio_enable_interrupt(uint16_t dev_id, int gpio, in rte_pmd_gpio_enable_interrupt() argument 413 return rte_pmd_gpio_set_pin_edge(dev_id, gpio, edge); in rte_pmd_gpio_enable_interrupt() [all …]
|
| H A D | cnxk_gpio_selftest.c | 106 ret = rte_pmd_gpio_set_pin_edge(dev_id, gpio, in cnxk_gpio_test_input() 112 ret = rte_pmd_gpio_set_pin_edge(dev_id, gpio, in cnxk_gpio_test_input() 155 cnxk_gpio_test_irq(uint16_t dev_id, int gpio) in cnxk_gpio_test_irq() argument 183 rte_pmd_gpio_disable_interrupt(dev_id, gpio); in cnxk_gpio_test_irq() 184 rte_pmd_gpio_unregister_irq(dev_id, gpio); in cnxk_gpio_test_irq() 243 ret = rte_pmd_gpio_set_pin_edge(dev_id, gpio, in cnxk_gpio_test_output() 303 cnxk_gpio_selftest(uint16_t dev_id) in cnxk_gpio_selftest() argument 318 queues = rte_rawdev_queue_count(dev_id); in cnxk_gpio_selftest() 322 ret = rte_rawdev_start(dev_id); in cnxk_gpio_selftest() 372 ret = rte_rawdev_queue_release(dev_id, i); in cnxk_gpio_selftest() [all …]
|