| /dpdk/drivers/net/bnxt/tf_core/ |
| H A D | tf_rm.c | 361 local_resv[f].stride = info.entry.stride; in tf_rm_check_residuals() 724 if (req_cnt[i] == resv[j].stride) { in tf_rm_create_db() 726 db[i].alloc.entry.stride = resv[j].stride; in tf_rm_create_db() 748 resv[j].stride, in tf_rm_create_db() 766 type_str, req_cnt[i], resv[j].stride); in tf_rm_create_db() 964 if (req_cnt[i] == resv[j].stride) { in tf_rm_create_db_no_reservation() 966 db[i].alloc.entry.stride = resv[j].stride; in tf_rm_create_db_no_reservation() 988 resv[j].stride, in tf_rm_create_db_no_reservation() 1006 type_str, req_cnt[i], resv[j].stride); in tf_rm_create_db_no_reservation() 1451 uint32_t stride; in tf_rm_check_indexes_in_range() local [all …]
|
| H A D | tf_tcam_shared.c | 88 int stride, in tf_tcam_shared_create_wc_pool() argument 97 pool_size = (BITALLOC_SIZEOF(stride) / sizeof(struct bitalloc)); in tf_tcam_shared_create_wc_pool() 112 stride, in tf_tcam_shared_create_wc_pool() 123 tcam_shared_wc->db[dir][id].info.stride = stride; in tf_tcam_shared_create_wc_pool() 138 tcam_shared_wc->db[dir][id].info.stride = 0; in tf_tcam_shared_free_wc_pool() 254 uint16_t start, stride; in tf_tcam_shared_bind() local 304 stride = info.entry.stride / 2; in tf_tcam_shared_bind() 309 stride, in tf_tcam_shared_bind() 312 start += stride; in tf_tcam_shared_bind() 316 stride, in tf_tcam_shared_bind() [all …]
|
| H A D | tf_msg.c | 550 resv[i].stride = tfp_le_to_cpu_16(resv_data[i].stride); in tf_msg_session_resc_alloc() 657 resv[i].stride = tfp_le_to_cpu_16(resv_data[i].stride); in tf_msg_session_resc_info() 731 resv_data[i].stride = tfp_cpu_to_le_16(resv[i].stride); in tf_msg_session_resc_flush()
|
| H A D | tf_em_internal.c | 291 iparms.info->entry.stride, in tf_em_int_bind()
|
| H A D | tf_core.h | 768 uint16_t stride; member
|
| H A D | tf_tbl.c | 607 if (dinfo[i].stride) in tf_tbl_get_resc_info()
|
| H A D | tf_tcam.c | 136 info.entry.stride % num_slices != 0) { in tf_tcam_bind()
|
| /dpdk/drivers/net/ngbe/ |
| H A D | ngbe_regs_group.h | 15 uint32_t stride; member 26 reg_buf[i] = rd32(hw, reg->base_addr + i * reg->stride); in ngbe_read_regs()
|
| /dpdk/drivers/net/txgbe/ |
| H A D | txgbe_regs_group.h | 15 uint32_t stride; member 27 reg->base_addr + i * reg->stride); in txgbe_read_regs()
|
| /dpdk/drivers/net/e1000/ |
| H A D | igb_regs.h | 12 uint32_t stride; member 165 reg->base_addr + i * reg->stride); in igb_read_regs()
|
| /dpdk/drivers/net/i40e/base/ |
| H A D | i40e_diag.h | 21 u32 stride; /* bytes between each element */ member
|
| H A D | i40e_diag.c | 107 + (j * i40e_reg_list[i].stride); in i40e_diag_reg_test()
|
| /dpdk/drivers/net/ixgbe/ |
| H A D | ixgbe_regs.h | 13 uint32_t stride; member 319 reg->base_addr + i * reg->stride); in ixgbe_read_regs()
|
| /dpdk/drivers/net/nfp/ |
| H A D | nfp_ethdev_vf.c | 289 int stride = 4; in nfp_netvf_init() local 404 hw->stride_rx = stride; in nfp_netvf_init() 405 hw->stride_tx = stride; in nfp_netvf_init()
|
| H A D | nfp_ethdev.c | 394 int stride = 4; in nfp_net_init() local 537 hw->stride_rx = stride; in nfp_net_init() 538 hw->stride_tx = stride; in nfp_net_init()
|
| /dpdk/doc/guides/nics/ |
| H A D | sfc_efx.rst | 108 Equal stride super-buffer mode 111 When the receive queue uses equal stride super-buffer DMA mode, one HW Rx 113 with some stride (equal to total size of rte_mbuf as mempool object). 119 Another limitation of a equal stride super-buffer mode, imposed by the 424 **ef10_essb** chooses SFNX2xxx equal stride super-buffer datapath 482 **dpdk** chooses DPDK firmware with equal stride super-buffer Rx mode 497 stride super-buffer Rx mode is used.
|
| H A D | mlx5.rst | 634 and each stride receives one packet. MPRQ can improve throughput for 639 configure large stride size enough to accommodate MTU as long as 655 Log 2 of the size of a stride for Multi-Packet Rx queue. Configuring a smaller 656 stride size can save some memory and reduce probability of a depletion of all 660 stride, valid only if ``mprq_en`` is set. With ``mprq_log_stride_size`` set 664 space left for a head room at the end of a stride which incurs some
|
| /dpdk/app/test-eventdev/ |
| H A D | test_perf_common.c | 559 perf_event_rx_adapter_setup(struct evt_options *opt, uint8_t stride, in perf_event_rx_adapter_setup() argument 580 queue_conf.ev.queue_id = prod * stride; in perf_event_rx_adapter_setup() 734 uint8_t stride, uint8_t nb_queues, in perf_event_dev_port_setup() argument 778 ret = perf_event_rx_adapter_setup(opt, stride, conf); in perf_event_dev_port_setup() 785 p->queue_id = prod * stride; in perf_event_dev_port_setup() 819 p->queue_id = prod * stride; in perf_event_dev_port_setup() 867 p->queue_id = prod * stride; in perf_event_dev_port_setup()
|
| H A D | test_perf_common.h | 175 uint8_t stride, uint8_t nb_queues,
|
| H A D | test_pipeline_common.c | 327 pipeline_event_rx_adapter_setup(struct evt_options *opt, uint8_t stride, in pipeline_event_rx_adapter_setup() argument 410 queue_conf.ev.queue_id = prod * stride; in pipeline_event_rx_adapter_setup()
|
| H A D | test_pipeline_common.h | 196 int pipeline_event_rx_adapter_setup(struct evt_options *opt, uint8_t stride,
|
| /dpdk/drivers/raw/dpaa2_qdma/ |
| H A D | dpaa2_qdma.h | 181 uint32_t stride; member
|
| /dpdk/drivers/net/qede/base/ |
| H A D | ecore_iov_api.h | 169 u16 stride; /* following VF stride */ member
|
| /dpdk/drivers/common/sfc_efx/base/ |
| H A D | ef10_tlv_layout.h | 204 uint16_t stride; member
|
| /dpdk/doc/guides/prog_guide/ |
| H A D | lpm6_lib.rst | 54 The implementation can be seen as a multi-bit trie where the *stride*
|