| /dpdk/drivers/net/atlantic/hw_atl/ |
| H A D | hw_atl_b0.c | 38 u32 buff_size = 0U; in hw_atl_b0_hw_qos_set() local 58 buff_size = HW_ATL_B0_TXBUF_MAX; in hw_atl_b0_hw_qos_set() 60 hw_atl_tpb_tx_pkt_buff_size_per_tc_set(self, buff_size, tc); in hw_atl_b0_hw_qos_set() 62 (buff_size * in hw_atl_b0_hw_qos_set() 66 (buff_size * in hw_atl_b0_hw_qos_set() 72 buff_size = HW_ATL_B0_RXBUF_MAX; in hw_atl_b0_hw_qos_set() 74 hw_atl_rpb_rx_pkt_buff_size_per_tc_set(self, buff_size, tc); in hw_atl_b0_hw_qos_set() 76 (buff_size * in hw_atl_b0_hw_qos_set() 80 (buff_size * in hw_atl_b0_hw_qos_set() 417 int index, int size, int buff_size, int cpu, int vec) in hw_atl_b0_hw_ring_rx_init() argument [all …]
|
| H A D | hw_atl_b0.h | 19 int index, int size, int buff_size, int cpu, int vec);
|
| /dpdk/drivers/net/atlantic/ |
| H A D | atl_rxtx.c | 63 uint16_t buff_size; member 359 buff_size = RTE_ALIGN_FLOOR(buff_size, 1024); in atl_rx_init() 360 if (buff_size > HW_ATL_B0_RXD_BUF_SIZE_MAX) { in atl_rx_init() 364 buff_size = HW_ATL_B0_RXD_BUF_SIZE_MAX; in atl_rx_init() 366 if (buff_size < 1024) { in atl_rx_init() 372 rxq->buff_size = buff_size; in atl_rx_init() 375 rxq->nb_rx_desc, buff_size, 0, in atl_rx_init() 1022 u16 remainder_len = pkt_len % rxq->buff_size; in atl_recv_pkts() 1024 remainder_len = rxq->buff_size; in atl_recv_pkts() 1027 rx_mbuf->data_len = pkt_len > rxq->buff_size ? in atl_recv_pkts() [all …]
|
| /dpdk/drivers/net/i40e/base/ |
| H A D | i40e_adminq.c | 857 u16 buff_size, argument 912 if (buff_size > hw->aq.asq_buf_size) { 916 buff_size); 955 i40e_memcpy(dma_buff->va, buff, buff_size, 957 desc_on_ring->datalen = CPU_TO_LE16(buff_size); 971 buff, buff_size); 1000 i40e_memcpy(buff, dma_buff->va, buff_size, 1068 u16 buff_size, argument 1074 status = i40e_asq_send_command_exec(hw, desc, buff, buff_size, 1097 u16 buff_size, argument [all …]
|
| H A D | i40e_prototype.h | 39 u16 buff_size, 45 u16 buff_size, 265 void *buff, u16 buff_size, 268 void *buff, u16 buff_size, u16 *data_size, 283 u8 mib_type, void *buff, u16 buff_size, 287 u8 mib_type, void *buff, u16 buff_size, 306 void *buff, u16 buff_size, 572 u8 table_id, u32 start_index, u16 buff_size, 643 u16 buff_size, u32 track_id, 648 u16 buff_size, u8 flags,
|
| H A D | i40e_common.c | 3892 void *buff, u16 buff_size, in i40e_aq_oem_post_update() argument 4342 if (buff_size > I40E_AQ_LARGE_BUF) in i40e_aq_discover_capabilities() 4480 if (buff_size == 0 || !buff) in i40e_aq_get_lldp_mib() 4494 if (buff_size > I40E_AQ_LARGE_BUF) in i40e_aq_get_lldp_mib() 4527 if (buff_size == 0 || !buff) in i40e_aq_set_lldp_mib() 4534 if (buff_size > I40E_AQ_LARGE_BUF) in i40e_aq_set_lldp_mib() 4724 void *buff, u16 buff_size, in i40e_aq_get_cee_dcb_config() argument 4730 if (buff_size == 0 || !buff) in i40e_aq_get_cee_dcb_config() 5370 void *buff, u16 buff_size, in i40e_aq_tx_sched_cmd() argument 6497 if (buff_size == 0 || !buff) in i40e_aq_debug_dump() [all …]
|
| H A D | i40e_nvm.c | 1502 u32 buff_size = 0; in i40e_nvmupd_exec_aq() local 1529 buff_size = max(aq_data_len, (u32)LE16_TO_CPU(aq_desc->datalen)); in i40e_nvmupd_exec_aq() 1530 if (buff_size) { in i40e_nvmupd_exec_aq() 1552 buff_size, &cmd_details); in i40e_nvmupd_exec_aq()
|
| /dpdk/drivers/common/iavf/ |
| H A D | iavf_adminq.c | 650 u16 buff_size, in iavf_asq_send_command() argument 707 if (buff_size > hw->aq.asq_buf_size) { in iavf_asq_send_command() 711 buff_size); in iavf_asq_send_command() 750 iavf_memcpy(dma_buff->va, buff, buff_size, in iavf_asq_send_command() 752 desc_on_ring->datalen = CPU_TO_LE16(buff_size); in iavf_asq_send_command() 766 buff, buff_size); in iavf_asq_send_command() 795 iavf_memcpy(buff, dma_buff->va, buff_size, in iavf_asq_send_command() 819 iavf_debug_aq(hw, IAVF_DEBUG_AQ_COMMAND, (void *)desc, buff, buff_size); in iavf_asq_send_command()
|
| H A D | iavf_prototype.h | 42 u16 buff_size, 94 u8 table_id, u32 start_index, u16 buff_size,
|
| /dpdk/drivers/crypto/scheduler/ |
| H A D | scheduler_pmd_ops.c | 91 uint32_t buff_size = rte_align32pow2( in update_order_ring() local 99 if (!buff_size) in update_order_ring() 111 buff_size, rte_socket_id(), in update_order_ring()
|
| /dpdk/drivers/compress/qat/ |
| H A D | qat_comp_pmd.c | 193 uint32_t buff_size) in qat_comp_setup_inter_buffers() argument 232 buff_size * QAT_NUM_BUFS_IN_IM_SGL; in qat_comp_setup_inter_buffers() 306 lb * buff_size; in qat_comp_setup_inter_buffers() 307 sgl->buffers[lb].len = buff_size; in qat_comp_setup_inter_buffers()
|
| H A D | qat_comp_pmd.h | 107 uint32_t buff_size);
|
| /dpdk/app/test-pmd/ |
| H A D | cmdline.c | 14646 uint32_t buff_size = 0; in cmd_ddp_info_parsed() local 14692 (uint8_t *)&buff_size, sizeof(buff_size), in cmd_ddp_info_parsed() 14694 if (!ret && buff_size) { in cmd_ddp_info_parsed() 14695 buff = (uint8_t *)malloc(buff_size); in cmd_ddp_info_parsed() 14698 buff, buff_size, in cmd_ddp_info_parsed() 14714 (uint8_t *)devs, buff_size, in cmd_ddp_info_parsed() 14739 proto = (struct rte_pmd_i40e_proto_info *)malloc(buff_size); in cmd_ddp_info_parsed() 14744 buff_size, in cmd_ddp_info_parsed() 14765 buff_size, in cmd_ddp_info_parsed() 14800 ptype = (struct rte_pmd_i40e_ptype_info *)malloc(buff_size); in cmd_ddp_info_parsed() [all …]
|
| /dpdk/drivers/net/i40e/ |
| H A D | i40e_ethdev.c | 11560 uint32_t buff_size; in i40e_update_customized_pctype() local 11587 pctype = rte_zmalloc("new_pctype", buff_size, 0); in i40e_update_customized_pctype() 11594 (uint8_t *)pctype, buff_size, in i40e_update_customized_pctype() 11693 uint32_t buff_size; in i40e_update_customized_ptype() local 11724 buff_size = ptype_num * sizeof(struct rte_pmd_i40e_ptype_info); in i40e_update_customized_ptype() 11725 ptype = rte_zmalloc("new_ptype", buff_size, 0); in i40e_update_customized_ptype() 11733 (uint8_t *)ptype, buff_size, in i40e_update_customized_ptype() 11742 ptype_mapping = rte_zmalloc("ptype_mapping", buff_size, 0); in i40e_update_customized_ptype() 11895 uint32_t buff_size; in i40e_update_customized_info() local 11919 proto = rte_zmalloc("new_proto", buff_size, 0); in i40e_update_customized_info() [all …]
|