| /dpdk/drivers/common/cnxk/ |
| H A D | roc_mbox.c | 54 mdev->msg_size = 0; in mbox_reset() 56 tx_hdr->msg_size = 0; in mbox_reset() 58 rx_hdr->msg_size = 0; in mbox_reset() 178 if (mdev->msg_size == 0) in mbox_alloc_msg_rsp() 184 mdev->msg_size)); in mbox_alloc_msg_rsp() 190 mdev->msg_size += size; in mbox_alloc_msg_rsp() 192 msghdr->next_msgoff = mdev->msg_size + msgs_offset(); in mbox_alloc_msg_rsp() 213 tx_hdr->msg_size = mdev->msg_size; in mbox_msg_send() 214 mdev->msg_size = 0; in mbox_msg_send() 346 rx_hdr->num_msgs, mdev->msg_size, in mbox_wait() [all …]
|
| H A D | roc_mbox_priv.h | 50 uint16_t msg_size; /* Total msg size to be sent */ member
|
| H A D | roc_mbox.h | 19 uint64_t __io msg_size; /* Total msgs size embedded */ member
|
| /dpdk/drivers/common/qat/ |
| H A D | qat_qp.c | 118 ADF_BYTES_TO_MSG_SIZE(qp->tx_q.msg_size)); in qat_qp_setup() 245 queue->msg_size = desc_size; in qat_queue_create() 502 max_head = qp->nb_descriptors * q->msg_size; in rxq_free_desc() 528 adf_verify_queue_size(uint32_t msg_size, uint32_t msg_num, in adf_verify_queue_size() argument 534 if ((msg_size * msg_num) == in adf_verify_queue_size() 539 QAT_LOG(ERR, "Invalid ring size %d", msg_size * msg_num); in adf_verify_queue_size() 831 head = adf_modulo(head + rx_queue->msg_size, in qat_dequeue_op_burst() 895 memcpy(out_data, resp_msg, queue->msg_size); in qat_cq_dequeue_response() 897 memset(out_data, 0, queue->msg_size); in qat_cq_dequeue_response() 900 queue->head = adf_modulo(queue->head + queue->msg_size, in qat_cq_dequeue_response() [all …]
|
| H A D | qat_qp.h | 27 uint32_t msg_size; member
|
| /dpdk/drivers/crypto/qat/dev/ |
| H A D | qat_sym_pmd_gen1.c | 476 tail = (tail + tx_queue->msg_size) & tx_queue->modulo_mask; in qat_sym_dp_enqueue_single_cipher_gen1() 545 tail = (tail + tx_queue->msg_size) & tx_queue->modulo_mask; in qat_sym_dp_enqueue_cipher_jobs_gen1() 584 tail = (tail + tx_queue->msg_size) & tx_queue->modulo_mask; in qat_sym_dp_enqueue_single_auth_gen1() 653 tail = (tail + tx_queue->msg_size) & tx_queue->modulo_mask; in qat_sym_dp_enqueue_auth_jobs_gen1() 692 tail = (tail + tx_queue->msg_size) & tx_queue->modulo_mask; in qat_sym_dp_enqueue_single_chain_gen1() 810 tail = (tail + tx_queue->msg_size) & tx_queue->modulo_mask; in qat_sym_dp_enqueue_single_aead_gen1() 948 head = (head + rx_queue->msg_size) & rx_queue->modulo_mask; in qat_sym_dp_dequeue_burst_gen1() 967 head = (head + rx_queue->msg_size) & in qat_sym_dp_dequeue_burst_gen1() 981 head = (head + rx_queue->msg_size) & in qat_sym_dp_dequeue_burst_gen1() 1008 dp_ctx->head = (dp_ctx->head + rx_queue->msg_size) & in qat_sym_dp_dequeue_single_gen1() [all …]
|
| H A D | qat_crypto_pmd_gen3.c | 422 tail = (tail + tx_queue->msg_size) & tx_queue->modulo_mask; in qat_sym_dp_enqueue_single_aead_gen3() 493 tail = (tail + tx_queue->msg_size) & tx_queue->modulo_mask; in qat_sym_dp_enqueue_aead_jobs_gen3() 532 tail = (tail + tx_queue->msg_size) & tx_queue->modulo_mask; in qat_sym_dp_enqueue_single_auth_gen3() 597 tail = (tail + tx_queue->msg_size) & tx_queue->modulo_mask; in qat_sym_dp_enqueue_auth_jobs_gen3()
|
| H A D | qat_crypto_pmd_gen4.c | 251 tail = (tail + tx_queue->msg_size) & tx_queue->modulo_mask; in qat_sym_dp_enqueue_single_aead_gen4() 322 tail = (tail + tx_queue->msg_size) & tx_queue->modulo_mask; in qat_sym_dp_enqueue_aead_jobs_gen4()
|
| /dpdk/drivers/net/hinic/base/ |
| H A D | hinic_pmd_mgmt.c | 80 u16 msg_size = (u16)(MGMT_MSG_RSVD_FOR_DEV + sizeof(u64) + in mgmt_msg_len() local 83 if (msg_size > MGMT_MSG_SIZE_MIN) in mgmt_msg_len() 84 msg_size = MGMT_MSG_SIZE_MIN + in mgmt_msg_len() 85 ALIGN((msg_size - MGMT_MSG_SIZE_MIN), in mgmt_msg_len() 88 msg_size = MGMT_MSG_SIZE_MIN; in mgmt_msg_len() 90 return msg_size; in mgmt_msg_len()
|
| /dpdk/drivers/baseband/la12xx/ |
| H A D | bbdev_la12xx.c | 137 uint32_t msg_size = sizeof(struct bbdev_ipc_enqueue_op); in ipc_queue_configure() local 146 channel_id, q_priv->queue_size, msg_size); in ipc_queue_configure() 152 ch->md.msg_size = msg_size; in ipc_queue_configure() 154 vaddr = rte_malloc(NULL, msg_size, RTE_CACHE_LINE_SIZE); in ipc_queue_configure()
|
| H A D | bbdev_la12xx_ipc.h | 104 uint32_t msg_size; /**< Size of the each buffer */ member
|
| /dpdk/drivers/net/ngbe/ |
| H A D | ngbe_pf.c | 625 uint16_t msg_size = NGBE_VF_MSG_SIZE_DEFAULT; in ngbe_rcv_msg_from_vf() local 701 msg_size = NGBE_VF_GET_QUEUE_MSG_SIZE; in ngbe_rcv_msg_from_vf() 725 ngbe_write_mbx(hw, msgbuf, msg_size, vf); in ngbe_rcv_msg_from_vf()
|
| /dpdk/drivers/net/txgbe/ |
| H A D | txgbe_pf.c | 765 uint16_t msg_size = TXGBE_VF_MSG_SIZE_DEFAULT; in txgbe_rcv_msg_from_vf() local 841 msg_size = TXGBE_VF_GET_QUEUE_MSG_SIZE; in txgbe_rcv_msg_from_vf() 865 txgbe_write_mbx(hw, msgbuf, msg_size, vf); in txgbe_rcv_msg_from_vf()
|
| /dpdk/drivers/net/ixgbe/ |
| H A D | ixgbe_pf.c | 828 uint16_t msg_size = IXGBE_VF_MSG_SIZE_DEFAULT; in ixgbe_rcv_msg_from_vf() local 904 msg_size = IXGBE_VF_GET_QUEUE_MSG_SIZE; in ixgbe_rcv_msg_from_vf() 928 ixgbe_write_mbx(hw, msgbuf, msg_size, vf); in ixgbe_rcv_msg_from_vf()
|
| /dpdk/drivers/compress/qat/ |
| H A D | qat_comp.c | 367 qp->op_cookies[parent_tail / txq->msg_size]; in qat_comp_build_multiple_requests() 433 tail = adf_modulo(tail + txq->msg_size, txq->modulo_mask); in qat_comp_build_multiple_requests() 434 cookie_index = tail / txq->msg_size; in qat_comp_build_multiple_requests()
|
| /dpdk/drivers/net/fm10k/base/ |
| H A D | fm10k_mbx.c | 674 u16 msg_size = fm10k_fifo_head_len(&mbx->rx); in fm10k_mbx_rx_ready() local 676 return msg_size && (fm10k_fifo_used(&mbx->rx) >= msg_size); in fm10k_mbx_rx_ready()
|
| /dpdk/drivers/net/qede/base/ |
| H A D | bcm_osal.h | 354 #define OSAL_VF_SEND_MSG2PF(dev, done, msg, reply_addr, msg_size, reply_size) 0 argument
|