| /dpdk/app/test/ |
| H A D | virtual_pmd.c | 141 struct virtual_ethdev_queue *tx_q; in virtual_ethdev_tx_queue_setup_success() local 143 tx_q = (struct virtual_ethdev_queue *)rte_zmalloc_socket(NULL, in virtual_ethdev_tx_queue_setup_success() 146 if (tx_q == NULL) in virtual_ethdev_tx_queue_setup_success() 149 tx_q->port_id = dev->data->port_id; in virtual_ethdev_tx_queue_setup_success() 150 tx_q->queue_id = tx_queue_id; in virtual_ethdev_tx_queue_setup_success() 152 dev->data->tx_queues[tx_queue_id] = tx_q; in virtual_ethdev_tx_queue_setup_success() 351 struct virtual_ethdev_queue *tx_q = queue; in virtual_ethdev_tx_burst_success() local 358 vrtl_eth_dev = &rte_eth_devices[tx_q->port_id]; in virtual_ethdev_tx_burst_success() 382 struct virtual_ethdev_queue *tx_q = NULL; in virtual_ethdev_tx_burst_fail() local 387 tx_q = queue; in virtual_ethdev_tx_burst_fail() [all …]
|
| /dpdk/examples/vhost/ |
| H A D | main.c | 1210 tx_q->m_table, tx_q->len); in do_drain_mbuf_table() 1212 free_pkts(&tx_q->m_table[count], tx_q->len - count); in do_drain_mbuf_table() 1214 tx_q->len = 0; in do_drain_mbuf_table() 1224 struct mbuf_table *tx_q; in virtio_tx_route() local 1297 tx_q->m_table[tx_q->len++] = m; in virtio_tx_route() 1304 do_drain_mbuf_table(tx_q); in virtio_tx_route() 1314 if (tx_q->len == 0) in drain_mbuf_table() 1323 tx_q->len); in drain_mbuf_table() 1437 struct mbuf_table *tx_q; in switch_worker() local 1444 tx_q->txq_id = i; in switch_worker() [all …]
|
| /dpdk/drivers/common/qat/ |
| H A D | qat_qp.c | 110 if (qat_queue_create(qat_dev, &(qp->tx_q), qat_qp_conf, in qat_qp_setup() 117 qp->max_inflights = ADF_MAX_INFLIGHTS(qp->tx_q.queue_size, in qat_qp_setup() 118 ADF_BYTES_TO_MSG_SIZE(qp->tx_q.msg_size)); in qat_qp_setup() 122 qat_queue_delete(&(qp->tx_q)); in qat_qp_setup() 130 qat_queue_delete(&(qp->tx_q)); in qat_qp_setup() 151 qat_queue_delete(&(qp->tx_q)); in qat_qp_setup() 316 qat_queue_delete(&(qp->tx_q)); in qat_qp_release() 322 ret = adf_queue_arb_disable(qat_dev_gen, &(qp->tx_q), in qat_qp_release() 565 queue = &(tmp_qp->tx_q); in qat_enqueue_op_burst() 655 queue = &(tmp_qp->tx_q); in qat_enqueue_comp_op_burst() [all …]
|
| H A D | qat_qp.h | 86 struct qat_queue tx_q; member
|
| /dpdk/drivers/common/qat/dev/ |
| H A D | qat_dev_gen1.c | 153 struct qat_queue *q_tx = &qp->tx_q, *q_rx = &qp->rx_q; in qat_qp_adf_configure_queues_gen1() 186 qat_qp_csr_build_ring_base_gen1(io_addr, &qp->tx_q); in qat_qp_csr_setup_gen1() 189 qat_qp_adf_arb_enable_gen1(&qp->tx_q, qp->mmap_bar_addr, in qat_qp_csr_setup_gen1()
|
| H A D | qat_dev_gen4.c | 196 struct qat_queue *q_tx = &qp->tx_q, *q_rx = &qp->rx_q; in qat_qp_adf_configure_queues_gen4() 230 qat_qp_build_ring_base_gen4(io_addr, &qp->tx_q); in qat_qp_csr_setup_gen4() 233 qat_qp_adf_arb_enable_gen4(&qp->tx_q, qp->mmap_bar_addr, in qat_qp_csr_setup_gen4()
|
| /dpdk/drivers/net/avp/ |
| H A D | avp_ethdev.c | 894 avp->tx_q[i] = avp_dev_translate_address(eth_dev, in avp_dev_create() 1700 struct rte_avp_fifo *tx_q; in avp_xmit_scattered_pkts() local 1717 tx_q = avp->tx_q[txq->queue_id]; in avp_xmit_scattered_pkts() 1731 count = avp_fifo_free_count(tx_q); in avp_xmit_scattered_pkts() 1765 nb_pkts, tx_q); in avp_xmit_scattered_pkts() 1805 n = avp_fifo_put(tx_q, (void **)&tx_bufs[0], nb_pkts); in avp_xmit_scattered_pkts() 1821 struct rte_avp_fifo *tx_q; in avp_xmit_pkts() local 1836 tx_q = avp->tx_q[txq->queue_id]; in avp_xmit_pkts() 1847 count = avp_fifo_free_count(tx_q); in avp_xmit_pkts() 1860 count, tx_q); in avp_xmit_pkts() [all …]
|
| /dpdk/drivers/crypto/bcmfs/ |
| H A D | bcmfs_qp.c | 178 bcmfs_queue_delete(&qp->tx_q, qp->qpair_id); in bcmfs_qp_release() 235 rc = bcmfs_queue_create(&qp->tx_q, qp_conf, qp->qpair_id, in bcmfs_qp_setup() 298 bcmfs_queue_delete(&qp->tx_q, queue_pair_id); in bcmfs_qp_setup()
|
| H A D | bcmfs_qp.h | 85 struct bcmfs_queue tx_q; member
|
| /dpdk/drivers/net/pfe/ |
| H A D | pfe_hif_lib.c | 201 hif_lib_client_cleanup_tx_queue(&client->tx_q[qno]); in hif_lib_client_release_tx_buffers() 219 queue = &client->tx_q[qno]; in hif_lib_client_init_tx_buffers() 496 struct hif_client_tx_queue *queue = &client->tx_q[qno]; in hif_lib_xmit_pkt() 522 struct hif_client_tx_queue *queue = &client->tx_q[qno]; in hif_lib_tx_get_next_complete()
|
| H A D | pfe_hif.h | 47 struct hif_tx_queue tx_q[HIF_CLIENT_QUEUES_MAX]; member
|
| H A D | pfe_hif_lib.h | 67 struct hif_client_tx_queue tx_q[HIF_CLIENT_QUEUES_MAX]; member
|
| H A D | pfe_hif.c | 252 tx_queue = &client->tx_q[i]; in pfe_hif_client_register() 487 struct hif_tx_queue *queue = &hif->client[client_id].tx_q[q_no]; in client_ack_txpacket()
|
| H A D | pfe_ethdev.c | 495 dev->data->tx_queues[queue_idx] = &priv->client.tx_q[queue_idx]; in pfe_tx_queue_setup() 496 priv->client.tx_q[queue_idx].queue_id = queue_idx; in pfe_tx_queue_setup()
|
| /dpdk/drivers/crypto/qat/dev/ |
| H A D | qat_sym_pmd_gen1.c | 466 struct qat_queue *tx_queue = &qp->tx_q; in qat_sym_dp_enqueue_single_cipher_gen1() 505 struct qat_queue *tx_queue = &qp->tx_q; in qat_sym_dp_enqueue_cipher_jobs_gen1() 574 struct qat_queue *tx_queue = &qp->tx_q; in qat_sym_dp_enqueue_single_auth_gen1() 613 struct qat_queue *tx_queue = &qp->tx_q; in qat_sym_dp_enqueue_auth_jobs_gen1() 682 struct qat_queue *tx_queue = &qp->tx_q; in qat_sym_dp_enqueue_single_chain_gen1() 723 struct qat_queue *tx_queue = &qp->tx_q; in qat_sym_dp_enqueue_chain_jobs_gen1() 799 struct qat_queue *tx_queue = &qp->tx_q; in qat_sym_dp_enqueue_single_aead_gen1() 838 struct qat_queue *tx_queue = &qp->tx_q; in qat_sym_dp_enqueue_aead_jobs_gen1() 1023 struct qat_queue *tx_queue = &qp->tx_q; in qat_sym_dp_enqueue_done_gen1()
|
| H A D | qat_crypto_pmd_gen3.c | 411 struct qat_queue *tx_queue = &qp->tx_q; in qat_sym_dp_enqueue_single_aead_gen3() 450 struct qat_queue *tx_queue = &qp->tx_q; in qat_sym_dp_enqueue_aead_jobs_gen3() 522 struct qat_queue *tx_queue = &qp->tx_q; in qat_sym_dp_enqueue_single_auth_gen3() 557 struct qat_queue *tx_queue = &qp->tx_q; in qat_sym_dp_enqueue_auth_jobs_gen3()
|
| H A D | qat_crypto_pmd_gen4.c | 240 struct qat_queue *tx_queue = &qp->tx_q; in qat_sym_dp_enqueue_single_aead_gen4() 279 struct qat_queue *tx_queue = &qp->tx_q; in qat_sym_dp_enqueue_aead_jobs_gen4()
|
| /dpdk/lib/kni/ |
| H A D | rte_kni.c | 68 struct rte_kni_fifo *tx_q; /**< TX queue */ member 275 kni->tx_q = kni->m_tx_q->addr; in rte_kni_alloc() 276 kni_fifo_init(kni->tx_q, KNI_FIFO_COUNT_MAX); in rte_kni_alloc() 444 kni_free_fifo(kni->tx_q); in rte_kni_release() 634 unsigned int ret = kni_fifo_get(kni->tx_q, (void **)mbufs, num); in rte_kni_rx_burst()
|
| /dpdk/kernel/linux/kni/ |
| H A D | kni_net.c | 321 if (kni_fifo_free_count(kni->tx_q) == 0 || in kni_net_tx() 349 ret = kni_fifo_put(kni->tx_q, &pkt_va, 1); in kni_net_tx() 477 num_tq = kni_fifo_free_count(kni->tx_q); in kni_net_rx_lo_fifo() 531 ret = kni_fifo_put(kni->tx_q, kni->alloc_va, num); in kni_net_rx_lo_fifo()
|
| H A D | kni_dev.h | 61 struct rte_kni_fifo *tx_q; member
|
| H A D | kni_misc.c | 359 kni->tx_q = iova_to_kva(current, dev_info.tx_phys); in kni_ioctl_create() 376 kni->tx_q = phys_to_virt(dev_info.tx_phys); in kni_ioctl_create() 391 (unsigned long long) dev_info.tx_phys, kni->tx_q); in kni_ioctl_create()
|
| /dpdk/drivers/crypto/bcmfs/hw/ |
| H A D | bcmfs5_rm.c | 389 struct bcmfs_queue *txq = &qp->tx_q; in bcmfs5_enqueue_single_request_qp() 442 struct bcmfs_queue *txq = &qp->tx_q; in bcmfs5_write_doorbell() 538 struct bcmfs_queue *tx_queue = &qp->tx_q; in bcmfs5_start_qp()
|
| H A D | bcmfs4_rm.c | 458 struct bcmfs_queue *txq = &qp->tx_q; in bcmfs4_enqueue_single_request_qp() 606 struct bcmfs_queue *tx_queue = &qp->tx_q; in bcmfs4_start_qp()
|
| /dpdk/drivers/crypto/qat/ |
| H A D | qat_sym.c | 384 dp_ctx->tail = qp->tx_q.tail; in qat_sym_configure_dp_ctx()
|
| /dpdk/drivers/compress/qat/ |
| H A D | qat_comp.c | 61 struct qat_queue *txq = &(cookie->qp->tx_q); in qat_comp_allocate_split_op_memzones() 359 struct qat_queue *txq = &(qp->tx_q); in qat_comp_build_multiple_requests()
|