| /dpdk/drivers/bus/dpaa/base/qbman/ |
| H A D | qman.c | 903 fq->cb.fqs(p, fq, &swapped_msg); in __poll_portal_slow() 908 fq = get_fq_table_entry(msg->fq.contextB); in __poll_portal_slow() 910 fq = (void *)(uintptr_t)msg->fq.contextB; in __poll_portal_slow() 914 fq->cb.fqs(p, fq, &swapped_msg); in __poll_portal_slow() 941 fq->cb.ern(p, fq, &swapped_msg); in __poll_portal_slow() 1046 res = fq->cb.dqrr(p, fq, dq); in __poll_portal_fast() 1060 res = fq->cb.dqrr(p, fq, dq); in __poll_portal_fast() 1810 msg.fq.fqid = fq->fqid; in qman_retire_fq() 1812 msg.fq.contextB = fq->key; in qman_retire_fq() 1814 msg.fq.contextB = (u32)(uintptr_t)fq; in qman_retire_fq() [all …]
|
| /dpdk/drivers/net/dpaa/ |
| H A D | dpaa_rxtx.c | 530 dpaa_intf = fq[0]->dpaa_intf; in dpaa_rx_cb_no_prefetch() 625 fq->qp_initialized = 1; in dpaa_eth_queue_portal_rx() 634 struct qman_fq *fq, in dpaa_rx_cb_parallel() argument 644 ev->flow_id = fq->ev.flow_id; in dpaa_rx_cb_parallel() 661 struct qman_fq *fq, in dpaa_rx_cb_atomic() argument 672 ev->flow_id = fq->ev.flow_id; in dpaa_rx_cb_atomic() 740 struct qman_fq *fq = q; in dpaa_eth_queue_rx() local 751 if (fq->fqid == ((struct dpaa_if *)fq->dpaa_intf)->rx_queues[0].fqid) in dpaa_eth_queue_rx() 755 if (likely(fq->is_static)) in dpaa_eth_queue_rx() 783 dq = qman_dequeue(fq); in dpaa_eth_queue_rx() [all …]
|
| H A D | dpaa_ethdev.c | 1686 ret = qman_create_fq(fqid, flags, fq); in dpaa_rx_queue_init() 1692 fq->is_static = false; in dpaa_rx_queue_init() 1705 fq->fqid, ret); in dpaa_rx_queue_init() 1713 ret = qman_init_fq(fq, 0, &opts); in dpaa_rx_queue_init() 1720 static int dpaa_tx_queue_init(struct qman_fq *fq, in dpaa_tx_queue_init() argument 1737 QMAN_FQ_FLAG_TO_DCPORTAL, fq); in dpaa_tx_queue_init() 1755 DPAA_PMD_DEBUG("init tx fq %p, fqid 0x%x", fq, fq->fqid); in dpaa_tx_queue_init() 1767 fq->fqid, ret); in dpaa_tx_queue_init() 1777 ret = qman_init_fq(fq, QMAN_INITFQ_FLAG_SCHED, &opts); in dpaa_tx_queue_init() 1800 ret = qman_create_fq(fqid, QMAN_FQ_FLAG_NO_ENQUEUE, fq); in dpaa_debug_queue_init() [all …]
|
| H A D | dpaa_ethdev.h | 185 struct qman_fq *fq, 191 struct qman_fq *fq,
|
| H A D | dpaa_rxtx.h | 283 void dpaa_rx_cb(struct qman_fq **fq, 288 void dpaa_rx_cb_no_prefetch(struct qman_fq **fq,
|
| /dpdk/drivers/bus/dpaa/include/ |
| H A D | fsl_qman.h | 1134 struct qman_fq *fq, 1139 struct qman_fq *fq, 1428 void qman_dqrr_consume(struct qman_fq *fq, 1607 void qman_destroy_fq(struct qman_fq *fq, u32 flags); 1614 u32 qman_fq_fqid(struct qman_fq *fq); 1673 int qman_schedule_fq(struct qman_fq *fq); 1694 int qman_retire_fq(struct qman_fq *fq, u32 *flags); 1704 int qman_oos_fq(struct qman_fq *fq); 1715 int qman_fq_flow_control(struct qman_fq *fq, int xon); 1730 int qman_query_fq_has_pkts(struct qman_fq *fq); [all …]
|
| /dpdk/drivers/net/af_xdp/ |
| H A D | rte_eth_af_xdp.c | 118 struct xsk_ring_prod fq; member 226 xsk_ring_prod__submit(fq, reserve_size); in reserve_fill_queue_zc() 234 struct xsk_ring_prod *fq) in reserve_fill_queue_cp() argument 260 xsk_ring_prod__submit(fq, reserve_size); in reserve_fill_queue_cp() 283 struct xsk_ring_prod *fq = &rxq->fq; in af_xdp_rx_zc() local 360 struct xsk_ring_prod *fq = &rxq->fq; in af_xdp_rx_cp() local 364 uint32_t free_thresh = fq->size >> 1; in af_xdp_rx_cp() 373 if (xsk_ring_prod__needs_wakeup(fq)) in af_xdp_rx_cp() 1084 &rxq->fq, &rxq->cq, &usr_config); in xdp_umem_configure() 1158 &rxq->fq, &rxq->cq, in xdp_umem_configure() [all …]
|
| /dpdk/drivers/bus/dpaa/ |
| H A D | dpaa_bus.c | 375 rte_dpaa_portal_fq_init(void *arg, struct qman_fq *fq) in rte_dpaa_portal_fq_init() argument 390 ret = fsl_qman_fq_portal_init(fq->qp); in rte_dpaa_portal_fq_init() 396 sdqcr = QM_SDQCR_CHANNELS_POOL_CONV(fq->ch_id); in rte_dpaa_portal_fq_init() 397 qman_static_dequeue_add(sdqcr, fq->qp); in rte_dpaa_portal_fq_init() 402 int rte_dpaa_portal_fq_close(struct qman_fq *fq) in rte_dpaa_portal_fq_close() argument 404 return fsl_qman_fq_portal_destroy(fq->qp); in rte_dpaa_portal_fq_close()
|
| H A D | rte_dpaa_bus.h | 240 int rte_dpaa_portal_fq_init(void *arg, struct qman_fq *fq); 243 int rte_dpaa_portal_fq_close(struct qman_fq *fq);
|
| /dpdk/drivers/crypto/dpaa_sec/ |
| H A D | dpaa_sec_raw_dp.c | 867 struct qman_fq *fq; in dpaa_sec_deq_raw() local 874 fq = &qp->outq; in dpaa_sec_deq_raw() 888 ret = qman_set_vdq(fq, num_rx_bufs, vdqcr_flags); in dpaa_sec_deq_raw() 897 dq = qman_dequeue(fq); in dpaa_sec_deq_raw() 927 qman_dqrr_consume(fq, dq); in dpaa_sec_deq_raw() 928 } while (fq->flags & QMAN_FQ_STATE_VDQCR); in dpaa_sec_deq_raw()
|
| H A D | dpaa_sec.c | 104 struct qman_fq *fq, in ern_sec_fq_handler() argument 190 dpaa_sec_init_tx(struct qman_fq *fq) in dpaa_sec_init_tx() argument 211 fq->cb.dqrr = dqrr_out_fq_cb_rx; in dpaa_sec_init_tx() 212 fq->cb.ern = ern_sec_fq_handler; in dpaa_sec_init_tx() 214 ret = qman_init_fq(fq, 0, &opts); in dpaa_sec_init_tx() 798 struct qman_fq *fq; in dpaa_sec_deq() local 804 fq = &qp->outq; in dpaa_sec_deq() 828 dq = qman_dequeue(fq); in dpaa_sec_deq() 874 qman_dqrr_consume(fq, dq); in dpaa_sec_deq() 2534 if (&qi->inq[i] == fq) { in dpaa_sec_detach_rxq() [all …]
|