Home
last modified time | relevance | path

Searched refs:fq (Results 1 – 11 of 11) sorted by relevance

/dpdk/drivers/bus/dpaa/base/qbman/
H A Dqman.c903 fq->cb.fqs(p, fq, &swapped_msg); in __poll_portal_slow()
908 fq = get_fq_table_entry(msg->fq.contextB); in __poll_portal_slow()
910 fq = (void *)(uintptr_t)msg->fq.contextB; in __poll_portal_slow()
914 fq->cb.fqs(p, fq, &swapped_msg); in __poll_portal_slow()
941 fq->cb.ern(p, fq, &swapped_msg); in __poll_portal_slow()
1046 res = fq->cb.dqrr(p, fq, dq); in __poll_portal_fast()
1060 res = fq->cb.dqrr(p, fq, dq); in __poll_portal_fast()
1810 msg.fq.fqid = fq->fqid; in qman_retire_fq()
1812 msg.fq.contextB = fq->key; in qman_retire_fq()
1814 msg.fq.contextB = (u32)(uintptr_t)fq; in qman_retire_fq()
[all …]
/dpdk/drivers/net/dpaa/
H A Ddpaa_rxtx.c530 dpaa_intf = fq[0]->dpaa_intf; in dpaa_rx_cb_no_prefetch()
625 fq->qp_initialized = 1; in dpaa_eth_queue_portal_rx()
634 struct qman_fq *fq, in dpaa_rx_cb_parallel() argument
644 ev->flow_id = fq->ev.flow_id; in dpaa_rx_cb_parallel()
661 struct qman_fq *fq, in dpaa_rx_cb_atomic() argument
672 ev->flow_id = fq->ev.flow_id; in dpaa_rx_cb_atomic()
740 struct qman_fq *fq = q; in dpaa_eth_queue_rx() local
751 if (fq->fqid == ((struct dpaa_if *)fq->dpaa_intf)->rx_queues[0].fqid) in dpaa_eth_queue_rx()
755 if (likely(fq->is_static)) in dpaa_eth_queue_rx()
783 dq = qman_dequeue(fq); in dpaa_eth_queue_rx()
[all …]
H A Ddpaa_ethdev.c1686 ret = qman_create_fq(fqid, flags, fq); in dpaa_rx_queue_init()
1692 fq->is_static = false; in dpaa_rx_queue_init()
1705 fq->fqid, ret); in dpaa_rx_queue_init()
1713 ret = qman_init_fq(fq, 0, &opts); in dpaa_rx_queue_init()
1720 static int dpaa_tx_queue_init(struct qman_fq *fq, in dpaa_tx_queue_init() argument
1737 QMAN_FQ_FLAG_TO_DCPORTAL, fq); in dpaa_tx_queue_init()
1755 DPAA_PMD_DEBUG("init tx fq %p, fqid 0x%x", fq, fq->fqid); in dpaa_tx_queue_init()
1767 fq->fqid, ret); in dpaa_tx_queue_init()
1777 ret = qman_init_fq(fq, QMAN_INITFQ_FLAG_SCHED, &opts); in dpaa_tx_queue_init()
1800 ret = qman_create_fq(fqid, QMAN_FQ_FLAG_NO_ENQUEUE, fq); in dpaa_debug_queue_init()
[all …]
H A Ddpaa_ethdev.h185 struct qman_fq *fq,
191 struct qman_fq *fq,
H A Ddpaa_rxtx.h283 void dpaa_rx_cb(struct qman_fq **fq,
288 void dpaa_rx_cb_no_prefetch(struct qman_fq **fq,
/dpdk/drivers/bus/dpaa/include/
H A Dfsl_qman.h1134 struct qman_fq *fq,
1139 struct qman_fq *fq,
1428 void qman_dqrr_consume(struct qman_fq *fq,
1607 void qman_destroy_fq(struct qman_fq *fq, u32 flags);
1614 u32 qman_fq_fqid(struct qman_fq *fq);
1673 int qman_schedule_fq(struct qman_fq *fq);
1694 int qman_retire_fq(struct qman_fq *fq, u32 *flags);
1704 int qman_oos_fq(struct qman_fq *fq);
1715 int qman_fq_flow_control(struct qman_fq *fq, int xon);
1730 int qman_query_fq_has_pkts(struct qman_fq *fq);
[all …]
/dpdk/drivers/net/af_xdp/
H A Drte_eth_af_xdp.c118 struct xsk_ring_prod fq; member
226 xsk_ring_prod__submit(fq, reserve_size); in reserve_fill_queue_zc()
234 struct xsk_ring_prod *fq) in reserve_fill_queue_cp() argument
260 xsk_ring_prod__submit(fq, reserve_size); in reserve_fill_queue_cp()
283 struct xsk_ring_prod *fq = &rxq->fq; in af_xdp_rx_zc() local
360 struct xsk_ring_prod *fq = &rxq->fq; in af_xdp_rx_cp() local
364 uint32_t free_thresh = fq->size >> 1; in af_xdp_rx_cp()
373 if (xsk_ring_prod__needs_wakeup(fq)) in af_xdp_rx_cp()
1084 &rxq->fq, &rxq->cq, &usr_config); in xdp_umem_configure()
1158 &rxq->fq, &rxq->cq, in xdp_umem_configure()
[all …]
/dpdk/drivers/bus/dpaa/
H A Ddpaa_bus.c375 rte_dpaa_portal_fq_init(void *arg, struct qman_fq *fq) in rte_dpaa_portal_fq_init() argument
390 ret = fsl_qman_fq_portal_init(fq->qp); in rte_dpaa_portal_fq_init()
396 sdqcr = QM_SDQCR_CHANNELS_POOL_CONV(fq->ch_id); in rte_dpaa_portal_fq_init()
397 qman_static_dequeue_add(sdqcr, fq->qp); in rte_dpaa_portal_fq_init()
402 int rte_dpaa_portal_fq_close(struct qman_fq *fq) in rte_dpaa_portal_fq_close() argument
404 return fsl_qman_fq_portal_destroy(fq->qp); in rte_dpaa_portal_fq_close()
H A Drte_dpaa_bus.h240 int rte_dpaa_portal_fq_init(void *arg, struct qman_fq *fq);
243 int rte_dpaa_portal_fq_close(struct qman_fq *fq);
/dpdk/drivers/crypto/dpaa_sec/
H A Ddpaa_sec_raw_dp.c867 struct qman_fq *fq; in dpaa_sec_deq_raw() local
874 fq = &qp->outq; in dpaa_sec_deq_raw()
888 ret = qman_set_vdq(fq, num_rx_bufs, vdqcr_flags); in dpaa_sec_deq_raw()
897 dq = qman_dequeue(fq); in dpaa_sec_deq_raw()
927 qman_dqrr_consume(fq, dq); in dpaa_sec_deq_raw()
928 } while (fq->flags & QMAN_FQ_STATE_VDQCR); in dpaa_sec_deq_raw()
H A Ddpaa_sec.c104 struct qman_fq *fq, in ern_sec_fq_handler() argument
190 dpaa_sec_init_tx(struct qman_fq *fq) in dpaa_sec_init_tx() argument
211 fq->cb.dqrr = dqrr_out_fq_cb_rx; in dpaa_sec_init_tx()
212 fq->cb.ern = ern_sec_fq_handler; in dpaa_sec_init_tx()
214 ret = qman_init_fq(fq, 0, &opts); in dpaa_sec_init_tx()
798 struct qman_fq *fq; in dpaa_sec_deq() local
804 fq = &qp->outq; in dpaa_sec_deq()
828 dq = qman_dequeue(fq); in dpaa_sec_deq()
874 qman_dqrr_consume(fq, dq); in dpaa_sec_deq()
2534 if (&qi->inq[i] == fq) { in dpaa_sec_detach_rxq()
[all …]