Home
last modified time | relevance | path

Searched refs:fq (Results 1 – 13 of 13) sorted by relevance

/f-stack/dpdk/drivers/bus/dpaa/base/qbman/
H A Dqman.c903 fq->cb.fqs(p, fq, &swapped_msg); in __poll_portal_slow()
908 fq = get_fq_table_entry(msg->fq.contextB); in __poll_portal_slow()
910 fq = (void *)(uintptr_t)msg->fq.contextB; in __poll_portal_slow()
914 fq->cb.fqs(p, fq, &swapped_msg); in __poll_portal_slow()
941 fq->cb.ern(p, fq, &swapped_msg); in __poll_portal_slow()
1046 res = fq->cb.dqrr(p, fq, dq); in __poll_portal_fast()
1060 res = fq->cb.dqrr(p, fq, dq); in __poll_portal_fast()
1810 msg.fq.fqid = fq->fqid; in qman_retire_fq()
1812 msg.fq.contextB = fq->key; in qman_retire_fq()
1814 msg.fq.contextB = (u32)(uintptr_t)fq; in qman_retire_fq()
[all …]
/f-stack/dpdk/drivers/net/dpaa/
H A Ddpaa_rxtx.c530 dpaa_intf = fq[0]->dpaa_intf; in dpaa_rx_cb_no_prefetch()
625 fq->qp_initialized = 1; in dpaa_eth_queue_portal_rx()
634 struct qman_fq *fq, in dpaa_rx_cb_parallel() argument
644 ev->flow_id = fq->ev.flow_id; in dpaa_rx_cb_parallel()
661 struct qman_fq *fq, in dpaa_rx_cb_atomic() argument
672 ev->flow_id = fq->ev.flow_id; in dpaa_rx_cb_atomic()
740 struct qman_fq *fq = q; in dpaa_eth_queue_rx() local
751 if (fq->fqid == ((struct dpaa_if *)fq->dpaa_intf)->rx_queues[0].fqid) in dpaa_eth_queue_rx()
755 if (likely(fq->is_static)) in dpaa_eth_queue_rx()
783 dq = qman_dequeue(fq); in dpaa_eth_queue_rx()
[all …]
H A Ddpaa_ethdev.c1683 ret = qman_create_fq(fqid, flags, fq); in dpaa_rx_queue_init()
1689 fq->is_static = false; in dpaa_rx_queue_init()
1702 fq->fqid, ret); in dpaa_rx_queue_init()
1710 ret = qman_init_fq(fq, 0, &opts); in dpaa_rx_queue_init()
1717 static int dpaa_tx_queue_init(struct qman_fq *fq, in dpaa_tx_queue_init() argument
1734 QMAN_FQ_FLAG_TO_DCPORTAL, fq); in dpaa_tx_queue_init()
1748 DPAA_PMD_DEBUG("init tx fq %p, fqid 0x%x", fq, fq->fqid); in dpaa_tx_queue_init()
1760 fq->fqid, ret); in dpaa_tx_queue_init()
1770 ret = qman_init_fq(fq, QMAN_INITFQ_FLAG_SCHED, &opts); in dpaa_tx_queue_init()
1793 ret = qman_create_fq(fqid, QMAN_FQ_FLAG_NO_ENQUEUE, fq); in dpaa_debug_queue_init()
[all …]
H A Ddpaa_ethdev.h182 struct qman_fq *fq,
188 struct qman_fq *fq,
H A Ddpaa_rxtx.h289 void dpaa_rx_cb(struct qman_fq **fq,
294 void dpaa_rx_cb_no_prefetch(struct qman_fq **fq,
/f-stack/dpdk/drivers/bus/dpaa/include/
H A Dfsl_qman.h1134 struct qman_fq *fq,
1139 struct qman_fq *fq,
1428 void qman_dqrr_consume(struct qman_fq *fq,
1607 void qman_destroy_fq(struct qman_fq *fq, u32 flags);
1614 u32 qman_fq_fqid(struct qman_fq *fq);
1673 int qman_schedule_fq(struct qman_fq *fq);
1694 int qman_retire_fq(struct qman_fq *fq, u32 *flags);
1704 int qman_oos_fq(struct qman_fq *fq);
1715 int qman_fq_flow_control(struct qman_fq *fq, int xon);
1730 int qman_query_fq_has_pkts(struct qman_fq *fq);
[all …]
/f-stack/dpdk/drivers/net/af_xdp/
H A Drte_eth_af_xdp.c97 struct xsk_ring_prod fq; member
197 xsk_ring_prod__submit(fq, reserve_size); in reserve_fill_queue_zc()
205 struct xsk_ring_prod *fq) in reserve_fill_queue_cp() argument
254 struct xsk_ring_prod *fq = &rxq->fq; in af_xdp_rx_zc() local
272 if (xsk_ring_prod__needs_wakeup(fq)) in af_xdp_rx_zc()
326 struct xsk_ring_prod *fq = &rxq->fq; in af_xdp_rx_cp() local
330 uint32_t free_thresh = fq->size >> 1; in af_xdp_rx_cp()
335 NULL, fq); in af_xdp_rx_cp()
343 if (xsk_ring_prod__needs_wakeup(fq)) in af_xdp_rx_cp()
976 &rxq->fq, &rxq->cq, &usr_config);
[all …]
/f-stack/freebsd/contrib/ncsw/Peripherals/QM/
H A Dqm.h398 #define FQLOCK(fq) \ argument
400 if (fq->flags & QMAN_FQ_FLAG_LOCKED) \
401 XX_LockSpinlock(&fq->fqlock); \
403 #define FQUNLOCK(fq) \ argument
405 if (fq->flags & QMAN_FQ_FLAG_LOCKED) \
406 XX_UnlockSpinlock(&fq->fqlock); \
H A Dfsl_qman.h491 } _PackedType fq; /* FQRN/FQRNI/FQRL/FQPN */ member
979 struct qman_fq *fq,
1077 void qman_free_fq(struct qman_fq *fq, uint32_t flags);
1083 uint32_t qman_fq_fqid(struct qman_fq *fq);
1095 void qman_fq_state(struct qman_fq *fq, enum qman_fq_state *state, uint32_t *flags);
H A Dqm_portal_fqr.c419 if (p_Msg->fq.fqs & QM_MR_FQS_NOTEMPTY) in fq_state_change()
421 if (p_Msg->fq.fqs & QM_MR_FQS_ORLPRESENT) in fq_state_change()
610 struct qman_fq *p_FqFqs = ptr_from_aligned_int(p_Msg->fq.contextB); in LoopMessageRing()
/f-stack/dpdk/drivers/bus/dpaa/
H A Ddpaa_bus.c335 rte_dpaa_portal_fq_init(void *arg, struct qman_fq *fq) in rte_dpaa_portal_fq_init() argument
350 ret = fsl_qman_fq_portal_init(fq->qp); in rte_dpaa_portal_fq_init()
356 sdqcr = QM_SDQCR_CHANNELS_POOL_CONV(fq->ch_id); in rte_dpaa_portal_fq_init()
357 qman_static_dequeue_add(sdqcr, fq->qp); in rte_dpaa_portal_fq_init()
362 int rte_dpaa_portal_fq_close(struct qman_fq *fq) in rte_dpaa_portal_fq_close() argument
364 return fsl_qman_fq_portal_destroy(fq->qp); in rte_dpaa_portal_fq_close()
H A Drte_dpaa_bus.h240 int rte_dpaa_portal_fq_init(void *arg, struct qman_fq *fq);
243 int rte_dpaa_portal_fq_close(struct qman_fq *fq);
/f-stack/dpdk/drivers/crypto/dpaa_sec/
H A Ddpaa_sec.c93 struct qman_fq *fq, in ern_sec_fq_handler() argument
183 dpaa_sec_init_tx(struct qman_fq *fq) in dpaa_sec_init_tx() argument
204 fq->cb.dqrr = dqrr_out_fq_cb_rx; in dpaa_sec_init_tx()
205 fq->cb.ern = ern_sec_fq_handler; in dpaa_sec_init_tx()
207 ret = qman_init_fq(fq, 0, &opts); in dpaa_sec_init_tx()
618 struct qman_fq *fq; in dpaa_sec_deq() local
624 fq = &qp->outq; in dpaa_sec_deq()
648 dq = qman_dequeue(fq); in dpaa_sec_deq()
689 qman_dqrr_consume(fq, dq); in dpaa_sec_deq()
2284 if (&qi->inq[i] == fq) { in dpaa_sec_detach_rxq()
[all …]