Home
last modified time | relevance | path

Searched refs:avail (Results 1 – 25 of 61) sorted by relevance

123

/dpdk/drivers/crypto/virtio/
H A Dvirtqueue.h117 vq->vq_ring.avail->idx = vq->vq_avail_idx; in vq_update_avail_idx()
132 if (unlikely(vq->vq_ring.avail->ring[avail_idx] != desc_idx)) in vq_update_avail_ring()
133 vq->vq_ring.avail->ring[avail_idx] = desc_idx; in vq_update_avail_ring()
166 (vq)->vq_desc_head_idx, (vq)->vq_ring.avail->idx, \
168 (vq)->vq_ring.avail->flags, (vq)->vq_ring.used->flags); \
H A Dvirtio_ring.h63 struct vring_avail *avail; member
97 #define vring_used_event(vr) ((vr)->avail->ring[(vr)->num])
119 vr->avail = (struct vring_avail *) (p + in vring_init()
122 RTE_ALIGN_CEIL((uintptr_t)(&vr->avail->ring[num]), align); in vring_init()
H A Dvirtqueue.c21 vq->vq_ring.avail->flags |= VRING_AVAIL_F_NO_INTERRUPT; in virtqueue_disable_intr()
/dpdk/drivers/bus/vmbus/
H A Dvmbus_bufring.c124 uint32_t avail; in vmbus_txbr_write() local
132 avail = vmbus_br_availwrite(tbr, old_windex); in vmbus_txbr_write()
135 if (avail <= total) in vmbus_txbr_write()
195 uint32_t avail; in vmbus_rxbr_peek() local
201 avail = vmbus_br_availread(rbr); in vmbus_rxbr_peek()
202 if (avail < dlen + sizeof(uint64_t)) in vmbus_rxbr_peek()
/dpdk/lib/ring/
H A Drte_ring_peek_elem_pvt.h151 uint32_t avail, head, next; in __rte_ring_do_dequeue_start() local
156 behavior, &head, &next, &avail); in __rte_ring_do_dequeue_start()
160 &head, &avail); in __rte_ring_do_dequeue_start()
168 avail = 0; in __rte_ring_do_dequeue_start()
175 *available = avail - n; in __rte_ring_do_dequeue_start()
H A Drte_ring_peek_zc.h335 uint32_t avail, head, next; in __rte_ring_do_dequeue_zc_elem_start() local
340 behavior, &head, &next, &avail); in __rte_ring_do_dequeue_zc_elem_start()
344 &head, &avail); in __rte_ring_do_dequeue_zc_elem_start()
352 avail = 0; in __rte_ring_do_dequeue_zc_elem_start()
360 *available = avail - n; in __rte_ring_do_dequeue_zc_elem_start()
/dpdk/drivers/net/virtio/
H A Dvirtio_ring.h94 struct vring_avail *avail; member
128 #define vring_used_event(vr) ((vr)->avail->ring[(vr)->num])
157 vr->avail = (struct vring_avail *) (p + in vring_init_split()
160 RTE_ALIGN_CEIL((uintptr_t)(&vr->avail->ring[num]), align); in vring_init_split()
H A Dvirtqueue.h367 uint16_t used, avail, flags; in desc_is_used() local
371 avail = !!(flags & VRING_PACKED_DESC_F_AVAIL); in desc_is_used()
373 return avail == used && used == vq->vq_packed.used_wrap_counter; in desc_is_used()
428 vq->vq_split.ring.avail->flags |= VRING_AVAIL_F_NO_INTERRUPT; in virtqueue_disable_intr_split()
462 vq->vq_split.ring.avail->flags &= (~VRING_AVAIL_F_NO_INTERRUPT); in virtqueue_enable_intr_split()
555 vq->vq_split.ring.avail->idx = vq->vq_avail_idx; in vq_update_avail_idx()
557 __atomic_store_n(&vq->vq_split.ring.avail->idx, in vq_update_avail_idx()
562 vq->vq_split.ring.avail->idx = vq->vq_avail_idx; in vq_update_avail_idx()
578 if (unlikely(vq->vq_split.ring.avail->ring[avail_idx] != desc_idx)) in vq_update_avail_ring()
579 vq->vq_split.ring.avail->ring[avail_idx] = desc_idx; in vq_update_avail_ring()
[all …]
/dpdk/drivers/bus/dpaa/base/qbman/
H A Dbman.c241 static void update_rcr_ci(struct bman_portal *p, int avail) in update_rcr_ci() argument
243 if (avail) in update_rcr_ci()
256 u8 avail; in bman_release() local
266 avail = bm_rcr_get_avail(&p->p); in bman_release()
267 if (avail < 2) in bman_release()
268 update_rcr_ci(p, avail); in bman_release()
/dpdk/drivers/net/ionic/
H A Dionic_dev.h231 uint16_t avail = q->tail_idx; in ionic_q_space_avail() local
233 if (q->head_idx >= avail) in ionic_q_space_avail()
234 avail += q->num_descs - q->head_idx - 1; in ionic_q_space_avail()
236 avail -= q->head_idx + 1; in ionic_q_space_avail()
238 return avail; in ionic_q_space_avail()
/dpdk/drivers/net/vmxnet3/
H A Dvmxnet3_rxtx.c73 uint32_t avail = 0; in vmxnet3_rxq_dump() local
90 (uint32_t)rxq->cmd_ring[0].size, avail, in vmxnet3_rxq_dump()
92 rxq->cmd_ring[0].size - avail); in vmxnet3_rxq_dump()
97 rxq->cmd_ring[1].size - avail); in vmxnet3_rxq_dump()
104 uint32_t avail = 0; in vmxnet3_txq_dump() local
118 (uint32_t)txq->cmd_ring.size, avail, in vmxnet3_txq_dump()
417 uint32_t first2fill, avail, dw2; in vmxnet3_xmit_pkts() local
426 if (count > avail) { in vmxnet3_xmit_pkts()
428 if (unlikely(avail == 0)) { in vmxnet3_xmit_pkts()
1003 uint32_t avail; in vmxnet3_recv_pkts() local
[all …]
/dpdk/app/test/
H A Dtest_ring_mpmc_stress.c9 uint32_t *avail) in _st_ring_dequeue_bulk() argument
11 return rte_ring_mc_dequeue_bulk(r, obj, n, avail); in _st_ring_dequeue_bulk()
H A Dtest_ring_rts_stress.c9 uint32_t *avail) in _st_ring_dequeue_bulk() argument
11 return rte_ring_mc_rts_dequeue_bulk(r, obj, n, avail); in _st_ring_dequeue_bulk()
H A Dtest_ring_hts_stress.c9 uint32_t *avail) in _st_ring_dequeue_bulk() argument
11 return rte_ring_mc_hts_dequeue_bulk(r, obj, n, avail); in _st_ring_dequeue_bulk()
H A Dtest_ring_mt_peek_stress.c10 uint32_t *avail) in _st_ring_dequeue_bulk() argument
14 m = rte_ring_dequeue_bulk_start(r, obj, n, avail); in _st_ring_dequeue_bulk()
H A Dtest_ring_mt_peek_stress_zc.c11 uint32_t *avail) in _st_ring_dequeue_bulk() argument
16 m = rte_ring_dequeue_zc_bulk_start(r, n, &zcd, avail); in _st_ring_dequeue_bulk()
H A Dtest_ring_st_peek_stress.c10 uint32_t *avail) in _st_ring_dequeue_bulk() argument
18 m = rte_ring_dequeue_bulk_start(r, obj, n, avail); in _st_ring_dequeue_bulk()
H A Dtest_ring_st_peek_stress_zc.c11 uint32_t *avail) in _st_ring_dequeue_bulk() argument
20 m = rte_ring_dequeue_zc_bulk_start(r, n, &zcd, avail); in _st_ring_dequeue_bulk()
/dpdk/drivers/mempool/octeontx/
H A Docteontx_fpavf.c494 uint64_t cnt, limit, avail; in octeontx_fpa_bufpool_free_count() local
515 avail = fpavf_read64((void *)((uintptr_t)pool_bar + in octeontx_fpa_bufpool_free_count()
518 return RTE_MIN(avail, (limit - cnt)); in octeontx_fpa_bufpool_free_count()
613 uint64_t cnt, avail; in octeontx_fpa_bufpool_destroy() local
645 avail = fpavf_read64((void *)((uintptr_t)pool_bar + in octeontx_fpa_bufpool_destroy()
649 fpavf_write64(avail, (void *)((uintptr_t)pool_bar + in octeontx_fpa_bufpool_destroy()
651 fpavf_write64(avail + 1, (void *)((uintptr_t)pool_bar + in octeontx_fpa_bufpool_destroy()
659 while (avail--) { in octeontx_fpa_bufpool_destroy()
668 gaura, avail); in octeontx_fpa_bufpool_destroy()
/dpdk/examples/vhost/
H A Dvirtio_net.c200 avail_idx = __atomic_load_n(&vr->avail->idx, __ATOMIC_ACQUIRE); in vs_enqueue_pkts()
209 rte_prefetch0(&vr->avail->ring[start_idx & (vr->size - 1)]); in vs_enqueue_pkts()
212 desc_indexes[i] = vr->avail->ring[used_idx]; in vs_enqueue_pkts()
381 free_entries = __atomic_load_n(&vr->avail->idx, __ATOMIC_ACQUIRE) - in vs_dequeue_pkts()
389 rte_prefetch0(&vr->avail->ring[avail_idx]); in vs_dequeue_pkts()
405 desc_indexes[i] = vr->avail->ring[avail_idx]; in vs_dequeue_pkts()
/dpdk/drivers/net/ice/base/
H A Dice_acl_ctrl.c414 ice_bitmap_set(tbl->avail, first_e, last_e - first_e + 1); in ice_acl_create_tbl()
490 bool avail = true; in ice_acl_alloc_partition() local
498 for (w = row; w < row + width && avail; w++) { in ice_acl_alloc_partition()
502 avail &= ice_is_bit_set(hw->acl_tbl->avail, b); in ice_acl_alloc_partition()
505 if (!avail) { in ice_acl_alloc_partition()
723 ice_set_bit(b, hw->acl_tbl->avail); in ice_acl_commit_partition()
725 ice_clear_bit(b, hw->acl_tbl->avail); in ice_acl_commit_partition()
/dpdk/drivers/baseband/fpga_lte_fec/
H A Dfpga_lte_fec.c1945 int32_t avail; in fpga_enqueue_enc() local
1956 avail = (q->head_free_desc > q->tail) ? in fpga_enqueue_enc()
1967 if (unlikely(avail - cbs_in_op < 0)) in fpga_enqueue_enc()
1969 avail -= cbs_in_op; in fpga_enqueue_enc()
1976 if (unlikely(avail - 1 < 0)) in fpga_enqueue_enc()
1978 avail -= 1; in fpga_enqueue_enc()
2015 int32_t avail; in fpga_enqueue_dec() local
2037 if (unlikely(avail - cbs_in_op < 0)) in fpga_enqueue_dec()
2039 avail -= cbs_in_op; in fpga_enqueue_dec()
2046 if (unlikely(avail - 1 < 0)) in fpga_enqueue_dec()
[all …]
/dpdk/drivers/net/avp/
H A Davp_ethdev.c1391 unsigned int count, avail, n; in avp_recv_scattered_pkts() local
1417 avail = avp_fifo_count(rx_q); in avp_recv_scattered_pkts()
1420 count = RTE_MIN(count, avail); in avp_recv_scattered_pkts()
1492 unsigned int count, avail, n; in avp_recv_pkts() local
1514 avail = avp_fifo_count(rx_q); in avp_recv_pkts()
1517 count = RTE_MIN(count, avail); in avp_recv_pkts()
1701 unsigned int count, avail, n; in avp_xmit_scattered_pkts() local
1725 avail = avp_fifo_count(alloc_q); in avp_xmit_scattered_pkts()
1822 unsigned int count, avail, n; in avp_xmit_pkts() local
1844 avail = avp_fifo_count(alloc_q); in avp_xmit_pkts()
[all …]
/dpdk/drivers/net/cxgbe/
H A Dsge.c221 q->avail--; in free_rx_bufs()
240 q->avail--; in unmap_rx_buf()
313 unsigned int cred = q->avail; in refill_fl_usembufs()
348 q->avail++; in refill_fl_usembufs()
356 out: cred = q->avail - cred; in refill_fl_usembufs()
417 reclaim_tx_desc(q, avail); in reclaim_completed_tx()
418 q->in_use -= avail; in reclaim_completed_tx()
419 avail = reclaimable(q); in reclaim_completed_tx()
420 } while (avail); in reclaim_completed_tx()
1899 fl->avail = 0; in t4_sge_alloc_rxq()
[all …]
/dpdk/drivers/baseband/acc100/
H A Drte_acc100_pmd.c3286 avail -= 1; in acc100_enqueue_enc_cb()
3340 if (unlikely(avail < 1)) in acc100_enqueue_ldpc_enc_cb()
3342 avail--; in acc100_enqueue_ldpc_enc_cb()
3394 avail -= cbs_in_tb; in acc100_enqueue_enc_tb()
3455 avail -= 1; in acc100_enqueue_dec_cb()
3509 avail -= cbs_in_tb; in acc100_enqueue_ldpc_dec_tb()
3539 if (unlikely(avail < 1)) in acc100_enqueue_ldpc_dec_cb()
3541 avail -= 1; in acc100_enqueue_ldpc_dec_cb()
3592 avail -= cbs_in_tb; in acc100_enqueue_dec_tb()
3970 dequeue_num = (avail < num) ? avail : num; in acc100_dequeue_enc()
[all …]

123