| /dpdk/drivers/crypto/virtio/ |
| H A D | virtqueue.h | 117 vq->vq_ring.avail->idx = vq->vq_avail_idx; in vq_update_avail_idx() 132 if (unlikely(vq->vq_ring.avail->ring[avail_idx] != desc_idx)) in vq_update_avail_ring() 133 vq->vq_ring.avail->ring[avail_idx] = desc_idx; in vq_update_avail_ring() 166 (vq)->vq_desc_head_idx, (vq)->vq_ring.avail->idx, \ 168 (vq)->vq_ring.avail->flags, (vq)->vq_ring.used->flags); \
|
| H A D | virtio_ring.h | 63 struct vring_avail *avail; member 97 #define vring_used_event(vr) ((vr)->avail->ring[(vr)->num]) 119 vr->avail = (struct vring_avail *) (p + in vring_init() 122 RTE_ALIGN_CEIL((uintptr_t)(&vr->avail->ring[num]), align); in vring_init()
|
| H A D | virtqueue.c | 21 vq->vq_ring.avail->flags |= VRING_AVAIL_F_NO_INTERRUPT; in virtqueue_disable_intr()
|
| /dpdk/drivers/bus/vmbus/ |
| H A D | vmbus_bufring.c | 124 uint32_t avail; in vmbus_txbr_write() local 132 avail = vmbus_br_availwrite(tbr, old_windex); in vmbus_txbr_write() 135 if (avail <= total) in vmbus_txbr_write() 195 uint32_t avail; in vmbus_rxbr_peek() local 201 avail = vmbus_br_availread(rbr); in vmbus_rxbr_peek() 202 if (avail < dlen + sizeof(uint64_t)) in vmbus_rxbr_peek()
|
| /dpdk/lib/ring/ |
| H A D | rte_ring_peek_elem_pvt.h | 151 uint32_t avail, head, next; in __rte_ring_do_dequeue_start() local 156 behavior, &head, &next, &avail); in __rte_ring_do_dequeue_start() 160 &head, &avail); in __rte_ring_do_dequeue_start() 168 avail = 0; in __rte_ring_do_dequeue_start() 175 *available = avail - n; in __rte_ring_do_dequeue_start()
|
| H A D | rte_ring_peek_zc.h | 335 uint32_t avail, head, next; in __rte_ring_do_dequeue_zc_elem_start() local 340 behavior, &head, &next, &avail); in __rte_ring_do_dequeue_zc_elem_start() 344 &head, &avail); in __rte_ring_do_dequeue_zc_elem_start() 352 avail = 0; in __rte_ring_do_dequeue_zc_elem_start() 360 *available = avail - n; in __rte_ring_do_dequeue_zc_elem_start()
|
| /dpdk/drivers/net/virtio/ |
| H A D | virtio_ring.h | 94 struct vring_avail *avail; member 128 #define vring_used_event(vr) ((vr)->avail->ring[(vr)->num]) 157 vr->avail = (struct vring_avail *) (p + in vring_init_split() 160 RTE_ALIGN_CEIL((uintptr_t)(&vr->avail->ring[num]), align); in vring_init_split()
|
| H A D | virtqueue.h | 367 uint16_t used, avail, flags; in desc_is_used() local 371 avail = !!(flags & VRING_PACKED_DESC_F_AVAIL); in desc_is_used() 373 return avail == used && used == vq->vq_packed.used_wrap_counter; in desc_is_used() 428 vq->vq_split.ring.avail->flags |= VRING_AVAIL_F_NO_INTERRUPT; in virtqueue_disable_intr_split() 462 vq->vq_split.ring.avail->flags &= (~VRING_AVAIL_F_NO_INTERRUPT); in virtqueue_enable_intr_split() 555 vq->vq_split.ring.avail->idx = vq->vq_avail_idx; in vq_update_avail_idx() 557 __atomic_store_n(&vq->vq_split.ring.avail->idx, in vq_update_avail_idx() 562 vq->vq_split.ring.avail->idx = vq->vq_avail_idx; in vq_update_avail_idx() 578 if (unlikely(vq->vq_split.ring.avail->ring[avail_idx] != desc_idx)) in vq_update_avail_ring() 579 vq->vq_split.ring.avail->ring[avail_idx] = desc_idx; in vq_update_avail_ring() [all …]
|
| /dpdk/drivers/bus/dpaa/base/qbman/ |
| H A D | bman.c | 241 static void update_rcr_ci(struct bman_portal *p, int avail) in update_rcr_ci() argument 243 if (avail) in update_rcr_ci() 256 u8 avail; in bman_release() local 266 avail = bm_rcr_get_avail(&p->p); in bman_release() 267 if (avail < 2) in bman_release() 268 update_rcr_ci(p, avail); in bman_release()
|
| /dpdk/drivers/net/ionic/ |
| H A D | ionic_dev.h | 231 uint16_t avail = q->tail_idx; in ionic_q_space_avail() local 233 if (q->head_idx >= avail) in ionic_q_space_avail() 234 avail += q->num_descs - q->head_idx - 1; in ionic_q_space_avail() 236 avail -= q->head_idx + 1; in ionic_q_space_avail() 238 return avail; in ionic_q_space_avail()
|
| /dpdk/drivers/net/vmxnet3/ |
| H A D | vmxnet3_rxtx.c | 73 uint32_t avail = 0; in vmxnet3_rxq_dump() local 90 (uint32_t)rxq->cmd_ring[0].size, avail, in vmxnet3_rxq_dump() 92 rxq->cmd_ring[0].size - avail); in vmxnet3_rxq_dump() 97 rxq->cmd_ring[1].size - avail); in vmxnet3_rxq_dump() 104 uint32_t avail = 0; in vmxnet3_txq_dump() local 118 (uint32_t)txq->cmd_ring.size, avail, in vmxnet3_txq_dump() 417 uint32_t first2fill, avail, dw2; in vmxnet3_xmit_pkts() local 426 if (count > avail) { in vmxnet3_xmit_pkts() 428 if (unlikely(avail == 0)) { in vmxnet3_xmit_pkts() 1003 uint32_t avail; in vmxnet3_recv_pkts() local [all …]
|
| /dpdk/app/test/ |
| H A D | test_ring_mpmc_stress.c | 9 uint32_t *avail) in _st_ring_dequeue_bulk() argument 11 return rte_ring_mc_dequeue_bulk(r, obj, n, avail); in _st_ring_dequeue_bulk()
|
| H A D | test_ring_rts_stress.c | 9 uint32_t *avail) in _st_ring_dequeue_bulk() argument 11 return rte_ring_mc_rts_dequeue_bulk(r, obj, n, avail); in _st_ring_dequeue_bulk()
|
| H A D | test_ring_hts_stress.c | 9 uint32_t *avail) in _st_ring_dequeue_bulk() argument 11 return rte_ring_mc_hts_dequeue_bulk(r, obj, n, avail); in _st_ring_dequeue_bulk()
|
| H A D | test_ring_mt_peek_stress.c | 10 uint32_t *avail) in _st_ring_dequeue_bulk() argument 14 m = rte_ring_dequeue_bulk_start(r, obj, n, avail); in _st_ring_dequeue_bulk()
|
| H A D | test_ring_mt_peek_stress_zc.c | 11 uint32_t *avail) in _st_ring_dequeue_bulk() argument 16 m = rte_ring_dequeue_zc_bulk_start(r, n, &zcd, avail); in _st_ring_dequeue_bulk()
|
| H A D | test_ring_st_peek_stress.c | 10 uint32_t *avail) in _st_ring_dequeue_bulk() argument 18 m = rte_ring_dequeue_bulk_start(r, obj, n, avail); in _st_ring_dequeue_bulk()
|
| H A D | test_ring_st_peek_stress_zc.c | 11 uint32_t *avail) in _st_ring_dequeue_bulk() argument 20 m = rte_ring_dequeue_zc_bulk_start(r, n, &zcd, avail); in _st_ring_dequeue_bulk()
|
| /dpdk/drivers/mempool/octeontx/ |
| H A D | octeontx_fpavf.c | 494 uint64_t cnt, limit, avail; in octeontx_fpa_bufpool_free_count() local 515 avail = fpavf_read64((void *)((uintptr_t)pool_bar + in octeontx_fpa_bufpool_free_count() 518 return RTE_MIN(avail, (limit - cnt)); in octeontx_fpa_bufpool_free_count() 613 uint64_t cnt, avail; in octeontx_fpa_bufpool_destroy() local 645 avail = fpavf_read64((void *)((uintptr_t)pool_bar + in octeontx_fpa_bufpool_destroy() 649 fpavf_write64(avail, (void *)((uintptr_t)pool_bar + in octeontx_fpa_bufpool_destroy() 651 fpavf_write64(avail + 1, (void *)((uintptr_t)pool_bar + in octeontx_fpa_bufpool_destroy() 659 while (avail--) { in octeontx_fpa_bufpool_destroy() 668 gaura, avail); in octeontx_fpa_bufpool_destroy()
|
| /dpdk/examples/vhost/ |
| H A D | virtio_net.c | 200 avail_idx = __atomic_load_n(&vr->avail->idx, __ATOMIC_ACQUIRE); in vs_enqueue_pkts() 209 rte_prefetch0(&vr->avail->ring[start_idx & (vr->size - 1)]); in vs_enqueue_pkts() 212 desc_indexes[i] = vr->avail->ring[used_idx]; in vs_enqueue_pkts() 381 free_entries = __atomic_load_n(&vr->avail->idx, __ATOMIC_ACQUIRE) - in vs_dequeue_pkts() 389 rte_prefetch0(&vr->avail->ring[avail_idx]); in vs_dequeue_pkts() 405 desc_indexes[i] = vr->avail->ring[avail_idx]; in vs_dequeue_pkts()
|
| /dpdk/drivers/net/ice/base/ |
| H A D | ice_acl_ctrl.c | 414 ice_bitmap_set(tbl->avail, first_e, last_e - first_e + 1); in ice_acl_create_tbl() 490 bool avail = true; in ice_acl_alloc_partition() local 498 for (w = row; w < row + width && avail; w++) { in ice_acl_alloc_partition() 502 avail &= ice_is_bit_set(hw->acl_tbl->avail, b); in ice_acl_alloc_partition() 505 if (!avail) { in ice_acl_alloc_partition() 723 ice_set_bit(b, hw->acl_tbl->avail); in ice_acl_commit_partition() 725 ice_clear_bit(b, hw->acl_tbl->avail); in ice_acl_commit_partition()
|
| /dpdk/drivers/baseband/fpga_lte_fec/ |
| H A D | fpga_lte_fec.c | 1945 int32_t avail; in fpga_enqueue_enc() local 1956 avail = (q->head_free_desc > q->tail) ? in fpga_enqueue_enc() 1967 if (unlikely(avail - cbs_in_op < 0)) in fpga_enqueue_enc() 1969 avail -= cbs_in_op; in fpga_enqueue_enc() 1976 if (unlikely(avail - 1 < 0)) in fpga_enqueue_enc() 1978 avail -= 1; in fpga_enqueue_enc() 2015 int32_t avail; in fpga_enqueue_dec() local 2037 if (unlikely(avail - cbs_in_op < 0)) in fpga_enqueue_dec() 2039 avail -= cbs_in_op; in fpga_enqueue_dec() 2046 if (unlikely(avail - 1 < 0)) in fpga_enqueue_dec() [all …]
|
| /dpdk/drivers/net/avp/ |
| H A D | avp_ethdev.c | 1391 unsigned int count, avail, n; in avp_recv_scattered_pkts() local 1417 avail = avp_fifo_count(rx_q); in avp_recv_scattered_pkts() 1420 count = RTE_MIN(count, avail); in avp_recv_scattered_pkts() 1492 unsigned int count, avail, n; in avp_recv_pkts() local 1514 avail = avp_fifo_count(rx_q); in avp_recv_pkts() 1517 count = RTE_MIN(count, avail); in avp_recv_pkts() 1701 unsigned int count, avail, n; in avp_xmit_scattered_pkts() local 1725 avail = avp_fifo_count(alloc_q); in avp_xmit_scattered_pkts() 1822 unsigned int count, avail, n; in avp_xmit_pkts() local 1844 avail = avp_fifo_count(alloc_q); in avp_xmit_pkts() [all …]
|
| /dpdk/drivers/net/cxgbe/ |
| H A D | sge.c | 221 q->avail--; in free_rx_bufs() 240 q->avail--; in unmap_rx_buf() 313 unsigned int cred = q->avail; in refill_fl_usembufs() 348 q->avail++; in refill_fl_usembufs() 356 out: cred = q->avail - cred; in refill_fl_usembufs() 417 reclaim_tx_desc(q, avail); in reclaim_completed_tx() 418 q->in_use -= avail; in reclaim_completed_tx() 419 avail = reclaimable(q); in reclaim_completed_tx() 420 } while (avail); in reclaim_completed_tx() 1899 fl->avail = 0; in t4_sge_alloc_rxq() [all …]
|
| /dpdk/drivers/baseband/acc100/ |
| H A D | rte_acc100_pmd.c | 3286 avail -= 1; in acc100_enqueue_enc_cb() 3340 if (unlikely(avail < 1)) in acc100_enqueue_ldpc_enc_cb() 3342 avail--; in acc100_enqueue_ldpc_enc_cb() 3394 avail -= cbs_in_tb; in acc100_enqueue_enc_tb() 3455 avail -= 1; in acc100_enqueue_dec_cb() 3509 avail -= cbs_in_tb; in acc100_enqueue_ldpc_dec_tb() 3539 if (unlikely(avail < 1)) in acc100_enqueue_ldpc_dec_cb() 3541 avail -= 1; in acc100_enqueue_ldpc_dec_cb() 3592 avail -= cbs_in_tb; in acc100_enqueue_dec_tb() 3970 dequeue_num = (avail < num) ? avail : num; in acc100_dequeue_enc() [all …]
|