| /f-stack/dpdk/drivers/event/sw/ |
| H A D | iq_chunk.h | 63 iq->head_idx = 0; in iq_init() 89 iq->head_idx++; in iq_pop() 96 iq->head_idx = 0; in iq_pop() 103 return &iq->head->events[iq->head_idx]; in iq_peek() 119 index = iq->head_idx; in iq_dequeue_burst() 144 iq->head_idx = 0; in iq_dequeue_burst() 147 iq->head_idx = index; in iq_dequeue_burst() 166 uint16_t avail_space = iq->head_idx; in iq_put_back() 175 iq->head_idx = idx; in iq_put_back() 187 iq->head_idx = SW_EVS_PER_Q_CHUNK - remaining; in iq_put_back() [all …]
|
| H A D | sw_evdev.h | 99 uint16_t head_idx; member
|
| /f-stack/dpdk/drivers/net/enic/ |
| H A D | enic_rxtx.c | 447 int head_idx; in enic_xmit_pkts() local 461 head_idx = wq->head_idx; in enic_xmit_pkts() 496 desc_p = descs + head_idx; in enic_xmit_pkts() 547 head_idx = enic_ring_incr(desc_count, head_idx); in enic_xmit_pkts() 575 head_idx = enic_ring_incr(desc_count, head_idx); in enic_xmit_pkts() 585 wq->head_idx = head_idx; in enic_xmit_pkts() 656 head_idx = wq->head_idx; in enic_simple_xmit_pkts() 660 n = desc_count - head_idx; in enic_simple_xmit_pkts() 682 head_idx += nb_pkts; in enic_simple_xmit_pkts() 684 head_idx -= desc_count; in enic_simple_xmit_pkts() [all …]
|
| /f-stack/dpdk/drivers/net/enic/base/ |
| H A D | vnic_wq.c | 38 wq->head_idx = 0; in vnic_wq_alloc_bufs() 105 wq->head_idx = fetch_index; in vnic_wq_init_start() 106 wq->tail_idx = wq->head_idx; in vnic_wq_init_start() 165 wq->head_idx = 0; in vnic_wq_clean()
|
| H A D | vnic_wq.h | 46 unsigned int head_idx; member
|
| /f-stack/dpdk/drivers/net/ionic/ |
| H A D | ionic_dev.c | 440 q->head_idx = 0; in ionic_q_init() 464 writeq(IONIC_DBELL_QID(q->hw_index) | q->head_idx, q->db); in ionic_q_flush() 471 struct ionic_desc_info *head = &q->info[q->head_idx]; in ionic_q_post() 476 q->head_idx = (q->head_idx + 1) & (q->num_descs - 1); in ionic_q_post() 487 if (q->head_idx >= avail) in ionic_q_space_avail() 488 avail += q->num_descs - q->head_idx - 1; in ionic_q_space_avail() 490 avail -= q->head_idx + 1; in ionic_q_space_avail() 569 q_desc = &q_desc_base[adminq->head_idx]; in ionic_adminq_post()
|
| H A D | ionic_rxtx.c | 304 struct ionic_txq_desc *desc = &desc_base[q->head_idx]; in ionic_tx_tso_next() 305 struct ionic_txq_sg_desc *sg_desc = &sg_desc_base[q->head_idx]; in ionic_tx_tso_next() 439 struct ionic_txq_desc *desc = &desc_base[q->head_idx]; in ionic_tx() 440 struct ionic_txq_sg_desc *sg_desc = &sg_desc_base[q->head_idx]; in ionic_tx() 516 next_q_head_idx = (q->head_idx + 1) & (q->num_descs - 1); in ionic_xmit_pkts() 619 while (q->tail_idx != q->head_idx) { in ionic_rx_empty() 868 struct ionic_rxq_desc *new = &desc_base[q->head_idx]; in ionic_rx_recycle() 904 desc = &desc_base[q->head_idx]; in ionic_rx_fill() 914 sg_desc = &sg_desc_base[q->head_idx]; in ionic_rx_fill() 941 ring_doorbell = ((q->head_idx + 1) & in ionic_rx_fill()
|
| H A D | ionic_dev.h | 160 uint32_t head_idx; member
|
| /f-stack/dpdk/drivers/crypto/virtio/ |
| H A D | virtio_rxtx.c | 198 uint16_t head_idx; in virtqueue_crypto_sym_enqueue_xmit() local 223 head_idx = txvq->vq_desc_head_idx; in virtqueue_crypto_sym_enqueue_xmit() 224 if (unlikely(head_idx >= txvq->vq_nentries)) in virtqueue_crypto_sym_enqueue_xmit() 229 dxp = &txvq->vq_descx[head_idx]; in virtqueue_crypto_sym_enqueue_xmit() 329 start_dp[head_idx].addr = indirect_op_data_req_phys_addr + in virtqueue_crypto_sym_enqueue_xmit() 331 start_dp[head_idx].len = num_entry * sizeof(struct vring_desc); in virtqueue_crypto_sym_enqueue_xmit() 332 start_dp[head_idx].flags = VRING_DESC_F_INDIRECT; in virtqueue_crypto_sym_enqueue_xmit() 334 idx = start_dp[head_idx].next; in virtqueue_crypto_sym_enqueue_xmit() 339 vq_update_avail_ring(txvq, head_idx); in virtqueue_crypto_sym_enqueue_xmit()
|
| /f-stack/dpdk/lib/librte_eventdev/ |
| H A D | rte_event_timer_adapter.c | 464 size_t head_idx; in event_buffer_add() local 471 head_idx = bufp->head & EVENT_BUFFER_MASK; in event_buffer_add() 472 buf_eventp = &bufp->events[head_idx]; in event_buffer_add() 487 size_t head_idx, tail_idx; in event_buffer_flush() local 491 head_idx = bufp->head & EVENT_BUFFER_MASK; in event_buffer_flush() 494 RTE_ASSERT(head_idx < EVENT_BUFFER_SZ && tail_idx < EVENT_BUFFER_SZ); in event_buffer_flush() 499 if (head_idx > tail_idx) in event_buffer_flush() 500 n = head_idx - tail_idx; in event_buffer_flush() 501 else if (head_idx < tail_idx) in event_buffer_flush()
|
| /f-stack/dpdk/drivers/net/virtio/ |
| H A D | virtqueue.h | 696 uint16_t idx, id, head_idx, head_flags; in virtqueue_enqueue_xmit_packed() local 709 head_idx = vq->vq_avail_idx; in virtqueue_enqueue_xmit_packed() 710 idx = head_idx; in virtqueue_enqueue_xmit_packed() 711 prev = head_idx; in virtqueue_enqueue_xmit_packed() 775 if (likely(idx != head_idx)) { in virtqueue_enqueue_xmit_packed() 792 idx = head_idx; in virtqueue_enqueue_xmit_packed()
|
| H A D | virtio_rxtx.c | 258 uint16_t head_idx, idx, i = 0; in virtqueue_enqueue_refill_inorder() local 265 head_idx = vq->vq_desc_head_idx & (vq->vq_nentries - 1); in virtqueue_enqueue_refill_inorder() 269 idx = head_idx & (vq->vq_nentries - 1); in virtqueue_enqueue_refill_inorder() 284 head_idx++; in virtqueue_enqueue_refill_inorder() 536 uint16_t head_idx, idx; in virtqueue_enqueue_xmit() local 541 head_idx = vq->vq_desc_head_idx; in virtqueue_enqueue_xmit() 542 idx = head_idx; in virtqueue_enqueue_xmit() 605 idx = vq->vq_split.ring.desc[head_idx].next; in virtqueue_enqueue_xmit() 610 vq_update_avail_ring(vq, head_idx); in virtqueue_enqueue_xmit()
|
| H A D | virtio_rxtx_packed_avx.c | 522 uint16_t head_idx = vq->vq_avail_idx; in virtio_recv_refill_packed_vec() local 560 virtqueue_store_flags_packed(&start_dp[head_idx], head_flag, in virtio_recv_refill_packed_vec()
|
| /f-stack/dpdk/lib/librte_vhost/ |
| H A D | virtio_net.c | 161 uint16_t head_idx = vq->last_used_idx; in vhost_flush_enqueue_shadow_packed() local 200 head_idx = vq->last_used_idx; in vhost_flush_enqueue_shadow_packed() 207 vq->desc_packed[head_idx].flags = head_flags; in vhost_flush_enqueue_shadow_packed() 210 head_idx * in vhost_flush_enqueue_shadow_packed() 611 uint16_t head_idx = 0; in reserve_avail_buf_split() local 635 &head_idx, &len, in reserve_avail_buf_split() 639 update_shadow_used_ring_split(vq, head_idx, len); in reserve_avail_buf_split() 2163 uint16_t head_idx; in virtio_dev_tx_split() local 2171 &head_idx, &buf_len, in virtio_dev_tx_split() 2175 update_shadow_used_ring_split(vq, head_idx, 0); in virtio_dev_tx_split()
|