| /f-stack/dpdk/drivers/net/hns3/ |
| H A D | hns3_rxtx_vec_neon.h | 35 uint16_t next_to_use; in hns3_xmit_fixed_burst_vec() local 50 next_to_use = txq->next_to_use; in hns3_xmit_fixed_burst_vec() 51 tx_desc = &txq->tx_ring[next_to_use]; in hns3_xmit_fixed_burst_vec() 52 tx_entry = &txq->sw_ring[next_to_use]; in hns3_xmit_fixed_burst_vec() 59 n = txq->nb_tx_desc - next_to_use; in hns3_xmit_fixed_burst_vec() 67 next_to_use = 0; in hns3_xmit_fixed_burst_vec() 77 next_to_use += nb_commit; in hns3_xmit_fixed_burst_vec() 78 txq->next_to_use = next_to_use; in hns3_xmit_fixed_burst_vec() 130 uint16_t rx_id = rxq->next_to_use; in hns3_recv_burst_vec() 284 rxq->next_to_use += nb_rx; in hns3_recv_burst_vec() [all …]
|
| H A D | hns3_cmd.c | 20 int ntu = ring->next_to_use; in hns3_ring_space() 30 int ntu = ring->next_to_use; in is_valid_csq_clean_head() 203 csq->next_to_use, csq->next_to_clean); in hns3_cmd_csq_clean() 222 return head == hw->cmq.csq.next_to_use; in hns3_cmd_csq_done() 384 ntc = hw->cmq.csq.next_to_use; in hns3_cmd_send() 387 desc_to_use = &hw->cmq.csq.desc[hw->cmq.csq.next_to_use]; in hns3_cmd_send() 389 (hw->cmq.csq.next_to_use)++; in hns3_cmd_send() 390 if (hw->cmq.csq.next_to_use == hw->cmq.csq.desc_num) in hns3_cmd_send() 391 hw->cmq.csq.next_to_use = 0; in hns3_cmd_send() 509 hw->cmq.csq.next_to_use = 0; in hns3_cmd_init() [all …]
|
| H A D | hns3_rxtx_vec_sve.c | 84 uint16_t rx_id = rxq->next_to_use; in hns3_recv_burst_vec_sve() 228 rxq->next_to_use += nb_rx; in hns3_recv_burst_vec_sve() 229 if (rxq->next_to_use >= rxq->nb_rx_desc) in hns3_recv_burst_vec_sve() 230 rxq->next_to_use = 0; in hns3_recv_burst_vec_sve() 289 struct hns3_desc *rxdp = &rxq->rx_ring[rxq->next_to_use]; in hns3_recv_pkts_vec_sve() 305 hns3_rx_prefetch_mbuf_sve(&rxq->sw_ring[rxq->next_to_use]); in hns3_recv_pkts_vec_sve() 361 struct hns3_desc *txdp = &txq->tx_ring[txq->next_to_use]; in hns3_tx_fill_hw_ring_sve() 437 if (txq->next_to_use + nb_pkts > txq->nb_tx_desc) { in hns3_xmit_fixed_burst_vec_sve() 438 nb_tx = txq->nb_tx_desc - txq->next_to_use; in hns3_xmit_fixed_burst_vec_sve() 440 txq->next_to_use = 0; in hns3_xmit_fixed_burst_vec_sve() [all …]
|
| H A D | hns3_rxtx_vec.c | 106 struct hns3_desc *rxdp = &rxq->rx_ring[rxq->next_to_use]; in hns3_recv_pkts_vec() 122 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 0].mbuf); in hns3_recv_pkts_vec() 123 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 1].mbuf); in hns3_recv_pkts_vec() 124 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 2].mbuf); in hns3_recv_pkts_vec() 125 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 3].mbuf); in hns3_recv_pkts_vec()
|
| H A D | hns3_mbx.c | 200 return tail == hw->cmq.crq.next_to_use; in hns3_cmd_crq_empty() 373 desc = &crq->desc[crq->next_to_use]; in hns3_dev_handle_mbx_msg() 377 flag = rte_le_to_cpu_16(crq->desc[crq->next_to_use].flag); in hns3_dev_handle_mbx_msg() 384 crq->desc[crq->next_to_use].flag = 0; in hns3_dev_handle_mbx_msg() 436 crq->desc[crq->next_to_use].flag = 0; in hns3_dev_handle_mbx_msg() 441 hns3_write_dev(hw, HNS3_CMDQ_RX_HEAD_REG, crq->next_to_use); in hns3_dev_handle_mbx_msg()
|
| H A D | hns3_mbx.h | 158 ((crq)->next_to_use = ((crq)->next_to_use + 1) % (crq)->desc_num)
|
| H A D | hns3_rxtx.c | 42 for (i = rxq->next_to_use; in hns3_rx_queue_release_mbufs() 971 rxq->next_to_use = 0; in hns3_init_rxq() 990 rxq->next_to_use = 0; in hns3_init_fake_rxq() 1010 txq->next_to_use = 0; in hns3_init_txq() 1766 rxq->next_to_use = 0; in hns3_rx_queue_setup() 2100 rx_id = rxq->next_to_use; in hns3_recv_pkts() 2175 rxq->next_to_use = rx_id; in hns3_recv_pkts() 2216 rx_id = rxq->next_to_use; in hns3_recv_scattered_pkts() 2588 txq->next_to_use = 0; in hns3_tx_queue_setup() 3539 txq->next_to_use = 0; in hns3_xmit_pkts_simple() [all …]
|
| H A D | hns3_rxtx.h | 291 uint16_t next_to_use; /* index of next BD to be polled */ member 349 uint16_t next_to_use; member
|
| H A D | hns3_cmd.h | 34 uint32_t next_to_use; member
|
| /f-stack/dpdk/drivers/net/enetc/ |
| H A D | enetc.h | 63 int next_to_use; member 109 if (bdr->next_to_clean > bdr->next_to_use) in enetc_bd_unused() 110 return bdr->next_to_clean - bdr->next_to_use - 1; in enetc_bd_unused() 112 return bdr->bd_count + bdr->next_to_clean - bdr->next_to_use - 1; in enetc_bd_unused()
|
| H A D | enetc_rxtx.c | 89 i = tx_ring->next_to_use; in enetc_xmit_pkts() 120 tx_ring->next_to_use = i; in enetc_xmit_pkts() 134 i = rx_ring->next_to_use; in enetc_refill_rx_ring() 167 rx_ring->next_to_use = i; in enetc_refill_rx_ring()
|
| H A D | enetc_ethdev.c | 239 txr->next_to_use = 0; in enetc_alloc_txbdr() 386 rxr->next_to_use = 0; in enetc_alloc_rxbdr() 498 while (i != rx_ring->next_to_use) { in enetc_rx_queue_release()
|
| /f-stack/dpdk/drivers/net/ice/base/ |
| H A D | ice_controlq.h | 18 (u16)((((R)->next_to_clean > (R)->next_to_use) ? 0 : (R)->count) + \ 19 (R)->next_to_clean - (R)->next_to_use - 1) 54 u16 next_to_use; member
|
| H A D | ice_controlq.c | 336 cq->sq.next_to_use = 0; in ice_init_sq() 398 cq->rq.next_to_use = 0; in ice_init_rq() 870 return rd32(hw, cq->sq.head) == cq->sq.next_to_use; in ice_sq_done() 937 details = ICE_CTL_Q_DETAILS(cq->sq, cq->sq.next_to_use); in ice_sq_send_cmd_nolock() 955 desc_on_ring = ICE_CTL_Q_DESC(cq->sq, cq->sq.next_to_use); in ice_sq_send_cmd_nolock() 963 dma_buf = &cq->sq.r.sq_bi[cq->sq.next_to_use]; in ice_sq_send_cmd_nolock() 982 (cq->sq.next_to_use)++; in ice_sq_send_cmd_nolock() 983 if (cq->sq.next_to_use == cq->sq.count) in ice_sq_send_cmd_nolock() 984 cq->sq.next_to_use = 0; in ice_sq_send_cmd_nolock() 985 wr32(hw, cq->sq.tail, cq->sq.next_to_use); in ice_sq_send_cmd_nolock() [all …]
|
| /f-stack/dpdk/drivers/common/iavf/ |
| H A D | iavf_adminq.c | 341 hw->aq.asq.next_to_use = 0; in iavf_init_asq() 404 hw->aq.arq.next_to_use = 0; in iavf_init_arq() 628 return rd32(hw, hw->aq.asq.head) == hw->aq.asq.next_to_use; in iavf_asq_done() 676 details = IAVF_ADMINQ_DETAILS(hw->aq.asq, hw->aq.asq.next_to_use); in iavf_asq_send_command() 736 desc_on_ring = IAVF_ADMINQ_DESC(hw->aq.asq, hw->aq.asq.next_to_use); in iavf_asq_send_command() 744 dma_buff = &(hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use]); in iavf_asq_send_command() 763 (hw->aq.asq.next_to_use)++; in iavf_asq_send_command() 764 if (hw->aq.asq.next_to_use == hw->aq.asq.count) in iavf_asq_send_command() 765 hw->aq.asq.next_to_use = 0; in iavf_asq_send_command() 767 wr32(hw, hw->aq.asq.tail, hw->aq.asq.next_to_use); in iavf_asq_send_command() [all …]
|
| H A D | iavf_adminq.h | 31 u16 next_to_use; member
|
| H A D | iavf_type.h | 72 ((((R)->next_to_clean > (R)->next_to_use) ? 0 : (R)->count) + \ 73 (R)->next_to_clean - (R)->next_to_use - 1)
|
| /f-stack/dpdk/drivers/net/i40e/base/ |
| H A D | i40e_adminq.c | 392 hw->aq.asq.next_to_use = 0; in i40e_init_asq() 455 hw->aq.arq.next_to_use = 0; in i40e_init_arq() 560 hw->aq.asq.next_to_use = 0; in i40e_resume_aq() 565 hw->aq.arq.next_to_use = 0; in i40e_resume_aq() 832 return rd32(hw, hw->aq.asq.head) == hw->aq.asq.next_to_use; in i40e_asq_done() 948 dma_buff = &(hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use]); 967 (hw->aq.asq.next_to_use)++; 968 if (hw->aq.asq.next_to_use == hw->aq.asq.count) 969 hw->aq.asq.next_to_use = 0; 971 wr32(hw, hw->aq.asq.tail, hw->aq.asq.next_to_use); [all …]
|
| H A D | i40e_adminq.h | 31 u16 next_to_use; member
|
| H A D | i40e_type.h | 90 ((((R)->next_to_clean > (R)->next_to_use) ? 0 : (R)->count) + \ 91 (R)->next_to_clean - (R)->next_to_use - 1)
|
| /f-stack/dpdk/drivers/net/ena/ |
| H A D | ena_ethdev.c | 1233 ring->next_to_use = 0; in ena_queue_start() 1291 txq->next_to_use = 0; in ena_tx_queue_setup() 1387 rxq->next_to_use = 0; in ena_rx_queue_setup() 1456 uint16_t next_to_use = rxq->next_to_use; in ena_populate_rx_queue() local 1484 req_id = rxq->empty_rx_reqs[next_to_use]; in ena_populate_rx_queue() 1496 next_to_use = ENA_IDX_NEXT_MASKED(next_to_use, rxq->size_mask); in ena_populate_rx_queue() 1512 rxq->next_to_use = next_to_use; in ena_populate_rx_queue() 2483 uint16_t next_to_use; in ena_xmit_mbuf() local 2494 next_to_use = tx_ring->next_to_use; in ena_xmit_mbuf() 2496 req_id = tx_ring->empty_tx_reqs[next_to_use]; in ena_xmit_mbuf() [all …]
|
| H A D | ena_ethdev.h | 98 u16 next_to_use; member
|
| /f-stack/freebsd/contrib/ena-com/ |
| H A D | ena_com.h | 158 u16 next_to_use; member 1044 (bounce_buf_ctrl->next_to_use++ & (buffers_num - 1)) * size; in ena_com_get_next_bounce_buffer() 1047 (bounce_buf_ctrl->next_to_use & (buffers_num - 1)) * size); in ena_com_get_next_bounce_buffer()
|
| H A D | ena_com.c | 397 io_sq->bounce_buf_ctrl.next_to_use = 0; in ena_com_init_io_sq()
|
| /f-stack/dpdk/drivers/net/ena/base/ |
| H A D | ena_com.h | 130 u16 next_to_use; member 996 (bounce_buf_ctrl->next_to_use++ & (buffers_num - 1)) * size; in ena_com_get_next_bounce_buffer() 999 (bounce_buf_ctrl->next_to_use & (buffers_num - 1)) * size); in ena_com_get_next_bounce_buffer()
|