Home
last modified time | relevance | path

Searched refs:elts (Results 1 – 15 of 15) sorted by relevance

/dpdk/drivers/net/mlx5/
H A Dmlx5_rxtx_vec_neon.h52 pkts[pos] = elts[pos]; in rxq_copy_mbuf_v()
72 struct rte_mbuf **elts) in rxq_cq_decompress_v() argument
254 elts[pos]->hash.fdir.hi = in rxq_cq_decompress_v()
292 elts[pos]->packet_type = in rxq_cq_decompress_v()
295 elts[pos + 1]->packet_type = in rxq_cq_decompress_v()
305 elts[pos]->packet_type |= in rxq_cq_decompress_v()
331 elts[pos]->hash.rss = 0; in rxq_cq_decompress_v()
332 elts[pos + 1]->hash.rss = 0; in rxq_cq_decompress_v()
333 elts[pos + 2]->hash.rss = 0; in rxq_cq_decompress_v()
334 elts[pos + 3]->hash.rss = 0; in rxq_cq_decompress_v()
[all …]
H A Dmlx5_rxtx_vec.c94 struct rte_mbuf **elts = &(*rxq->elts)[elts_idx]; in mlx5_rx_replenish_bulk_mbuf() local
156 struct rte_mbuf **elts = &(*rxq->elts)[elts_idx]; in mlx5_rx_mprq_replenish_bulk_mbuf() local
205 struct rte_mbuf **elts = &(*rxq->elts)[elts_idx]; in rxq_copy_mprq_mbuf_v() local
224 if (!elts[i]->pkt_len) { in rxq_copy_mprq_mbuf_v()
234 rxq_code = mprq_buf_to_pkt(rxq, elts[i], elts[i]->pkt_len, in rxq_copy_mprq_mbuf_v()
295 struct rte_mbuf **elts; in rxq_burst_v() local
322 elts = &(*rxq->elts)[elts_idx]; in rxq_burst_v()
350 &elts[nocmp_n]); in rxq_burst_v()
435 struct rte_mbuf **elts; in rxq_burst_mprq_v() local
462 elts = &(*rxq->elts)[elts_idx]; in rxq_burst_mprq_v()
[all …]
H A Dmlx5_rxtx_vec_sse.h54 pkts[pos] = elts[pos]; in rxq_copy_mbuf_v()
74 struct rte_mbuf **elts) in rxq_cq_decompress_v() argument
239 elts[pos]->hash.fdir.hi = in rxq_cq_decompress_v()
278 elts[pos]->packet_type = in rxq_cq_decompress_v()
291 elts[pos]->packet_type |= in rxq_cq_decompress_v()
313 elts[pos]->ol_flags = in rxq_cq_decompress_v()
315 elts[pos + 1]->ol_flags = in rxq_cq_decompress_v()
317 elts[pos + 2]->ol_flags = in rxq_cq_decompress_v()
319 elts[pos + 3]->ol_flags = in rxq_cq_decompress_v()
321 elts[pos]->hash.rss = 0; in rxq_cq_decompress_v()
[all …]
H A Dmlx5_rxtx_vec_altivec.h53 (signed int const *)&elts[pos]); in rxq_copy_mbuf_v()
57 pkts[pos] = elts[pos]; in rxq_copy_mbuf_v()
77 struct rte_mbuf **elts) in rxq_cq_decompress_v() argument
353 elts[pos]->hash.fdir.hi = in rxq_cq_decompress_v()
407 elts[pos]->packet_type = in rxq_cq_decompress_v()
420 elts[pos]->packet_type |= in rxq_cq_decompress_v()
456 elts[pos]->ol_flags = in rxq_cq_decompress_v()
458 elts[pos + 1]->ol_flags = in rxq_cq_decompress_v()
460 elts[pos + 2]->ol_flags = in rxq_cq_decompress_v()
462 elts[pos + 3]->ol_flags = in rxq_cq_decompress_v()
[all …]
H A Dmlx5_flow_aso.c362 sq->elts[start_head & mask].burst_size = max; in mlx5_aso_sq_enqueue_burst()
370 sq->elts[sq->head & mask].pool = pool; in mlx5_aso_sq_enqueue_burst()
392 return sq->elts[start_head & mask].burst_size; in mlx5_aso_sq_enqueue_burst()
667 sq->elts[sq->head & mask].mtr = aso_mtr; in mlx5_aso_mtr_sq_enqueue_single()
904 sq->elts[sq->head & mask].ct = ct; in mlx5_aso_ct_sq_enqueue_single()
905 sq->elts[sq->head & mask].query_data = NULL; in mlx5_aso_ct_sq_enqueue_single()
1016 ct = sq->elts[idx].ct; in mlx5_aso_ct_status_update()
1019 if (sq->elts[idx].query_data) in mlx5_aso_ct_status_update()
1020 rte_memcpy(sq->elts[idx].query_data, in mlx5_aso_ct_status_update()
1072 sq->elts[wqe_idx].ct = ct; in mlx5_aso_ct_sq_query_single()
[all …]
H A Dmlx5_rx.c363 struct rte_mbuf *buf = (*rxq->elts)[i]; in mlx5_rxq_initialize()
493 elt = &(*rxq->elts)[elt_idx]; in mlx5_rx_err_handle()
499 elt = &(*rxq->elts) in mlx5_rx_err_handle()
508 elt = &(*rxq->elts)[i]; in mlx5_rx_err_handle()
515 (*rxq->elts)[elts_n + i] = in mlx5_rx_err_handle()
818 struct rte_mbuf *rep = (*rxq->elts)[idx]; in mlx5_rx_burst()
839 MLX5_ASSERT(pkt != (*rxq->elts)[idx]); in mlx5_rx_burst()
877 (*rxq->elts)[idx] = rep; in mlx5_rx_burst()
H A Dmlx5_tx.h169 struct rte_mbuf *elts[0]; member
629 &txq->elts[txq->elts_tail & txq->elts_m], in mlx5_tx_free_elts()
657 struct rte_mbuf **elts = (struct rte_mbuf **)txq->elts; in mlx5_tx_copy_elts() local
671 rte_memcpy((void *)elts, (void *)(pkts + part), in mlx5_tx_copy_elts()
1592 txq->elts[txq->elts_head++ & txq->elts_m] = loc->mbuf; in mlx5_tx_mseg_build()
1618 txq->elts[txq->elts_head++ & txq->elts_m] = loc->mbuf; in mlx5_tx_mseg_build()
1904 txq->elts[txq->elts_head++ & txq->elts_m] = loc->mbuf; in mlx5_tx_packet_multi_send()
2299 txq->elts[txq->elts_head++ & txq->elts_m] = loc->mbuf; in mlx5_tx_burst_tso()
2928 txq->elts[txq->elts_head++ & txq->elts_m] = loc->mbuf; in mlx5_tx_burst_empw_inline()
3239 txq->elts[txq->elts_head++ & txq->elts_m] = in mlx5_tx_burst_single_send()
[all …]
H A Dmlx5_txq.c44 txq_ctrl->txq.elts[i] = NULL; in txq_alloc_elts()
65 struct rte_mbuf *(*elts)[elts_n] = &txq_ctrl->txq.elts; in txq_free_elts() local
74 struct rte_mbuf *elt = (*elts)[elts_tail & elts_m]; in txq_free_elts()
80 memset(&(*elts)[elts_tail & elts_m], in txq_free_elts()
82 sizeof((*elts)[elts_tail & elts_m])); in txq_free_elts()
H A Dmlx5_rxq.c178 (*rxq_ctrl->rxq.elts)[i] = buf; in rxq_alloc_elts_sprq()
206 (*rxq->elts)[elts_n + j] = &rxq->fake_mbuf; in rxq_alloc_elts_sprq()
223 if ((*rxq_ctrl->rxq.elts)[i] != NULL) in rxq_alloc_elts_sprq()
224 rte_pktmbuf_free_seg((*rxq_ctrl->rxq.elts)[i]); in rxq_alloc_elts_sprq()
225 (*rxq_ctrl->rxq.elts)[i] = NULL; in rxq_alloc_elts_sprq()
314 if (rxq->elts == NULL) in rxq_free_elts_sprq()
322 (*rxq->elts)[(elts_ci + i) & q_mask] = NULL; in rxq_free_elts_sprq()
326 if ((*rxq->elts)[i] != NULL) in rxq_free_elts_sprq()
327 rte_pktmbuf_free_seg((*rxq->elts)[i]); in rxq_free_elts_sprq()
328 (*rxq->elts)[i] = NULL; in rxq_free_elts_sprq()
[all …]
H A Dmlx5_rx.h121 struct rte_mbuf *(*elts)[]; member
H A Dmlx5.h571 struct mlx5_aso_sq_elem elts[1 << MLX5_ASO_QUEUE_LOG_DESC]; member
/dpdk/drivers/net/mlx4/
H A Dmlx4_rxq.c501 struct rte_mbuf *(*elts)[elts_n] = rxq->elts; in mlx4_rxq_attach() local
579 (*elts)[i] = NULL; in mlx4_rxq_attach()
603 (*elts)[i] = buf; in mlx4_rxq_attach()
652 struct rte_mbuf *(*elts)[1 << rxq->elts_n] = rxq->elts; in mlx4_rxq_detach() local
666 if (!(*elts)[i]) in mlx4_rxq_detach()
669 (*elts)[i] = NULL; in mlx4_rxq_detach()
748 .size = sizeof(*elts), in mlx4_rx_queue_setup()
749 .addr = (void **)&elts, in mlx4_rx_queue_setup()
780 if (desc != RTE_DIM(*elts)) { in mlx4_rx_queue_setup()
781 desc = RTE_DIM(*elts); in mlx4_rx_queue_setup()
[all …]
H A Dmlx4_txq.c209 struct txq_elt (*elts)[txq->elts_n] = txq->elts; in mlx4_txq_free_elts() local
214 struct txq_elt *elt = &(*elts)[n]; in mlx4_txq_free_elts()
319 struct txq_elt (*elts)[rte_align32pow2(desc)]; in mlx4_tx_queue_setup() local
331 .size = sizeof(*elts), in mlx4_tx_queue_setup()
332 .addr = (void **)&elts, in mlx4_tx_queue_setup()
364 if (desc != RTE_DIM(*elts)) { in mlx4_tx_queue_setup()
365 desc = RTE_DIM(*elts); in mlx4_tx_queue_setup()
385 .elts = elts, in mlx4_tx_queue_setup()
501 (&(*txq->elts)[0])->wqe = in mlx4_tx_queue_setup()
H A Dmlx4_rxtx.c345 first_txbb = (&(*txq->elts)[elts_tail & elts_m])->eocb; in mlx4_txq_complete()
349 (&(*txq->elts)[elts_tail & elts_m])->eocb); in mlx4_txq_complete()
903 elt = &(*txq->elts)[elts_head & elts_m]; in mlx4_tx_burst()
908 struct txq_elt *elt_next = &(*txq->elts)[++elts_head & elts_m]; in mlx4_tx_burst()
1230 struct rte_mbuf *rep = (*rxq->elts)[idx]; in mlx4_rx_burst()
1250 MLX4_ASSERT(pkt != (*rxq->elts)[idx]); in mlx4_rx_burst()
1298 (*rxq->elts)[idx] = rep; in mlx4_rx_burst()
H A Dmlx4_rxtx.h51 struct rte_mbuf *(*elts)[]; /**< Rx elements. */ member
107 struct txq_elt (*elts)[]; /**< Tx elements. */ member