| /dpdk/drivers/net/mlx5/ |
| H A D | mlx5_rxtx_vec_neon.h | 52 pkts[pos] = elts[pos]; in rxq_copy_mbuf_v() 72 struct rte_mbuf **elts) in rxq_cq_decompress_v() argument 254 elts[pos]->hash.fdir.hi = in rxq_cq_decompress_v() 292 elts[pos]->packet_type = in rxq_cq_decompress_v() 295 elts[pos + 1]->packet_type = in rxq_cq_decompress_v() 305 elts[pos]->packet_type |= in rxq_cq_decompress_v() 331 elts[pos]->hash.rss = 0; in rxq_cq_decompress_v() 332 elts[pos + 1]->hash.rss = 0; in rxq_cq_decompress_v() 333 elts[pos + 2]->hash.rss = 0; in rxq_cq_decompress_v() 334 elts[pos + 3]->hash.rss = 0; in rxq_cq_decompress_v() [all …]
|
| H A D | mlx5_rxtx_vec.c | 94 struct rte_mbuf **elts = &(*rxq->elts)[elts_idx]; in mlx5_rx_replenish_bulk_mbuf() local 156 struct rte_mbuf **elts = &(*rxq->elts)[elts_idx]; in mlx5_rx_mprq_replenish_bulk_mbuf() local 205 struct rte_mbuf **elts = &(*rxq->elts)[elts_idx]; in rxq_copy_mprq_mbuf_v() local 224 if (!elts[i]->pkt_len) { in rxq_copy_mprq_mbuf_v() 234 rxq_code = mprq_buf_to_pkt(rxq, elts[i], elts[i]->pkt_len, in rxq_copy_mprq_mbuf_v() 295 struct rte_mbuf **elts; in rxq_burst_v() local 322 elts = &(*rxq->elts)[elts_idx]; in rxq_burst_v() 350 &elts[nocmp_n]); in rxq_burst_v() 435 struct rte_mbuf **elts; in rxq_burst_mprq_v() local 462 elts = &(*rxq->elts)[elts_idx]; in rxq_burst_mprq_v() [all …]
|
| H A D | mlx5_rxtx_vec_sse.h | 54 pkts[pos] = elts[pos]; in rxq_copy_mbuf_v() 74 struct rte_mbuf **elts) in rxq_cq_decompress_v() argument 239 elts[pos]->hash.fdir.hi = in rxq_cq_decompress_v() 278 elts[pos]->packet_type = in rxq_cq_decompress_v() 291 elts[pos]->packet_type |= in rxq_cq_decompress_v() 313 elts[pos]->ol_flags = in rxq_cq_decompress_v() 315 elts[pos + 1]->ol_flags = in rxq_cq_decompress_v() 317 elts[pos + 2]->ol_flags = in rxq_cq_decompress_v() 319 elts[pos + 3]->ol_flags = in rxq_cq_decompress_v() 321 elts[pos]->hash.rss = 0; in rxq_cq_decompress_v() [all …]
|
| H A D | mlx5_rxtx_vec_altivec.h | 53 (signed int const *)&elts[pos]); in rxq_copy_mbuf_v() 57 pkts[pos] = elts[pos]; in rxq_copy_mbuf_v() 77 struct rte_mbuf **elts) in rxq_cq_decompress_v() argument 353 elts[pos]->hash.fdir.hi = in rxq_cq_decompress_v() 407 elts[pos]->packet_type = in rxq_cq_decompress_v() 420 elts[pos]->packet_type |= in rxq_cq_decompress_v() 456 elts[pos]->ol_flags = in rxq_cq_decompress_v() 458 elts[pos + 1]->ol_flags = in rxq_cq_decompress_v() 460 elts[pos + 2]->ol_flags = in rxq_cq_decompress_v() 462 elts[pos + 3]->ol_flags = in rxq_cq_decompress_v() [all …]
|
| H A D | mlx5_flow_aso.c | 362 sq->elts[start_head & mask].burst_size = max; in mlx5_aso_sq_enqueue_burst() 370 sq->elts[sq->head & mask].pool = pool; in mlx5_aso_sq_enqueue_burst() 392 return sq->elts[start_head & mask].burst_size; in mlx5_aso_sq_enqueue_burst() 667 sq->elts[sq->head & mask].mtr = aso_mtr; in mlx5_aso_mtr_sq_enqueue_single() 904 sq->elts[sq->head & mask].ct = ct; in mlx5_aso_ct_sq_enqueue_single() 905 sq->elts[sq->head & mask].query_data = NULL; in mlx5_aso_ct_sq_enqueue_single() 1016 ct = sq->elts[idx].ct; in mlx5_aso_ct_status_update() 1019 if (sq->elts[idx].query_data) in mlx5_aso_ct_status_update() 1020 rte_memcpy(sq->elts[idx].query_data, in mlx5_aso_ct_status_update() 1072 sq->elts[wqe_idx].ct = ct; in mlx5_aso_ct_sq_query_single() [all …]
|
| H A D | mlx5_rx.c | 363 struct rte_mbuf *buf = (*rxq->elts)[i]; in mlx5_rxq_initialize() 493 elt = &(*rxq->elts)[elt_idx]; in mlx5_rx_err_handle() 499 elt = &(*rxq->elts) in mlx5_rx_err_handle() 508 elt = &(*rxq->elts)[i]; in mlx5_rx_err_handle() 515 (*rxq->elts)[elts_n + i] = in mlx5_rx_err_handle() 818 struct rte_mbuf *rep = (*rxq->elts)[idx]; in mlx5_rx_burst() 839 MLX5_ASSERT(pkt != (*rxq->elts)[idx]); in mlx5_rx_burst() 877 (*rxq->elts)[idx] = rep; in mlx5_rx_burst()
|
| H A D | mlx5_tx.h | 169 struct rte_mbuf *elts[0]; member 629 &txq->elts[txq->elts_tail & txq->elts_m], in mlx5_tx_free_elts() 657 struct rte_mbuf **elts = (struct rte_mbuf **)txq->elts; in mlx5_tx_copy_elts() local 671 rte_memcpy((void *)elts, (void *)(pkts + part), in mlx5_tx_copy_elts() 1592 txq->elts[txq->elts_head++ & txq->elts_m] = loc->mbuf; in mlx5_tx_mseg_build() 1618 txq->elts[txq->elts_head++ & txq->elts_m] = loc->mbuf; in mlx5_tx_mseg_build() 1904 txq->elts[txq->elts_head++ & txq->elts_m] = loc->mbuf; in mlx5_tx_packet_multi_send() 2299 txq->elts[txq->elts_head++ & txq->elts_m] = loc->mbuf; in mlx5_tx_burst_tso() 2928 txq->elts[txq->elts_head++ & txq->elts_m] = loc->mbuf; in mlx5_tx_burst_empw_inline() 3239 txq->elts[txq->elts_head++ & txq->elts_m] = in mlx5_tx_burst_single_send() [all …]
|
| H A D | mlx5_txq.c | 44 txq_ctrl->txq.elts[i] = NULL; in txq_alloc_elts() 65 struct rte_mbuf *(*elts)[elts_n] = &txq_ctrl->txq.elts; in txq_free_elts() local 74 struct rte_mbuf *elt = (*elts)[elts_tail & elts_m]; in txq_free_elts() 80 memset(&(*elts)[elts_tail & elts_m], in txq_free_elts() 82 sizeof((*elts)[elts_tail & elts_m])); in txq_free_elts()
|
| H A D | mlx5_rxq.c | 178 (*rxq_ctrl->rxq.elts)[i] = buf; in rxq_alloc_elts_sprq() 206 (*rxq->elts)[elts_n + j] = &rxq->fake_mbuf; in rxq_alloc_elts_sprq() 223 if ((*rxq_ctrl->rxq.elts)[i] != NULL) in rxq_alloc_elts_sprq() 224 rte_pktmbuf_free_seg((*rxq_ctrl->rxq.elts)[i]); in rxq_alloc_elts_sprq() 225 (*rxq_ctrl->rxq.elts)[i] = NULL; in rxq_alloc_elts_sprq() 314 if (rxq->elts == NULL) in rxq_free_elts_sprq() 322 (*rxq->elts)[(elts_ci + i) & q_mask] = NULL; in rxq_free_elts_sprq() 326 if ((*rxq->elts)[i] != NULL) in rxq_free_elts_sprq() 327 rte_pktmbuf_free_seg((*rxq->elts)[i]); in rxq_free_elts_sprq() 328 (*rxq->elts)[i] = NULL; in rxq_free_elts_sprq() [all …]
|
| H A D | mlx5_rx.h | 121 struct rte_mbuf *(*elts)[]; member
|
| H A D | mlx5.h | 571 struct mlx5_aso_sq_elem elts[1 << MLX5_ASO_QUEUE_LOG_DESC]; member
|
| /dpdk/drivers/net/mlx4/ |
| H A D | mlx4_rxq.c | 501 struct rte_mbuf *(*elts)[elts_n] = rxq->elts; in mlx4_rxq_attach() local 579 (*elts)[i] = NULL; in mlx4_rxq_attach() 603 (*elts)[i] = buf; in mlx4_rxq_attach() 652 struct rte_mbuf *(*elts)[1 << rxq->elts_n] = rxq->elts; in mlx4_rxq_detach() local 666 if (!(*elts)[i]) in mlx4_rxq_detach() 669 (*elts)[i] = NULL; in mlx4_rxq_detach() 748 .size = sizeof(*elts), in mlx4_rx_queue_setup() 749 .addr = (void **)&elts, in mlx4_rx_queue_setup() 780 if (desc != RTE_DIM(*elts)) { in mlx4_rx_queue_setup() 781 desc = RTE_DIM(*elts); in mlx4_rx_queue_setup() [all …]
|
| H A D | mlx4_txq.c | 209 struct txq_elt (*elts)[txq->elts_n] = txq->elts; in mlx4_txq_free_elts() local 214 struct txq_elt *elt = &(*elts)[n]; in mlx4_txq_free_elts() 319 struct txq_elt (*elts)[rte_align32pow2(desc)]; in mlx4_tx_queue_setup() local 331 .size = sizeof(*elts), in mlx4_tx_queue_setup() 332 .addr = (void **)&elts, in mlx4_tx_queue_setup() 364 if (desc != RTE_DIM(*elts)) { in mlx4_tx_queue_setup() 365 desc = RTE_DIM(*elts); in mlx4_tx_queue_setup() 385 .elts = elts, in mlx4_tx_queue_setup() 501 (&(*txq->elts)[0])->wqe = in mlx4_tx_queue_setup()
|
| H A D | mlx4_rxtx.c | 345 first_txbb = (&(*txq->elts)[elts_tail & elts_m])->eocb; in mlx4_txq_complete() 349 (&(*txq->elts)[elts_tail & elts_m])->eocb); in mlx4_txq_complete() 903 elt = &(*txq->elts)[elts_head & elts_m]; in mlx4_tx_burst() 908 struct txq_elt *elt_next = &(*txq->elts)[++elts_head & elts_m]; in mlx4_tx_burst() 1230 struct rte_mbuf *rep = (*rxq->elts)[idx]; in mlx4_rx_burst() 1250 MLX4_ASSERT(pkt != (*rxq->elts)[idx]); in mlx4_rx_burst() 1298 (*rxq->elts)[idx] = rep; in mlx4_rx_burst()
|
| H A D | mlx4_rxtx.h | 51 struct rte_mbuf *(*elts)[]; /**< Rx elements. */ member 107 struct txq_elt (*elts)[]; /**< Tx elements. */ member
|