| /dpdk/drivers/net/nfp/ |
| H A D | nfp_rxtx.c | 494 const struct rte_memzone *tz; in nfp_net_rx_queue_setup() local 555 tz = rte_eth_dma_zone_reserve(dev, "rx_ring", queue_idx, in nfp_net_rx_queue_setup() 560 if (tz == NULL) { in nfp_net_rx_queue_setup() 568 rxq->dma = (uint64_t)tz->iova; in nfp_net_rx_queue_setup() 569 rxq->rxds = (struct nfp_net_rx_desc *)tz->addr; in nfp_net_rx_queue_setup() 683 const struct rte_memzone *tz; in nfp_net_tx_queue_setup() local 741 tz = rte_eth_dma_zone_reserve(dev, "tx_ring", queue_idx, in nfp_net_tx_queue_setup() 745 if (tz == NULL) { in nfp_net_tx_queue_setup() 766 txq->dma = (uint64_t)tz->iova; in nfp_net_tx_queue_setup() 767 txq->txds = (struct nfp_net_tx_desc *)tz->addr; in nfp_net_tx_queue_setup()
|
| /dpdk/drivers/net/bnx2x/ |
| H A D | bnx2x_rxtx.c | 239 const struct rte_memzone *tz; in bnx2x_dev_tx_queue_setup() local 273 tz = ring_dma_zone_reserve(dev, "tx_hw_ring", queue_idx, tsize, socket_id); in bnx2x_dev_tx_queue_setup() 274 if (tz == NULL) { in bnx2x_dev_tx_queue_setup() 278 fp->tx_desc_mapping = txq->tx_ring_phys_addr = (uint64_t)tz->iova; in bnx2x_dev_tx_queue_setup() 279 txq->tx_ring = (union eth_tx_bd_types *) tz->addr; in bnx2x_dev_tx_queue_setup()
|
| /dpdk/drivers/net/axgbe/ |
| H A D | axgbe_rxtx.c | 531 const struct rte_memzone *tz; in axgbe_dev_tx_queue_setup() local 570 tz = rte_eth_dma_zone_reserve(dev, "tx_ring", queue_idx, in axgbe_dev_tx_queue_setup() 572 if (!tz) { in axgbe_dev_tx_queue_setup() 576 memset(tz->addr, 0, tsize); in axgbe_dev_tx_queue_setup() 577 txq->ring_phys_addr = (uint64_t)tz->iova; in axgbe_dev_tx_queue_setup() 578 txq->desc = tz->addr; in axgbe_dev_tx_queue_setup()
|
| /dpdk/drivers/net/i40e/ |
| H A D | i40e_rxtx.c | 2277 const struct rte_memzone *tz; in i40e_dev_tx_queue_setup() local 2411 if (!tz) { in i40e_dev_tx_queue_setup() 2417 txq->mz = tz; in i40e_dev_tx_queue_setup() 2431 txq->tx_ring_phys_addr = tz->iova; in i40e_dev_tx_queue_setup() 2432 txq->tx_ring = (struct i40e_tx_desc *)tz->addr; in i40e_dev_tx_queue_setup() 3048 const struct rte_memzone *tz = NULL; in i40e_fdir_setup_tx_resources() local 3074 tz = rte_eth_dma_zone_reserve(dev, "fdir_tx_ring", in i40e_fdir_setup_tx_resources() 3077 if (!tz) { in i40e_fdir_setup_tx_resources() 3083 txq->mz = tz; in i40e_fdir_setup_tx_resources() 3089 txq->tx_ring_phys_addr = tz->iova; in i40e_fdir_setup_tx_resources() [all …]
|
| /dpdk/drivers/net/e1000/ |
| H A D | em_rxtx.c | 1204 const struct rte_memzone *tz; in eth_em_tx_queue_setup() local 1281 tz = rte_eth_dma_zone_reserve(dev, "tx_ring", queue_idx, tsize, in eth_em_tx_queue_setup() 1283 if (tz == NULL) in eth_em_tx_queue_setup() 1291 txq->mz = tz; in eth_em_tx_queue_setup() 1310 txq->tx_ring_phys_addr = tz->iova; in eth_em_tx_queue_setup() 1311 txq->tx_ring = (struct e1000_data_desc *) tz->addr; in eth_em_tx_queue_setup()
|
| H A D | igb_rxtx.c | 1489 const struct rte_memzone *tz; in eth_igb_tx_queue_setup() local 1543 tz = rte_eth_dma_zone_reserve(dev, "tx_ring", queue_idx, size, in eth_igb_tx_queue_setup() 1545 if (tz == NULL) { in eth_igb_tx_queue_setup() 1550 txq->mz = tz; in eth_igb_tx_queue_setup() 1563 txq->tx_ring_phys_addr = tz->iova; in eth_igb_tx_queue_setup() 1565 txq->tx_ring = (union e1000_adv_tx_desc *) tz->addr; in eth_igb_tx_queue_setup()
|
| /dpdk/drivers/net/ice/ |
| H A D | ice_rxtx.c | 1216 const struct rte_memzone *tz; in ice_tx_queue_setup() local 1343 if (!tz) { in ice_tx_queue_setup() 1349 txq->mz = tz; in ice_tx_queue_setup() 1364 txq->tx_ring_dma = tz->iova; in ice_tx_queue_setup() 1365 txq->tx_ring = tz->addr; in ice_tx_queue_setup() 2224 const struct rte_memzone *tz = NULL; in ice_fdir_setup_tx_resources() local 2250 tz = rte_eth_dma_zone_reserve(dev, "fdir_tx_ring", in ice_fdir_setup_tx_resources() 2253 if (!tz) { in ice_fdir_setup_tx_resources() 2259 txq->mz = tz; in ice_fdir_setup_tx_resources() 2265 txq->tx_ring_dma = tz->iova; in ice_fdir_setup_tx_resources() [all …]
|
| /dpdk/drivers/net/igc/ |
| H A D | igc_txrx.c | 1953 const struct rte_memzone *tz; in eth_igc_tx_queue_setup() local 2001 tz = rte_eth_dma_zone_reserve(dev, "tx_ring", queue_idx, size, in eth_igc_tx_queue_setup() 2003 if (tz == NULL) { in eth_igc_tx_queue_setup() 2018 txq->tx_ring_phys_addr = tz->iova; in eth_igc_tx_queue_setup() 2020 txq->tx_ring = (union igc_adv_tx_desc *)tz->addr; in eth_igc_tx_queue_setup()
|
| /dpdk/drivers/net/cxgbe/ |
| H A D | sge.c | 1392 const struct rte_memzone *tz; in alloc_ring() local 1408 tz = rte_eth_dma_zone_reserve(dev, z_name, queue_id, len, 4096, in alloc_ring() 1410 if (!tz) in alloc_ring() 1413 memset(tz->addr, 0, len); in alloc_ring() 1427 *phys = (uint64_t)tz->iova; in alloc_ring() 1428 return tz->addr; in alloc_ring()
|
| /dpdk/drivers/net/ngbe/ |
| H A D | ngbe_rxtx.c | 1963 const struct rte_memzone *tz; in ngbe_dev_tx_queue_setup() local 2019 tz = rte_eth_dma_zone_reserve(dev, "tx_ring", queue_idx, in ngbe_dev_tx_queue_setup() 2022 if (tz == NULL) { in ngbe_dev_tx_queue_setup() 2043 txq->tx_ring_phys_addr = TMZ_PADDR(tz); in ngbe_dev_tx_queue_setup() 2044 txq->tx_ring = (struct ngbe_tx_desc *)TMZ_VADDR(tz); in ngbe_dev_tx_queue_setup()
|
| /dpdk/drivers/net/ixgbe/ |
| H A D | ixgbe_rxtx.c | 2624 const struct rte_memzone *tz; in ixgbe_dev_tx_queue_setup() local 2757 tz = rte_eth_dma_zone_reserve(dev, "tx_ring", queue_idx, in ixgbe_dev_tx_queue_setup() 2760 if (tz == NULL) { in ixgbe_dev_tx_queue_setup() 2765 txq->mz = tz; in ixgbe_dev_tx_queue_setup() 2796 txq->tx_ring_phys_addr = tz->iova; in ixgbe_dev_tx_queue_setup() 2797 txq->tx_ring = (union ixgbe_adv_tx_desc *) tz->addr; in ixgbe_dev_tx_queue_setup()
|
| /dpdk/drivers/net/txgbe/ |
| H A D | txgbe_rxtx.c | 2262 const struct rte_memzone *tz; in txgbe_dev_tx_queue_setup() local 2330 tz = rte_eth_dma_zone_reserve(dev, "tx_ring", queue_idx, in txgbe_dev_tx_queue_setup() 2333 if (tz == NULL) { in txgbe_dev_tx_queue_setup() 2368 txq->tx_ring_phys_addr = TMZ_PADDR(tz); in txgbe_dev_tx_queue_setup() 2369 txq->tx_ring = (struct txgbe_tx_desc *)TMZ_VADDR(tz); in txgbe_dev_tx_queue_setup()
|