| /f-stack/dpdk/app/test/ |
| H A D | test_memzone.c | 70 if (mz != NULL) { in test_memzone_invalid_alignment() 77 if (mz != NULL) { in test_memzone_invalid_alignment() 91 if (mz != NULL) { in test_memzone_reserving_zone_size_bigger_than_the_maximum() 99 if (mz != NULL) { in test_memzone_reserving_zone_size_bigger_than_the_maximum() 578 if (mz->addr != RTE_PTR_ALIGN(mz->addr, align)) { in test_memzone_reserve_max_aligned() 586 if (mz->len < minlen || mz->len > maxlen) { in test_memzone_reserve_max_aligned() 779 if ((mz->len & RTE_CACHE_LINE_MASK) != 0 || mz->len < len || in check_memzone_bounded() 787 ((mz->iova + mz->len - 1) & bmask)) { in check_memzone_bounded() 862 if (mz[0] > mz[1]) in test_memzone_free() 880 if (mz[2] > mz[1]) { in test_memzone_free() [all …]
|
| H A D | test_timer_secondary.c | 115 const struct rte_memzone *mz; in test_timer_secondary() local 126 mz = rte_memzone_reserve(TEST_INFO_MZ_NAME, sizeof(*test_info), in test_timer_secondary() 128 test_info = mz->addr; in test_timer_secondary() 173 mz = rte_memzone_lookup(TEST_INFO_MZ_NAME); in test_timer_secondary() 174 test_info = mz->addr; in test_timer_secondary()
|
| /f-stack/dpdk/lib/librte_mbuf/ |
| H A D | rte_mbuf_pool_ops.c | 15 const struct rte_memzone *mz; in rte_mbuf_set_platform_mempool_ops() local 24 if (mz == NULL) { in rte_mbuf_set_platform_mempool_ops() 27 if (mz == NULL) in rte_mbuf_set_platform_mempool_ops() 29 strcpy(mz->addr, ops_name); in rte_mbuf_set_platform_mempool_ops() 37 (char *)mz->addr); in rte_mbuf_set_platform_mempool_ops() 47 if (mz == NULL) in rte_mbuf_platform_mempool_ops() 49 return mz->addr; in rte_mbuf_platform_mempool_ops() 64 if (mz == NULL) { in rte_mbuf_set_user_mempool_ops() 67 if (mz == NULL) in rte_mbuf_set_user_mempool_ops() 82 if (mz == NULL) in rte_mbuf_user_mempool_ops() [all …]
|
| /f-stack/dpdk/lib/librte_eal/common/ |
| H A D | eal_common_memzone.c | 50 return mz; in memzone_lookup_thread_unsafe() 164 mz = NULL; in memzone_reserve_aligned_thread_unsafe() 177 strlcpy(mz->name, name, sizeof(mz->name)); in memzone_reserve_aligned_thread_unsafe() 185 mz->flags = 0; in memzone_reserve_aligned_thread_unsafe() 187 return mz; in memzone_reserve_aligned_thread_unsafe() 206 bound, mz); in rte_memzone_reserve_thread_safe() 210 return mz; in rte_memzone_reserve_thread_safe() 330 mz->name, in dump_memzone() 331 mz->len, in dump_memzone() 332 mz->addr, in dump_memzone() [all …]
|
| /f-stack/dpdk/drivers/net/qede/base/ |
| H A D | bcm_osal.c | 130 const struct rte_memzone *mz; in osal_dma_alloc_coherent() local 150 if (!mz) { in osal_dma_alloc_coherent() 157 *phys = mz->iova; in osal_dma_alloc_coherent() 158 ecore_mz_mapping[ecore_mz_count++] = mz; in osal_dma_alloc_coherent() 162 mz->len, (unsigned long)mz->iova, mz->addr, core_id); in osal_dma_alloc_coherent() 163 return mz->addr; in osal_dma_alloc_coherent() 169 const struct rte_memzone *mz; in osal_dma_alloc_coherent_aligned() local 189 if (!mz) { in osal_dma_alloc_coherent_aligned() 196 *phys = mz->iova; in osal_dma_alloc_coherent_aligned() 201 mz->len, (unsigned long)mz->iova, mz->addr, core_id); in osal_dma_alloc_coherent_aligned() [all …]
|
| /f-stack/dpdk/examples/l2fwd-event/ |
| H A D | l2fwd_common.h | 106 const struct rte_memzone *mz; in l2fwd_get_rsrc() local 108 mz = rte_memzone_lookup(name); in l2fwd_get_rsrc() 109 if (mz != NULL) in l2fwd_get_rsrc() 110 return mz->addr; in l2fwd_get_rsrc() 112 mz = rte_memzone_reserve(name, sizeof(struct l2fwd_resources), 0, 0); in l2fwd_get_rsrc() 113 if (mz != NULL) { in l2fwd_get_rsrc() 114 struct l2fwd_resources *rsrc = mz->addr; in l2fwd_get_rsrc() 123 return mz->addr; in l2fwd_get_rsrc()
|
| /f-stack/dpdk/drivers/crypto/nitrox/ |
| H A D | nitrox_qp.c | 20 const struct rte_memzone *mz; in nitrox_setup_cmdq() local 25 mz = rte_memzone_reserve_aligned(mz_name, cmdq_size, socket_id, in nitrox_setup_cmdq() 29 if (!mz) { in nitrox_setup_cmdq() 35 qp->cmdq.mz = mz; in nitrox_setup_cmdq() 38 qp->cmdq.ring = mz->addr; in nitrox_setup_cmdq() 40 setup_nps_pkt_input_ring(bar_addr, qp->qno, qp->count, mz->iova); in nitrox_setup_cmdq() 67 return rte_memzone_free(qp->cmdq.mz); in nitrox_release_cmdq()
|
| /f-stack/dpdk/drivers/common/iavf/ |
| H A D | iavf_impl.c | 22 const struct rte_memzone *mz = NULL; in iavf_allocate_dma_mem_d() local 29 mz = rte_memzone_reserve_bounded(z_name, size, SOCKET_ID_ANY, in iavf_allocate_dma_mem_d() 32 if (!mz) in iavf_allocate_dma_mem_d() 36 mem->va = mz->addr; in iavf_allocate_dma_mem_d() 37 mem->pa = mz->iova; in iavf_allocate_dma_mem_d() 38 mem->zone = (const void *)mz; in iavf_allocate_dma_mem_d()
|
| /f-stack/dpdk/drivers/common/qat/ |
| H A D | qat_device.c | 74 if (qat_pci_devs[i].mz && in qat_pci_get_named_dev() 76 qat_pci_devs[i].mz->addr)->name, name) in qat_pci_get_named_dev() 79 qat_pci_devs[i].mz->addr; in qat_pci_get_named_dev() 92 if (qat_pci_devs[dev_id].mz == NULL) in qat_pci_find_free_device_index() 180 if (mz == NULL) { in qat_pci_device_allocate() 186 qat_dev = mz->addr; in qat_pci_device_allocate() 187 qat_pci_devs[qat_dev->qat_dev_id].mz = mz; in qat_pci_device_allocate() 207 qat_pci_devs[qat_dev_id].mz = rte_memzone_reserve(name, in qat_pci_device_allocate() 211 if (qat_pci_devs[qat_dev_id].mz == NULL) { in qat_pci_device_allocate() 217 qat_dev = qat_pci_devs[qat_dev_id].mz->addr; in qat_pci_device_allocate() [all …]
|
| /f-stack/dpdk/drivers/net/bnxt/ |
| H A D | bnxt_ring.c | 111 const struct rte_memzone *mz = NULL; in bnxt_alloc_rings() local 203 mz = rte_memzone_lookup(mz_name); in bnxt_alloc_rings() 204 if (!mz) { in bnxt_alloc_rings() 211 if (mz == NULL) in bnxt_alloc_rings() 214 memset(mz->addr, 0, mz->len); in bnxt_alloc_rings() 215 mz_phys_addr = mz->iova; in bnxt_alloc_rings() 218 txq->mz = mz; in bnxt_alloc_rings() 225 tx_ring->mem_zone = (const void *)mz; in bnxt_alloc_rings() 238 rxq->mz = mz; in bnxt_alloc_rings() 291 cp_ring->mem_zone = (const void *)mz; in bnxt_alloc_rings() [all …]
|
| /f-stack/dpdk/drivers/net/ena/base/ |
| H A D | ena_plat_dpdk.h | 178 const struct rte_memzone *mz = NULL; \ 189 handle = mz; \ 191 if (mz == NULL) { \ 195 memset(mz->addr, 0, size); \ 196 virt = mz->addr; \ 197 phys = mz->iova; \ 225 mem_handle = mz; \ 227 if (mz == NULL) { \ 231 memset(mz->addr, 0, size); \ 232 virt = mz->addr; \ [all …]
|
| /f-stack/dpdk/lib/librte_latencystats/ |
| H A D | rte_latencystats.c | 223 const struct rte_memzone *mz = NULL; in rte_latencystats_init() local 233 if (mz == NULL) { in rte_latencystats_init() 239 glob_stats = mz->addr; in rte_latencystats_init() 306 const struct rte_memzone *mz = NULL; in rte_latencystats_uninit() local 340 mz = rte_memzone_lookup(MZ_RTE_LATENCY_STATS); in rte_latencystats_uninit() 341 if (mz) in rte_latencystats_uninit() 342 rte_memzone_free(mz); in rte_latencystats_uninit() 369 const struct rte_memzone *mz; in rte_latencystats_get() local 370 mz = rte_memzone_lookup(MZ_RTE_LATENCY_STATS); in rte_latencystats_get() 371 if (mz == NULL) { in rte_latencystats_get() [all …]
|
| /f-stack/dpdk/drivers/crypto/bcmfs/ |
| H A D | bcmfs_qp.c | 40 const struct rte_memzone *mz; in bcmfs_queue_delete() local 50 mz = rte_memzone_lookup(queue->memz_name); in bcmfs_queue_delete() 51 if (mz != NULL) { in bcmfs_queue_delete() 54 status = rte_memzone_free(mz); in bcmfs_queue_delete() 68 const struct rte_memzone *mz; in queue_dma_zone_reserve() local 70 mz = rte_memzone_lookup(queue_name); in queue_dma_zone_reserve() 71 if (mz != NULL) { in queue_dma_zone_reserve() 72 if (((size_t)queue_size <= mz->len) && in queue_dma_zone_reserve() 74 socket_id == mz->socket_id)) { in queue_dma_zone_reserve() 77 return mz; in queue_dma_zone_reserve() [all …]
|
| /f-stack/dpdk/drivers/common/octeontx2/ |
| H A D | otx2_common.c | 32 const struct rte_memzone *mz; in otx2_intra_dev_get_cfg() local 35 mz = rte_memzone_lookup(name); in otx2_intra_dev_get_cfg() 36 if (mz != NULL) in otx2_intra_dev_get_cfg() 37 return mz->addr; in otx2_intra_dev_get_cfg() 40 mz = rte_memzone_reserve_aligned(name, sizeof(struct otx2_idev_cfg), in otx2_intra_dev_get_cfg() 42 if (mz != NULL) { in otx2_intra_dev_get_cfg() 43 idev = mz->addr; in otx2_intra_dev_get_cfg()
|
| /f-stack/dpdk/lib/librte_stack/ |
| H A D | rte_stack.c | 54 const struct rte_memzone *mz; in rte_stack_create() local 92 mz = rte_memzone_reserve_aligned(mz_name, sz, socket_id, in rte_stack_create() 94 if (mz == NULL) { in rte_stack_create() 101 s = mz->addr; in rte_stack_create() 112 rte_memzone_free(mz); in rte_stack_create() 116 s->memzone = mz; in rte_stack_create()
|
| /f-stack/dpdk/drivers/event/dlb2/pf/ |
| H A D | dlb2_pf.c | 244 *mz = rte_memzone_reserve_aligned(mz_name, size, socket_id, in dlb2_alloc_coherent_aligned() 246 if (*mz == NULL) { in dlb2_alloc_coherent_aligned() 252 *phys = (*mz)->iova; in dlb2_alloc_coherent_aligned() 253 return (*mz)->addr; in dlb2_alloc_coherent_aligned() 266 const struct rte_memzone *mz; in dlb2_pf_ldb_port_create() local 286 port_base = dlb2_alloc_coherent_aligned(&mz, &cq_base, alloc_sz, in dlb2_pf_ldb_port_create() 315 dlb2_port[response.id][DLB2_LDB_PORT].mz = mz; in dlb2_pf_ldb_port_create() 325 rte_memzone_free(mz); in dlb2_pf_ldb_port_create() 342 const struct rte_memzone *mz; in dlb2_pf_dir_port_create() local 391 dlb2_port[response.id][DLB2_DIR_PORT].mz = mz; in dlb2_pf_dir_port_create() [all …]
|
| /f-stack/dpdk/drivers/net/octeontx2/ |
| H A D | otx2_ethdev_sec.c | 146 const struct rte_memzone *mz; in lookup_mem_sa_tbl_clear() local 151 if (mz == NULL) in lookup_mem_sa_tbl_clear() 154 mem = mz->addr; in lookup_mem_sa_tbl_clear() 170 const struct rte_memzone *mz; in lookup_mem_sa_index_update() local 175 if (mz == NULL) { in lookup_mem_sa_index_update() 180 mem = mz->addr; in lookup_mem_sa_index_update() 205 const struct rte_memzone *mz; in in_sa_get() local 209 if (mz == NULL) { in in_sa_get() 214 sa = mz->addr; in in_sa_get() 764 if (mz == NULL) in eth_sec_ipsec_cfg() [all …]
|
| H A D | otx2_lookup.c | 328 const struct rte_memzone *mz; in otx2_nix_fastpath_lookup_mem_get() local 335 mz = rte_memzone_lookup(name); in otx2_nix_fastpath_lookup_mem_get() 336 if (mz != NULL) in otx2_nix_fastpath_lookup_mem_get() 337 return mz->addr; in otx2_nix_fastpath_lookup_mem_get() 340 mz = rte_memzone_reserve_aligned(name, LOOKUP_ARRAY_SZ, in otx2_nix_fastpath_lookup_mem_get() 342 if (mz != NULL) { in otx2_nix_fastpath_lookup_mem_get() 343 mem = mz->addr; in otx2_nix_fastpath_lookup_mem_get()
|
| /f-stack/dpdk/drivers/mempool/octeontx2/ |
| H A D | otx2_mempool.c | 345 const struct rte_memzone *mz; in otx2_npa_init() local 349 mz = rte_memzone_reserve_aligned(otx2_npa_dev_to_name(pci_dev, name), in otx2_npa_init() 352 if (mz == NULL) in otx2_npa_init() 355 dev = mz->addr; in otx2_npa_init() 374 rte_memzone_free(mz); in otx2_npa_init() 384 const struct rte_memzone *mz; in otx2_npa_fini() local 387 mz = rte_memzone_lookup(otx2_npa_dev_to_name(pci_dev, name)); in otx2_npa_fini() 388 if (mz == NULL) in otx2_npa_fini() 391 dev = mz->addr; in otx2_npa_fini() 406 rte_memzone_free(mz); in otx2_npa_fini()
|
| /f-stack/dpdk/drivers/event/dlb/pf/ |
| H A D | dlb_pf.c | 260 if (*mz == NULL) { in dlb_alloc_coherent_aligned() 266 *phys = (*mz)->iova; in dlb_alloc_coherent_aligned() 267 return (*mz)->addr; in dlb_alloc_coherent_aligned() 279 const struct rte_memzone *mz; in dlb_pf_ldb_port_create() local 305 port_base = dlb_alloc_coherent_aligned(&mz, &pc_dma_base, in dlb_pf_ldb_port_create() 340 dlb_port[response.id][DLB_LDB].mz = mz; in dlb_pf_ldb_port_create() 349 rte_memzone_free(mz); in dlb_pf_ldb_port_create() 363 const struct rte_memzone *mz; in dlb_pf_dir_port_create() local 383 port_base = dlb_alloc_coherent_aligned(&mz, &pc_dma_base, in dlb_pf_dir_port_create() 418 dlb_port[response.id][DLB_DIR].mz = mz; in dlb_pf_dir_port_create() [all …]
|
| /f-stack/dpdk/lib/librte_graph/ |
| H A D | graph_populate.c | 198 const struct rte_memzone *mz; in graph_fp_mem_create() local 202 mz = rte_memzone_reserve(graph->name, sz, graph->socket, 0); in graph_fp_mem_create() 203 if (mz == NULL) in graph_fp_mem_create() 207 graph->graph = mz->addr; in graph_fp_mem_create() 208 graph->mz = mz; in graph_fp_mem_create() 233 return rte_memzone_free(graph->mz); in graph_fp_mem_destroy()
|
| /f-stack/dpdk/drivers/net/virtio/ |
| H A D | virtio_rxtx.h | 33 const struct rte_memzone *mz; /**< mem zone to populate RX ring. */ member 48 const struct rte_memzone *mz; /**< mem zone to populate TX ring. */ member 57 const struct rte_memzone *mz; /**< mem zone to populate CTL ring. */ member
|
| /f-stack/dpdk/drivers/raw/ioat/ |
| H A D | ioat_rawdev.c | 148 const struct rte_memzone *mz = NULL; in ioat_rawdev_create() local 169 mz = rte_memzone_reserve(mz_name, sizeof(struct rte_ioat_rawdev), in ioat_rawdev_create() 171 if (mz == NULL) { in ioat_rawdev_create() 177 rawdev->dev_private = mz->addr; in ioat_rawdev_create() 185 ioat->mz = mz; in ioat_rawdev_create() 190 ioat->status_addr = ioat->mz->iova + in ioat_rawdev_create() 252 rte_memzone_free(ioat->mz); in ioat_rawdev_destroy()
|
| /f-stack/dpdk/drivers/net/vmxnet3/ |
| H A D | vmxnet3_ethdev.c | 170 if (mz) in gpa_zone_reserve() 176 if (mz) in gpa_zone_reserve() 177 return mz; in gpa_zone_reserve() 448 if (mz == NULL) { in vmxnet3_dev_configure() 452 memset(mz->addr, 0, mz->len); in vmxnet3_dev_configure() 467 if (mz == NULL) { in vmxnet3_dev_configure() 471 memset(mz->addr, 0, mz->len); in vmxnet3_dev_configure() 484 if (mz == NULL) { in vmxnet3_dev_configure() 489 memset(mz->addr, 0, mz->len); in vmxnet3_dev_configure() 534 if (mz == NULL) { in vmxnet3_dev_setup_memreg() [all …]
|
| /f-stack/dpdk/lib/librte_mempool/ |
| H A D | rte_mempool.c | 264 rte_memzone_free(mz); in rte_mempool_memchunk_mz_free() 575 if (mz == NULL) { in rte_mempool_populate_default() 581 iova = mz->iova; in rte_mempool_populate_default() 587 iova, mz->len, in rte_mempool_populate_default() 589 (void *)(uintptr_t)mz); in rte_mempool_populate_default() 592 mz->len, pg_sz, in rte_mempool_populate_default() 598 rte_memzone_free(mz); in rte_mempool_populate_default() 728 rte_memzone_free(mp->mz); in rte_mempool_free() 862 if (mz == NULL) in rte_mempool_create_empty() 866 mp = mz->addr; in rte_mempool_create_empty() [all …]
|