| /f-stack/dpdk/lib/librte_vhost/ |
| H A D | iotlb.c | 85 node->iova = iova; in vhost_user_iotlb_pending_insert() 104 if (node->iova < iova) in vhost_user_iotlb_pending_remove() 106 if (node->iova >= iova + size) in vhost_user_iotlb_pending_remove() 178 new_node->iova = iova; in vhost_user_iotlb_cache_insert() 190 if (node->iova == new_node->iova) { in vhost_user_iotlb_cache_insert() 223 if (unlikely(iova + size < node->iova)) in vhost_user_iotlb_cache_remove() 226 if (iova < node->iova + node->size) { in vhost_user_iotlb_cache_remove() 248 if (unlikely(iova < node->iova)) in vhost_user_iotlb_cache_find() 251 if (iova >= node->iova + node->size) in vhost_user_iotlb_cache_find() 259 offset = iova - node->iova; in vhost_user_iotlb_cache_find() [all …]
|
| H A D | iotlb.h | 36 void vhost_user_iotlb_cache_insert(struct vhost_virtqueue *vq, uint64_t iova, 40 uint64_t iova, uint64_t size); 41 uint64_t vhost_user_iotlb_cache_find(struct vhost_virtqueue *vq, uint64_t iova, 43 bool vhost_user_iotlb_pending_miss(struct vhost_virtqueue *vq, uint64_t iova, 45 void vhost_user_iotlb_pending_insert(struct vhost_virtqueue *vq, uint64_t iova, 47 void vhost_user_iotlb_pending_remove(struct vhost_virtqueue *vq, uint64_t iova,
|
| H A D | vhost.h | 235 __u64 iova; member 433 uint64_t iova, uint64_t len); 438 uint64_t iova, uint64_t len); 487 uint64_t iova, uint64_t len) in vhost_log_cache_write_iova() argument 493 __vhost_log_cache_write_iova(dev, vq, iova, len); in vhost_log_cache_write_iova() 495 __vhost_log_cache_write(dev, vq, iova, len); in vhost_log_cache_write_iova() 500 uint64_t iova, uint64_t len) in vhost_log_write_iova() argument 506 __vhost_log_write_iova(dev, vq, iova, len); in vhost_log_write_iova() 508 __vhost_log_write(dev, iova, len); in vhost_log_write_iova() 691 uint64_t iova, uint64_t *len, uint8_t perm); [all …]
|
| /f-stack/dpdk/app/test/ |
| H A D | test_external_mem.c | 54 expected_iova = (iova == NULL) ? RTE_BAD_IOVA : iova[i]; in check_mem() 55 if (ms->iova != expected_iova) { in check_mem() 232 if (iova != NULL) { in test_malloc_invalid_param() 241 iova, n_pages - 1, pgsz) >= 0 || in test_malloc_invalid_param() 248 iova, n_pages + 1, pgsz) >= 0 || in test_malloc_invalid_param() 302 iova, n_pages, pgsz) != 0) { in test_malloc_basic() 475 if (iova != NULL) { in test_extmem_invalid_param() 484 iova, n_pages - 1, pgsz) >= 0 || in test_extmem_invalid_param() 537 rte_iova_t iova[len / pgsz]; in test_external_mem() local 543 n_pages = RTE_DIM(iova); in test_external_mem() [all …]
|
| H A D | test_memzone.c | 654 if ((memzone_aligned_128->iova & 127) != 0) in test_memzone_aligned() 665 if ((memzone_aligned_256->iova & 255) != 0) in test_memzone_aligned() 676 if ((memzone_aligned_512->iova & 511) != 0) in test_memzone_aligned() 687 if ((memzone_aligned_1024->iova & 1023) != 0) in test_memzone_aligned() 767 if ((mz->iova & ((rte_iova_t)align - 1)) != 0) { in check_memzone_bounded() 786 if ((mz->iova & bmask) != in check_memzone_bounded() 787 ((mz->iova + mz->len - 1) & bmask)) { in check_memzone_bounded() 992 if ((memzone1->iova & RTE_CACHE_LINE_MASK) != 0) in test_memzone_basic() 1012 memzone2->iova, memzone2->len)) in test_memzone_basic() 1016 memzone3->iova, memzone3->len)) in test_memzone_basic() [all …]
|
| /f-stack/dpdk/lib/librte_eal/linux/ |
| H A D | eal_vfio.c | 114 if (umm_a->iova < umm_b->iova) in user_mem_map_cmp() 116 if (umm_a->iova > umm_b->iova) in user_mem_map_cmp() 157 tmp.iova = src->iova + src->len; in adjust_map() 162 end->iova = tmp.iova + tmp.len; in adjust_map() 173 if (left->iova + left->len != right->iova) in merge_map() 204 if (iova < map->iova || iova >= map_iova_end) in find_user_mem_map() 1360 dma_map.iova = iova; in vfio_type1_dma_mem_map() 1386 dma_unmap.iova = iova; in vfio_type1_dma_mem_map() 1454 dma_map.iova = iova; in vfio_spapr_dma_do_map() 1471 dma_unmap.iova = iova; in vfio_spapr_dma_do_map() [all …]
|
| /f-stack/dpdk/drivers/crypto/nitrox/ |
| H A D | nitrox_sym_reqmgr.c | 127 rte_iova_t iova; member 133 uint64_t iova[4]; member 147 rte_iova_t iova; member 163 rte_iova_t iova; member 170 sr->iova = iova; in softreq_init() 261 sglist[cnt].iova = iova; in fill_sglist() 323 sgcomp->iova[j] = rte_cpu_to_be_64(sglist->iova); in create_sgcomp() 357 sr->out.sglist[cnt].iova = sr->iova + offsetof(struct nitrox_softreq, in create_cipher_outbuf() 373 sr->out.sglist[cnt].iova = sr->iova + offsetof(struct nitrox_softreq, in create_cipher_outbuf() 560 sr->out.sglist[cnt].iova = sr->in.sglist[i].iova; in create_aead_inplace_outbuf() [all …]
|
| /f-stack/dpdk/kernel/linux/kni/ |
| H A D | kni_dev.h | 95 unsigned long iova) in iova_to_phys() argument 101 offset = iova & (PAGE_SIZE - 1); in iova_to_phys() 105 ret = get_user_pages_remote(tsk, tsk->mm, iova, 1, in iova_to_phys() 108 ret = get_user_pages_remote(tsk->mm, iova, 1, in iova_to_phys() 120 static inline void *iova_to_kva(struct task_struct *tsk, unsigned long iova) in iova_to_kva() argument 122 return phys_to_virt(iova_to_phys(tsk, iova)); in iova_to_kva()
|
| /f-stack/dpdk/drivers/net/virtio/virtio_user/ |
| H A D | vhost_vdpa.c | 58 uint64_t iova; member 85 uint64_t iova, size_t len) in vhost_vdpa_dma_map() argument 91 msg.iotlb.iova = iova; in vhost_vdpa_dma_map() 107 uint64_t iova, size_t len) in vhost_vdpa_dma_unmap() argument 113 msg.iotlb.iova = iova; in vhost_vdpa_dma_unmap() 135 return vhost_vdpa_dma_map(dev, ms->addr, ms->iova, len); in vhost_vdpa_map_contig() 149 if (ms->iova == RTE_BAD_IOVA) in vhost_vdpa_map() 156 return vhost_vdpa_dma_map(dev, ms->addr, ms->iova, ms->len); in vhost_vdpa_map()
|
| H A D | vhost.h | 109 uint64_t iova, size_t len); 111 uint64_t iova, size_t len);
|
| /f-stack/dpdk/lib/librte_eal/common/ |
| H A D | eal_common_memory.c | 335 rte_iova_t iova; member 343 if (vi->iova >= ms->iova && vi->iova < (ms->iova + ms->len)) { in find_virt() 344 size_t offset = vi->iova - ms->iova; in find_virt() 356 if (vi->iova >= ms->iova && vi->iova < (ms->iova + len)) { in find_virt_legacy() 357 size_t offset = vi->iova - ms->iova; in find_virt_legacy() 374 vi.iova = iova; in rte_mem_iova2virt() 439 ms->iova, in dump_memseg() 530 rte_iova_t iova; in check_iova() local 533 iova = (ms->iova + ms->len) - 1; in check_iova() 534 if (!(iova & *mask)) in check_iova() [all …]
|
| /f-stack/dpdk/drivers/compress/octeontx/ |
| H A D | otx_zip.c | 42 uint64_t iova; in zipvf_q_init() local 56 iova = rte_mem_virt2iova(cmdq_addr); in zipvf_q_init() 58 cmdq->iova = iova; in zipvf_q_init() 61 que_sbuf_addr.s.ptr = (cmdq->iova >> 7); in zipvf_q_init() 119 ncp.s.addr = cmdq->iova; in zipvf_push_command()
|
| /f-stack/dpdk/lib/librte_mempool/ |
| H A D | rte_mempool_ops_default.c | 92 unsigned int max_objs, void *vaddr, rte_iova_t iova, in rte_mempool_op_populate_helper() argument 129 (iova == RTE_BAD_IOVA) ? RTE_BAD_IOVA : (iova + off)); in rte_mempool_op_populate_helper() 139 void *vaddr, rte_iova_t iova, size_t len, in rte_mempool_op_populate_default() argument 143 return rte_mempool_op_populate_helper(mp, 0, max_objs, vaddr, iova, in rte_mempool_op_populate_default()
|
| H A D | rte_mempool.c | 156 void *obj, rte_iova_t iova) in mempool_add_elem() argument 164 hdr->iova = iova; in mempool_add_elem() 334 memhdr->iova = iova; in rte_mempool_populate_iova() 351 (iova == RTE_BAD_IOVA) ? RTE_BAD_IOVA : (iova + off), in rte_mempool_populate_iova() 392 rte_iova_t iova; in rte_mempool_populate_virt() local 403 iova = get_iova(addr + off); in rte_mempool_populate_virt() 417 iova_tmp != iova + phys_len) in rte_mempool_populate_virt() 478 rte_iova_t iova; in rte_mempool_populate_default() local 581 iova = mz->iova; in rte_mempool_populate_default() 583 iova = RTE_BAD_IOVA; in rte_mempool_populate_default() [all …]
|
| H A D | rte_mempool_trace.h | 70 rte_iova_t iova, size_t len, void *free_cb, void *opaque), 74 rte_trace_point_emit_u64(iova); 135 void *vaddr, uint64_t iova, size_t len, void *obj_cb, 141 rte_trace_point_emit_u64(iova);
|
| H A D | rte_mempool_ops.c | 129 void *vaddr, rte_iova_t iova, size_t len, in rte_mempool_ops_populate() argument 137 rte_mempool_trace_ops_populate(mp, max_objs, vaddr, iova, len, obj_cb, in rte_mempool_ops_populate() 141 iova, len, obj_cb, in rte_mempool_ops_populate() 144 return ops->populate(mp, max_objs, vaddr, iova, len, obj_cb, in rte_mempool_ops_populate()
|
| H A D | rte_mempool.h | 141 rte_iova_t iova; /**< IO address of the object. */ member 187 rte_iova_t iova; /**< IO address of the chunk */ member 518 void *opaque, void *vaddr, rte_iova_t iova); 550 void *vaddr, rte_iova_t iova, size_t len, 592 void *vaddr, rte_iova_t iova, size_t len, 603 void *vaddr, rte_iova_t iova, size_t len, 824 void *vaddr, rte_iova_t iova, size_t len, 1090 rte_iova_t iova, size_t len, rte_mempool_memchunk_free_cb_t *free_cb, 1678 return hdr->iova; in rte_mempool_virt2iova()
|
| /f-stack/dpdk/drivers/bus/dpaa/ |
| H A D | rte_dpaa_bus.h | 159 rte_iova_t iova; member 182 if (paddr >= ms->iova && paddr < in rte_dpaa_mem_ptov() 183 ms->iova + ms->len) in rte_dpaa_mem_ptov() 184 return RTE_PTR_ADD(ms->vaddr, (uintptr_t)(paddr - ms->iova)); in rte_dpaa_mem_ptov() 202 return ms->iova + RTE_PTR_DIFF(vaddr, ms->addr); in rte_dpaa_mem_vtop()
|
| /f-stack/dpdk/drivers/mempool/octeontx2/ |
| H A D | otx2_mempool_ops.c | 625 pool->stack_base = mz->iova; in npa_lf_aura_pool_pair_alloc() 819 otx2_npa_l1d_way_set_get(uint64_t iova) in otx2_npa_l1d_way_set_get() argument 821 return (iova >> rte_log2_u32(RTE_CACHE_LINE_SIZE)) & 0x7; in otx2_npa_l1d_way_set_get() 826 rte_iova_t iova, size_t len, in otx2_npa_populate() argument 837 if (iova == RTE_BAD_IOVA) in otx2_npa_populate() 850 iova += off; in otx2_npa_populate() 854 start_iova = iova; in otx2_npa_populate() 855 while (start_iova < iova + len) { in otx2_npa_populate() 862 iova); in otx2_npa_populate() 873 npa_lf_aura_op_range_set(mp->pool_id, iova, iova + len); in otx2_npa_populate() [all …]
|
| /f-stack/dpdk/drivers/mempool/octeontx/ |
| H A D | rte_mempool_octeontx.c | 157 void *vaddr, rte_iova_t iova, size_t len, in octeontx_fpavf_populate() argument 166 if (iova == RTE_BAD_IOVA) in octeontx_fpavf_populate() 178 iova += off; in octeontx_fpavf_populate() 190 max_objs, vaddr, iova, len, in octeontx_fpavf_populate()
|
| /f-stack/dpdk/drivers/bus/fslmc/ |
| H A D | fslmc_vfio.c | 182 .iova = 0x6030000, in vfio_map_irq_region() 223 iova_addr = ms->iova; in fslmc_memevent_cb() 283 dma_map.iova = iovaddr; in fslmc_map_dma() 285 dma_map.iova = dma_map.vaddr; in fslmc_map_dma() 324 dma_unmap.iova = vaddr; in fslmc_unmap_dma() 335 (uint64_t)dma_unmap.iova, (uint64_t)dma_unmap.size); in fslmc_unmap_dma() 354 if (ms->iova == RTE_BAD_IOVA) in fslmc_dmamap_seg() 357 ret = fslmc_map_dma(ms->addr_64, ms->iova, ms->len); in fslmc_dmamap_seg() 368 rte_fslmc_vfio_mem_dmamap(uint64_t vaddr, uint64_t iova, uint64_t size) in rte_fslmc_vfio_mem_dmamap() argument 391 dma_map.iova = iova; in rte_fslmc_vfio_mem_dmamap() [all …]
|
| /f-stack/dpdk/lib/librte_cryptodev/ |
| H A D | rte_crypto_sym.h | 37 rte_iova_t iova; member 60 rte_iova_t iova; member 944 vec[0].iova = rte_pktmbuf_iova_offset(mb, ofs); in rte_crypto_mbuf_to_vec() 959 vec[i].iova = rte_pktmbuf_iova(nseg); in rte_crypto_mbuf_to_vec()
|
| /f-stack/dpdk/drivers/net/qede/base/ |
| H A D | bcm_osal.c | 157 *phys = mz->iova; in osal_dma_alloc_coherent() 162 mz->len, (unsigned long)mz->iova, mz->addr, core_id); in osal_dma_alloc_coherent() 196 *phys = mz->iova; in osal_dma_alloc_coherent_aligned() 201 mz->len, (unsigned long)mz->iova, mz->addr, core_id); in osal_dma_alloc_coherent_aligned() 210 if (phys == ecore_mz_mapping[j]->iova) { in osal_dma_free_mem()
|
| /f-stack/dpdk/lib/librte_eal/include/ |
| H A D | rte_vfio.h | 330 uint64_t iova, uint64_t len); 354 uint64_t iova, uint64_t len);
|
| /f-stack/dpdk/drivers/crypto/qat/ |
| H A D | qat_sym_hw_dp.c | 36 data[0].iova; in qat_sym_dp_parse_data_vec() 56 list->buffers[i].addr = data[i].iova; in qat_sym_dp_parse_data_vec() 132 aad_iova = aad->iova; in enqueue_one_aead_job() 136 aad_iova = aad->iova; in enqueue_one_aead_job() 149 aad_iova = iv->iova; in enqueue_one_aead_job() 210 auth_param->auth_res_addr = digest->iova; in enqueue_one_aead_job() 418 auth_param->auth_res_addr = digest->iova; in enqueue_one_auth_job() 424 auth_param->u1.aad_adr = auth_iv->iova; in enqueue_one_auth_job() 560 auth_param->u1.aad_adr = auth_iv->iova; in enqueue_one_chain_job() 571 auth_iova_end = cvec->iova + in enqueue_one_chain_job() [all …]
|