Home
last modified time | relevance | path

Searched refs:iova (Results 1 – 25 of 149) sorted by relevance

123456

/f-stack/dpdk/lib/librte_vhost/
H A Diotlb.c85 node->iova = iova; in vhost_user_iotlb_pending_insert()
104 if (node->iova < iova) in vhost_user_iotlb_pending_remove()
106 if (node->iova >= iova + size) in vhost_user_iotlb_pending_remove()
178 new_node->iova = iova; in vhost_user_iotlb_cache_insert()
190 if (node->iova == new_node->iova) { in vhost_user_iotlb_cache_insert()
223 if (unlikely(iova + size < node->iova)) in vhost_user_iotlb_cache_remove()
226 if (iova < node->iova + node->size) { in vhost_user_iotlb_cache_remove()
248 if (unlikely(iova < node->iova)) in vhost_user_iotlb_cache_find()
251 if (iova >= node->iova + node->size) in vhost_user_iotlb_cache_find()
259 offset = iova - node->iova; in vhost_user_iotlb_cache_find()
[all …]
H A Diotlb.h36 void vhost_user_iotlb_cache_insert(struct vhost_virtqueue *vq, uint64_t iova,
40 uint64_t iova, uint64_t size);
41 uint64_t vhost_user_iotlb_cache_find(struct vhost_virtqueue *vq, uint64_t iova,
43 bool vhost_user_iotlb_pending_miss(struct vhost_virtqueue *vq, uint64_t iova,
45 void vhost_user_iotlb_pending_insert(struct vhost_virtqueue *vq, uint64_t iova,
47 void vhost_user_iotlb_pending_remove(struct vhost_virtqueue *vq, uint64_t iova,
H A Dvhost.h235 __u64 iova; member
433 uint64_t iova, uint64_t len);
438 uint64_t iova, uint64_t len);
487 uint64_t iova, uint64_t len) in vhost_log_cache_write_iova() argument
493 __vhost_log_cache_write_iova(dev, vq, iova, len); in vhost_log_cache_write_iova()
495 __vhost_log_cache_write(dev, vq, iova, len); in vhost_log_cache_write_iova()
500 uint64_t iova, uint64_t len) in vhost_log_write_iova() argument
506 __vhost_log_write_iova(dev, vq, iova, len); in vhost_log_write_iova()
508 __vhost_log_write(dev, iova, len); in vhost_log_write_iova()
691 uint64_t iova, uint64_t *len, uint8_t perm);
[all …]
/f-stack/dpdk/app/test/
H A Dtest_external_mem.c54 expected_iova = (iova == NULL) ? RTE_BAD_IOVA : iova[i]; in check_mem()
55 if (ms->iova != expected_iova) { in check_mem()
232 if (iova != NULL) { in test_malloc_invalid_param()
241 iova, n_pages - 1, pgsz) >= 0 || in test_malloc_invalid_param()
248 iova, n_pages + 1, pgsz) >= 0 || in test_malloc_invalid_param()
302 iova, n_pages, pgsz) != 0) { in test_malloc_basic()
475 if (iova != NULL) { in test_extmem_invalid_param()
484 iova, n_pages - 1, pgsz) >= 0 || in test_extmem_invalid_param()
537 rte_iova_t iova[len / pgsz]; in test_external_mem() local
543 n_pages = RTE_DIM(iova); in test_external_mem()
[all …]
H A Dtest_memzone.c654 if ((memzone_aligned_128->iova & 127) != 0) in test_memzone_aligned()
665 if ((memzone_aligned_256->iova & 255) != 0) in test_memzone_aligned()
676 if ((memzone_aligned_512->iova & 511) != 0) in test_memzone_aligned()
687 if ((memzone_aligned_1024->iova & 1023) != 0) in test_memzone_aligned()
767 if ((mz->iova & ((rte_iova_t)align - 1)) != 0) { in check_memzone_bounded()
786 if ((mz->iova & bmask) != in check_memzone_bounded()
787 ((mz->iova + mz->len - 1) & bmask)) { in check_memzone_bounded()
992 if ((memzone1->iova & RTE_CACHE_LINE_MASK) != 0) in test_memzone_basic()
1012 memzone2->iova, memzone2->len)) in test_memzone_basic()
1016 memzone3->iova, memzone3->len)) in test_memzone_basic()
[all …]
/f-stack/dpdk/lib/librte_eal/linux/
H A Deal_vfio.c114 if (umm_a->iova < umm_b->iova) in user_mem_map_cmp()
116 if (umm_a->iova > umm_b->iova) in user_mem_map_cmp()
157 tmp.iova = src->iova + src->len; in adjust_map()
162 end->iova = tmp.iova + tmp.len; in adjust_map()
173 if (left->iova + left->len != right->iova) in merge_map()
204 if (iova < map->iova || iova >= map_iova_end) in find_user_mem_map()
1360 dma_map.iova = iova; in vfio_type1_dma_mem_map()
1386 dma_unmap.iova = iova; in vfio_type1_dma_mem_map()
1454 dma_map.iova = iova; in vfio_spapr_dma_do_map()
1471 dma_unmap.iova = iova; in vfio_spapr_dma_do_map()
[all …]
/f-stack/dpdk/drivers/crypto/nitrox/
H A Dnitrox_sym_reqmgr.c127 rte_iova_t iova; member
133 uint64_t iova[4]; member
147 rte_iova_t iova; member
163 rte_iova_t iova; member
170 sr->iova = iova; in softreq_init()
261 sglist[cnt].iova = iova; in fill_sglist()
323 sgcomp->iova[j] = rte_cpu_to_be_64(sglist->iova); in create_sgcomp()
357 sr->out.sglist[cnt].iova = sr->iova + offsetof(struct nitrox_softreq, in create_cipher_outbuf()
373 sr->out.sglist[cnt].iova = sr->iova + offsetof(struct nitrox_softreq, in create_cipher_outbuf()
560 sr->out.sglist[cnt].iova = sr->in.sglist[i].iova; in create_aead_inplace_outbuf()
[all …]
/f-stack/dpdk/kernel/linux/kni/
H A Dkni_dev.h95 unsigned long iova) in iova_to_phys() argument
101 offset = iova & (PAGE_SIZE - 1); in iova_to_phys()
105 ret = get_user_pages_remote(tsk, tsk->mm, iova, 1, in iova_to_phys()
108 ret = get_user_pages_remote(tsk->mm, iova, 1, in iova_to_phys()
120 static inline void *iova_to_kva(struct task_struct *tsk, unsigned long iova) in iova_to_kva() argument
122 return phys_to_virt(iova_to_phys(tsk, iova)); in iova_to_kva()
/f-stack/dpdk/drivers/net/virtio/virtio_user/
H A Dvhost_vdpa.c58 uint64_t iova; member
85 uint64_t iova, size_t len) in vhost_vdpa_dma_map() argument
91 msg.iotlb.iova = iova; in vhost_vdpa_dma_map()
107 uint64_t iova, size_t len) in vhost_vdpa_dma_unmap() argument
113 msg.iotlb.iova = iova; in vhost_vdpa_dma_unmap()
135 return vhost_vdpa_dma_map(dev, ms->addr, ms->iova, len); in vhost_vdpa_map_contig()
149 if (ms->iova == RTE_BAD_IOVA) in vhost_vdpa_map()
156 return vhost_vdpa_dma_map(dev, ms->addr, ms->iova, ms->len); in vhost_vdpa_map()
H A Dvhost.h109 uint64_t iova, size_t len);
111 uint64_t iova, size_t len);
/f-stack/dpdk/lib/librte_eal/common/
H A Deal_common_memory.c335 rte_iova_t iova; member
343 if (vi->iova >= ms->iova && vi->iova < (ms->iova + ms->len)) { in find_virt()
344 size_t offset = vi->iova - ms->iova; in find_virt()
356 if (vi->iova >= ms->iova && vi->iova < (ms->iova + len)) { in find_virt_legacy()
357 size_t offset = vi->iova - ms->iova; in find_virt_legacy()
374 vi.iova = iova; in rte_mem_iova2virt()
439 ms->iova, in dump_memseg()
530 rte_iova_t iova; in check_iova() local
533 iova = (ms->iova + ms->len) - 1; in check_iova()
534 if (!(iova & *mask)) in check_iova()
[all …]
/f-stack/dpdk/drivers/compress/octeontx/
H A Dotx_zip.c42 uint64_t iova; in zipvf_q_init() local
56 iova = rte_mem_virt2iova(cmdq_addr); in zipvf_q_init()
58 cmdq->iova = iova; in zipvf_q_init()
61 que_sbuf_addr.s.ptr = (cmdq->iova >> 7); in zipvf_q_init()
119 ncp.s.addr = cmdq->iova; in zipvf_push_command()
/f-stack/dpdk/lib/librte_mempool/
H A Drte_mempool_ops_default.c92 unsigned int max_objs, void *vaddr, rte_iova_t iova, in rte_mempool_op_populate_helper() argument
129 (iova == RTE_BAD_IOVA) ? RTE_BAD_IOVA : (iova + off)); in rte_mempool_op_populate_helper()
139 void *vaddr, rte_iova_t iova, size_t len, in rte_mempool_op_populate_default() argument
143 return rte_mempool_op_populate_helper(mp, 0, max_objs, vaddr, iova, in rte_mempool_op_populate_default()
H A Drte_mempool.c156 void *obj, rte_iova_t iova) in mempool_add_elem() argument
164 hdr->iova = iova; in mempool_add_elem()
334 memhdr->iova = iova; in rte_mempool_populate_iova()
351 (iova == RTE_BAD_IOVA) ? RTE_BAD_IOVA : (iova + off), in rte_mempool_populate_iova()
392 rte_iova_t iova; in rte_mempool_populate_virt() local
403 iova = get_iova(addr + off); in rte_mempool_populate_virt()
417 iova_tmp != iova + phys_len) in rte_mempool_populate_virt()
478 rte_iova_t iova; in rte_mempool_populate_default() local
581 iova = mz->iova; in rte_mempool_populate_default()
583 iova = RTE_BAD_IOVA; in rte_mempool_populate_default()
[all …]
H A Drte_mempool_trace.h70 rte_iova_t iova, size_t len, void *free_cb, void *opaque),
74 rte_trace_point_emit_u64(iova);
135 void *vaddr, uint64_t iova, size_t len, void *obj_cb,
141 rte_trace_point_emit_u64(iova);
H A Drte_mempool_ops.c129 void *vaddr, rte_iova_t iova, size_t len, in rte_mempool_ops_populate() argument
137 rte_mempool_trace_ops_populate(mp, max_objs, vaddr, iova, len, obj_cb, in rte_mempool_ops_populate()
141 iova, len, obj_cb, in rte_mempool_ops_populate()
144 return ops->populate(mp, max_objs, vaddr, iova, len, obj_cb, in rte_mempool_ops_populate()
H A Drte_mempool.h141 rte_iova_t iova; /**< IO address of the object. */ member
187 rte_iova_t iova; /**< IO address of the chunk */ member
518 void *opaque, void *vaddr, rte_iova_t iova);
550 void *vaddr, rte_iova_t iova, size_t len,
592 void *vaddr, rte_iova_t iova, size_t len,
603 void *vaddr, rte_iova_t iova, size_t len,
824 void *vaddr, rte_iova_t iova, size_t len,
1090 rte_iova_t iova, size_t len, rte_mempool_memchunk_free_cb_t *free_cb,
1678 return hdr->iova; in rte_mempool_virt2iova()
/f-stack/dpdk/drivers/bus/dpaa/
H A Drte_dpaa_bus.h159 rte_iova_t iova; member
182 if (paddr >= ms->iova && paddr < in rte_dpaa_mem_ptov()
183 ms->iova + ms->len) in rte_dpaa_mem_ptov()
184 return RTE_PTR_ADD(ms->vaddr, (uintptr_t)(paddr - ms->iova)); in rte_dpaa_mem_ptov()
202 return ms->iova + RTE_PTR_DIFF(vaddr, ms->addr); in rte_dpaa_mem_vtop()
/f-stack/dpdk/drivers/mempool/octeontx2/
H A Dotx2_mempool_ops.c625 pool->stack_base = mz->iova; in npa_lf_aura_pool_pair_alloc()
819 otx2_npa_l1d_way_set_get(uint64_t iova) in otx2_npa_l1d_way_set_get() argument
821 return (iova >> rte_log2_u32(RTE_CACHE_LINE_SIZE)) & 0x7; in otx2_npa_l1d_way_set_get()
826 rte_iova_t iova, size_t len, in otx2_npa_populate() argument
837 if (iova == RTE_BAD_IOVA) in otx2_npa_populate()
850 iova += off; in otx2_npa_populate()
854 start_iova = iova; in otx2_npa_populate()
855 while (start_iova < iova + len) { in otx2_npa_populate()
862 iova); in otx2_npa_populate()
873 npa_lf_aura_op_range_set(mp->pool_id, iova, iova + len); in otx2_npa_populate()
[all …]
/f-stack/dpdk/drivers/mempool/octeontx/
H A Drte_mempool_octeontx.c157 void *vaddr, rte_iova_t iova, size_t len, in octeontx_fpavf_populate() argument
166 if (iova == RTE_BAD_IOVA) in octeontx_fpavf_populate()
178 iova += off; in octeontx_fpavf_populate()
190 max_objs, vaddr, iova, len, in octeontx_fpavf_populate()
/f-stack/dpdk/drivers/bus/fslmc/
H A Dfslmc_vfio.c182 .iova = 0x6030000, in vfio_map_irq_region()
223 iova_addr = ms->iova; in fslmc_memevent_cb()
283 dma_map.iova = iovaddr; in fslmc_map_dma()
285 dma_map.iova = dma_map.vaddr; in fslmc_map_dma()
324 dma_unmap.iova = vaddr; in fslmc_unmap_dma()
335 (uint64_t)dma_unmap.iova, (uint64_t)dma_unmap.size); in fslmc_unmap_dma()
354 if (ms->iova == RTE_BAD_IOVA) in fslmc_dmamap_seg()
357 ret = fslmc_map_dma(ms->addr_64, ms->iova, ms->len); in fslmc_dmamap_seg()
368 rte_fslmc_vfio_mem_dmamap(uint64_t vaddr, uint64_t iova, uint64_t size) in rte_fslmc_vfio_mem_dmamap() argument
391 dma_map.iova = iova; in rte_fslmc_vfio_mem_dmamap()
[all …]
/f-stack/dpdk/lib/librte_cryptodev/
H A Drte_crypto_sym.h37 rte_iova_t iova; member
60 rte_iova_t iova; member
944 vec[0].iova = rte_pktmbuf_iova_offset(mb, ofs); in rte_crypto_mbuf_to_vec()
959 vec[i].iova = rte_pktmbuf_iova(nseg); in rte_crypto_mbuf_to_vec()
/f-stack/dpdk/drivers/net/qede/base/
H A Dbcm_osal.c157 *phys = mz->iova; in osal_dma_alloc_coherent()
162 mz->len, (unsigned long)mz->iova, mz->addr, core_id); in osal_dma_alloc_coherent()
196 *phys = mz->iova; in osal_dma_alloc_coherent_aligned()
201 mz->len, (unsigned long)mz->iova, mz->addr, core_id); in osal_dma_alloc_coherent_aligned()
210 if (phys == ecore_mz_mapping[j]->iova) { in osal_dma_free_mem()
/f-stack/dpdk/lib/librte_eal/include/
H A Drte_vfio.h330 uint64_t iova, uint64_t len);
354 uint64_t iova, uint64_t len);
/f-stack/dpdk/drivers/crypto/qat/
H A Dqat_sym_hw_dp.c36 data[0].iova; in qat_sym_dp_parse_data_vec()
56 list->buffers[i].addr = data[i].iova; in qat_sym_dp_parse_data_vec()
132 aad_iova = aad->iova; in enqueue_one_aead_job()
136 aad_iova = aad->iova; in enqueue_one_aead_job()
149 aad_iova = iv->iova; in enqueue_one_aead_job()
210 auth_param->auth_res_addr = digest->iova; in enqueue_one_aead_job()
418 auth_param->auth_res_addr = digest->iova; in enqueue_one_auth_job()
424 auth_param->u1.aad_adr = auth_iv->iova; in enqueue_one_auth_job()
560 auth_param->u1.aad_adr = auth_iv->iova; in enqueue_one_chain_job()
571 auth_iova_end = cvec->iova + in enqueue_one_chain_job()
[all …]

123456