| /dpdk/drivers/common/iavf/ |
| H A D | iavf_impl.c | 37 mem->va = mz->addr; in iavf_allocate_dma_mem_d() 53 mem->va = NULL; in iavf_free_dma_mem_d() 68 mem->va = rte_zmalloc("iavf", size, 0); in iavf_allocate_virt_mem_d() 70 if (mem->va) in iavf_allocate_virt_mem_d() 83 rte_free(mem->va); in iavf_free_virt_mem_d() 84 mem->va = NULL; in iavf_free_virt_mem_d()
|
| H A D | iavf_adminq.h | 13 (&(((struct iavf_aq_desc *)((R).desc_buf.va))[i])) 54 (&(((struct iavf_asq_cmd_details *)((R).cmd_buf.va))[i]))
|
| H A D | iavf_osdep.h | 162 void *va; member 169 void *va; member
|
| /dpdk/lib/mempool/ |
| H A D | rte_mempool_ops_default.c | 96 char *va = vaddr; in rte_mempool_op_populate_helper() local 110 off = total_elt_sz - (((uintptr_t)(va - 1) % total_elt_sz) + 1); in rte_mempool_op_populate_helper() 115 if (check_obj_bounds(va + off, pg_sz, total_elt_sz) < 0) { in rte_mempool_op_populate_helper() 116 off += RTE_PTR_ALIGN_CEIL(va + off, pg_sz) - (va + off); in rte_mempool_op_populate_helper() 119 (((uintptr_t)(va + off - 1) % in rte_mempool_op_populate_helper() 127 obj = va + off; in rte_mempool_op_populate_helper()
|
| /dpdk/drivers/bus/dpaa/ |
| H A D | rte_dpaa_bus.h | 172 void *va; in rte_dpaa_mem_ptov() local 174 va = dpaax_iova_table_get_va(paddr); in rte_dpaa_mem_ptov() 175 if (likely(va != NULL)) in rte_dpaa_mem_ptov() 176 return va; in rte_dpaa_mem_ptov() 188 va = rte_mem_iova2virt(paddr); in rte_dpaa_mem_ptov() 190 dpaax_iova_table_update(paddr, va, RTE_CACHE_LINE_SIZE); in rte_dpaa_mem_ptov() 192 return va; in rte_dpaa_mem_ptov()
|
| /dpdk/drivers/compress/octeontx/ |
| H A D | otx_zip.c | 55 cmdq->va = (uint8_t *)cmdq_addr; in zipvf_q_init() 70 memset(cmdq->va, 0, ZIP_MAX_CMDQ_SIZE); in zipvf_q_init() 83 if (cmdq->va != NULL) { in zipvf_q_term() 84 memset(cmdq->va, 0, ZIP_MAX_CMDQ_SIZE); in zipvf_q_term() 85 rte_free(cmdq->va); in zipvf_q_term() 108 if ((((cmdq->sw_head - (uint64_t *)cmdq->va) * sizeof(uint64_t *)) + in zipvf_push_command() 120 cmdq->sw_head = (uint64_t *)cmdq->va; in zipvf_push_command()
|
| /dpdk/drivers/crypto/caam_jr/ |
| H A D | caam_jr_pvt.h | 259 void *va; in caam_jr_dma_ptov() local 260 va = dpaax_iova_table_get_va(paddr); in caam_jr_dma_ptov() 261 if (likely(va != NULL)) in caam_jr_dma_ptov() 262 return va; in caam_jr_dma_ptov()
|
| /dpdk/drivers/crypto/qat/dev/ |
| H A D | qat_crypto_pmd_gens.h | 254 cipher_iv->va = rte_crypto_op_ctod_offset(op, void *, in qat_sym_convert_op_to_vec_cipher() 303 auth_iv->va = rte_crypto_op_ctod_offset(op, void *, in qat_sym_convert_op_to_vec_auth() 321 auth_iv->va = NULL; in qat_sym_convert_op_to_vec_auth() 357 digest->va = (void *)op->sym->auth.digest.data; in qat_sym_convert_op_to_vec_auth() 392 digest->va = (void *)op->sym->auth.digest.data; in qat_sym_convert_op_to_vec_chain() 512 digest->va = (void *)op->sym->aead.digest.data; in qat_sym_convert_op_to_vec_aead() 743 aad_data = aad->va; in enqueue_one_aead_job_gen1() 756 aad_data = iv->va; in enqueue_one_aead_job_gen1() 799 (uint8_t *)iv->va + in enqueue_one_aead_job_gen1() 804 rte_memcpy((uint8_t *)aad->va + in enqueue_one_aead_job_gen1() [all …]
|
| /dpdk/lib/ipsec/ |
| H A D | esp_outb.c | 35 sop->auth.digest.data = icv->va; in sop_ciph_auth_prepare() 50 sop->aead.digest.data = icv->va; in sop_aead_prepare() 52 sop->aead.aad.data = icv->va + sa->icv_len; in sop_aead_prepare() 231 icv->va = rte_pktmbuf_mtod_offset(ml, void *, pdofs); in outb_tun_pkt_prepare() 252 psqh = (uint32_t *)(icv->va - sa->sqh_len); in outb_pkt_xprepare() 263 gaad = (struct aead_gcm_aad *)(icv->va + sa->icv_len); in outb_pkt_xprepare() 276 (icv->va + sa->icv_len); in outb_pkt_xprepare() 428 icv->va = rte_pktmbuf_mtod_offset(ml, void *, pdofs); in outb_trs_pkt_prepare() 579 iv[k].va = ivbuf[k]; in cpu_outb_pkt_prepare() 580 aad[k].va = icv.va + sa->icv_len; in cpu_outb_pkt_prepare() [all …]
|
| H A D | esp_inb.c | 34 sop->auth.digest.data = icv->va; in sop_ciph_auth_prepare() 49 sop->aead.digest.data = icv->va; in sop_aead_prepare() 51 sop->aead.aad.data = icv->va + sa->icv_len; in sop_aead_prepare() 222 insert_sqh(sqn_hi32(sqc), icv->va, sa->icv_len); in inb_pkt_xprepare() 231 aad = (struct aead_gcm_aad *)(icv->va + sa->icv_len); in inb_pkt_xprepare() 237 caad = (struct aead_ccm_aad *)(icv->va + sa->icv_len); in inb_pkt_xprepare() 244 (icv->va + sa->icv_len); in inb_pkt_xprepare() 326 icv->va = rte_pktmbuf_mtod_offset(ml, void *, icv_ofs); in inb_prepare() 790 iv[k].va = ivbuf[k]; in cpu_inb_pkt_prepare() 791 aad[k].va = icv.va + sa->icv_len; in cpu_inb_pkt_prepare() [all …]
|
| /dpdk/kernel/linux/kni/ |
| H A D | kni_net.c | 67 void *va; in pa2va() local 69 va = (void *)((unsigned long)pa + in pa2va() 72 return va; in pa2va() 248 kni->va[i] = pa2va(kni->pa[i], kva); in kni_fifo_trans_pa2va() 262 ret = kni_fifo_put(dst_va, kni->va, num_rx); in kni_fifo_trans_pa2va() 410 kni->va[i] = pa2va(kni->pa[i], kva); in kni_net_rx_normal() 452 ret = kni_fifo_put(kni->free_q, kni->va, num_rx); in kni_net_rx_normal() 509 kni->va[i] = pa2va(kni->pa[i], kva); in kni_net_rx_lo_fifo() 538 ret = kni_fifo_put(kni->free_q, kni->va, num); in kni_net_rx_lo_fifo() 589 kni->va[i] = pa2va(kni->pa[i], kva); in kni_net_rx_lo_fifo_skb() [all …]
|
| /dpdk/lib/eal/windows/ |
| H A D | eal_lcore.c | 45 va_list va; in log_early() local 47 va_start(va, format); in log_early() 48 vfprintf(stderr, format, va); in log_early() 49 va_end(va); in log_early()
|
| /dpdk/drivers/bus/fslmc/portal/ |
| H A D | dpaa2_hw_pvt.h | 381 void *va; in dpaa2_mem_ptov() local 386 va = (void *)dpaax_iova_table_get_va(paddr); in dpaa2_mem_ptov() 387 if (likely(va != NULL)) in dpaa2_mem_ptov() 388 return va; in dpaa2_mem_ptov() 391 va = rte_mem_iova2virt(paddr); in dpaa2_mem_ptov() 393 return va; in dpaa2_mem_ptov()
|
| /dpdk/drivers/crypto/ipsec_mb/ |
| H A D | pmd_aesni_gcm.c | 632 vec->iv[i].va, vec->aad[i].va, in aesni_gcm_sgl_encrypt() 635 s, gdata_ctx, vec->digest[i].va, ops); in aesni_gcm_sgl_encrypt() 653 vec->iv[i].va, vec->aad[i].va, in aesni_gcm_sgl_decrypt() 656 s, gdata_ctx, vec->digest[i].va, ops); in aesni_gcm_sgl_decrypt() 679 vec->iv[i].va, ops); in aesni_gmac_sgl_generate() 681 s, gdata_ctx, vec->digest[i].va, ops); in aesni_gmac_sgl_generate() 704 vec->iv[i].va, ops); in aesni_gmac_sgl_verify() 706 s, gdata_ctx, vec->digest[i].va, ops); in aesni_gmac_sgl_verify()
|
| /dpdk/drivers/crypto/bcmfs/ |
| H A D | bcmfs_sym_engine.h | 17 void *va; member 22 #define fsattr_va(__ptr) ((__ptr)->va)
|
| /dpdk/drivers/net/octeontx/base/ |
| H A D | octeontx_pkovf.c | 21 uint8_t *va; member 501 pko_vf_ctl.fc_iomem.va = rte_malloc(NULL, fc_mem_size, 128); in octeontx_pko_init_fc() 502 if (unlikely(!pko_vf_ctl.fc_iomem.va)) { in octeontx_pko_init_fc() 508 pko_vf_ctl.fc_iomem.va); in octeontx_pko_init_fc() 512 (struct octeontx_pko_fc_ctl_s *)pko_vf_ctl.fc_iomem.va; in octeontx_pko_init_fc() 541 rte_free(pko_vf_ctl.fc_iomem.va); in octeontx_pko_fc_free()
|
| /dpdk/drivers/net/ice/base/ |
| H A D | ice_controlq.c | 100 cq->sq.desc_buf.va = ice_alloc_dma_mem(hw, &cq->sq.desc_buf, size); in ice_alloc_ctrlq_sq_ring() 101 if (!cq->sq.desc_buf.va) in ice_alloc_ctrlq_sq_ring() 124 cq->rq.desc_buf.va = ice_alloc_dma_mem(hw, &cq->rq.desc_buf, size); in ice_alloc_ctrlq_rq_ring() 125 if (!cq->rq.desc_buf.va) in ice_alloc_ctrlq_rq_ring() 168 bi->va = ice_alloc_dma_mem(hw, bi, cq->rq_buf_size); in ice_alloc_rq_bufs() 169 if (!bi->va) in ice_alloc_rq_bufs() 229 bi->va = ice_alloc_dma_mem(hw, bi, cq->sq_buf_size); in ice_alloc_sq_bufs() 230 if (!bi->va) in ice_alloc_sq_bufs() 1015 ice_memcpy(dma_buf->va, buf, buf_size, ICE_NONDMA_TO_DMA); in ice_sq_send_cmd_nolock() 1058 ice_memcpy(buf, dma_buf->va, copy_size, in ice_sq_send_cmd_nolock() [all …]
|
| H A D | ice_osdep.h | 191 void *va; member 198 void *va; member 271 mem->va = mz->addr; in ice_alloc_dma_mem() 277 return mem->va; in ice_alloc_dma_mem() 289 mem->va = NULL; in ice_free_dma_mem()
|
| /dpdk/drivers/net/i40e/base/ |
| H A D | i40e_adminq.h | 13 (&(((struct i40e_aq_desc *)((R).desc_buf.va))[i])) 54 (&(((struct i40e_asq_cmd_details *)((R).cmd_buf.va))[i]))
|
| H A D | i40e_hmc.c | 69 sd_entry->u.pd_table.pd_entry_virt_mem.va; in i40e_add_sd_table_entry() 163 pd_addr = (u64 *)pd_table->pd_page_addr.va; in i40e_add_pd_table_entry() 229 pd_addr = (u64 *)pd_table->pd_page_addr.va; in i40e_remove_pd_bp()
|
| H A D | i40e_osdep.h | 181 void *va; member 192 void *va; member
|
| /dpdk/drivers/common/dpaax/ |
| H A D | dpaax_iova_table.c | 406 const void *va = RTE_PTR_ADD(addr, cur_len); in dpaax_memevent_cb() local 408 ms = rte_mem_virt2memseg(va, msl); in dpaax_memevent_cb() 417 va, virt_addr, phys_addr, map_len); in dpaax_memevent_cb() 431 type, va, map_len, ret); in dpaax_memevent_cb()
|
| /dpdk/drivers/gpu/cuda/ |
| H A D | gdrcopy.c | 14 static int (*sym_gdr_map)(gdr_t g, gdr_mh_t handle, void **va, size_t size); 15 static int (*sym_gdr_unmap)(gdr_t g, gdr_mh_t handle, void *va, size_t size);
|
| /dpdk/drivers/crypto/qat/ |
| H A D | qat_sym.h | 388 QAT_DP_HEXDUMP_LOG(DEBUG, "cipher iv:", cipher_iv->va, in qat_sym_debug_log_dump() 391 QAT_DP_HEXDUMP_LOG(DEBUG, "auth iv:", auth_iv->va, in qat_sym_debug_log_dump() 394 QAT_DP_HEXDUMP_LOG(DEBUG, "aad:", aad->va, in qat_sym_debug_log_dump() 397 QAT_DP_HEXDUMP_LOG(DEBUG, "digest:", digest->va, in qat_sym_debug_log_dump()
|
| /dpdk/drivers/bus/vmbus/linux/ |
| H A D | vmbus_uio.c | 141 void *va = NULL; in vmbus_find_max_end_va() local 143 rte_memseg_list_walk(find_max_end_va, &va); in vmbus_find_max_end_va() 144 return va; in vmbus_find_max_end_va()
|