| /dpdk/drivers/crypto/dpaa_sec/ |
| H A D | dpaa_sec_raw_dp.c | 91 struct rte_crypto_sgl *sgl, in build_dpaa_raw_dp_auth_fd() argument 115 for (i = 0; i < sgl->num; i++) in build_dpaa_raw_dp_auth_fd() 116 total_len += sgl->vec[i].len; in build_dpaa_raw_dp_auth_fd() 223 struct rte_crypto_sgl *sgl, in build_raw_cipher_auth_gcm_sg() argument 242 for (i = 0; i < sgl->num; i++) in build_raw_cipher_auth_gcm_sg() 243 data_len += sgl->vec[i].len; in build_raw_cipher_auth_gcm_sg() 380 struct rte_crypto_sgl *sgl, in build_dpaa_raw_dp_chain_fd() argument 405 data_len += sgl->vec[i].len; in build_dpaa_raw_dp_chain_fd() 532 struct rte_crypto_sgl *sgl, in build_dpaa_raw_dp_cipher_fd() argument 651 struct rte_crypto_sgl *sgl, in build_dpaa_raw_proto_sg() argument [all …]
|
| H A D | dpaa_sec.h | 138 struct rte_crypto_sgl *sgl,
|
| /dpdk/drivers/crypto/dpaa2_sec/ |
| H A D | dpaa2_sec_raw_dp.c | 26 struct rte_crypto_sgl *sgl, in build_raw_dp_chain_fd() argument 53 for (i = 0; i < sgl->num; i++) in build_raw_dp_chain_fd() 54 data_len += sgl->vec[i].len; in build_raw_dp_chain_fd() 62 FLE_SG_MEM_SIZE(2 * sgl->num), in build_raw_dp_chain_fd() 121 sge->length = sgl->vec[i].len; in build_raw_dp_chain_fd() 155 for (i = 1; i < sgl->num; i++) { in build_raw_dp_chain_fd() 159 sge->length = sgl->vec[i].len; in build_raw_dp_chain_fd() 204 for (i = 0; i < sgl->num; i++) in build_raw_dp_aead_fd() 205 data_len += sgl->vec[i].len; in build_raw_dp_aead_fd() 356 for (i = 0; i < sgl->num; i++) in build_raw_dp_auth_fd() [all …]
|
| H A D | dpaa2_sec_priv.h | 182 struct rte_crypto_sgl *sgl,
|
| /dpdk/drivers/compress/qat/ |
| H A D | qat_comp_pmd.c | 255 struct qat_inter_sgl *sgl; in qat_comp_setup_inter_buffers() local 293 sgl = (struct qat_inter_sgl *) mz->addr; in qat_comp_setup_inter_buffers() 294 sgl->num_bufs = QAT_NUM_BUFS_IN_IM_SGL; in qat_comp_setup_inter_buffers() 295 sgl->num_mapped_bufs = 0; in qat_comp_setup_inter_buffers() 296 sgl->resrvd = 0; in qat_comp_setup_inter_buffers() 301 QAT_LOG(DEBUG, " : virt address of sgl[%i] = %p", i, sgl); in qat_comp_setup_inter_buffers() 304 sgl->buffers[lb].addr = in qat_comp_setup_inter_buffers() 307 sgl->buffers[lb].len = buff_size; in qat_comp_setup_inter_buffers() 308 sgl->buffers[lb].resrvd = 0; in qat_comp_setup_inter_buffers() 312 lb, sgl->buffers[lb].addr, sgl->buffers[lb].len); in qat_comp_setup_inter_buffers()
|
| /dpdk/doc/guides/tools/ |
| H A D | comp_perf.rst | 30 if the max-num-sgl-segs x seg_sz > input size then segments number in 31 the chain will be lower than value passed into max-num-sgl-segs. 81 ``--max-num-sgl-segs N``: maximum number of segments for each mbuf (default: 16) 108 …--compress-level 1:1:9 --num-iter 10 --extended-input-sz 1048576 --max-num-sgl-segs 16 --huffman-…
|
| /dpdk/drivers/crypto/ipsec_mb/ |
| H A D | pmd_aesni_gcm.c | 581 struct rte_crypto_sgl *sgl, void *iv, void *aad, in aesni_gcm_process_gcm_sgl_op() argument 591 for (i = 0; i < sgl->num; i++) { in aesni_gcm_process_gcm_sgl_op() 592 struct rte_crypto_vec *vec = &sgl->vec[i]; in aesni_gcm_process_gcm_sgl_op() 614 struct rte_crypto_sgl *sgl, void *iv, in aesni_gcm_process_gmac_sgl_op() argument 617 ops.init(&s->gdata_key, gdata_ctx, iv, sgl->vec[0].base, in aesni_gcm_process_gmac_sgl_op() 618 sgl->vec[0].len); in aesni_gcm_process_gmac_sgl_op()
|
| H A D | pmd_aesni_mb.c | 1824 check_crypto_sgl(union rte_crypto_sym_ofs so, const struct rte_crypto_sgl *sgl) in check_crypto_sgl() argument 1827 if (sgl->num != 1) in check_crypto_sgl() 1829 else if (so.ofs.cipher.head + so.ofs.cipher.tail > sgl->vec[0].len) in check_crypto_sgl()
|
| /dpdk/drivers/net/cxgbe/ |
| H A D | sge.c | 547 struct ulptx_sgl *sgl, u64 *end, unsigned int start, in write_sgl() argument 557 sgl->len0 = htonl(len); in write_sgl() 558 sgl->addr0 = rte_cpu_to_be_64(addr[0]); in write_sgl() 560 sgl->cmd_nsge = htonl(V_ULPTX_CMD(ULP_TX_SC_DSGL) | in write_sgl() 569 to = (u8 *)end > (u8 *)q->stat ? buf : sgl->sge; in write_sgl() 587 (u8 *)sgl->sge); in write_sgl() 591 memcpy(sgl->sge, buf, part0); in write_sgl() 1016 sd->coalesce.sgl[idx & 1] = (struct ulptx_sgl *)(cpl + 1); in tx_do_packet_coalesce() 1238 txq->q.sdesc[last_desc].sgl = (struct ulptx_sgl *)(cpl + 1); in t4_eth_xmit()
|
| /dpdk/drivers/net/cxgbe/base/ |
| H A D | adapter.h | 182 struct ulptx_sgl *sgl[ETH_COALESCE_PKT_PER_DESC]; member 192 struct ulptx_sgl *sgl; member
|
| /dpdk/app/test/ |
| H A D | test_cryptodev.c | 243 sgl.vec = data_vec; in process_sym_raw_dp_op() 245 vec.src_sgl = &sgl; in process_sym_raw_dp_op() 366 sgl.num = n; in process_sym_raw_dp_op() 434 struct rte_crypto_sgl sgl; in process_cpu_aead_op() local 449 sgl.vec = vec; in process_cpu_aead_op() 450 sgl.num = n; in process_cpu_aead_op() 451 symvec.src_sgl = &sgl; in process_cpu_aead_op() 480 struct rte_crypto_sgl sgl; in process_cpu_crypt_auth_op() local 495 sgl.vec = vec; in process_cpu_crypt_auth_op() 496 sgl.num = n; in process_cpu_crypt_auth_op() [all …]
|
| /dpdk/drivers/net/bnx2x/ |
| H A D | ecore_hsi.h | 4536 __le16 sgl[8]; member
|