| /f-stack/dpdk/drivers/crypto/octeontx2/ |
| H A D | otx2_ipsec_po_ops.h | 63 struct rte_mbuf *m_src = sym_op->m_src; in process_outb_sa() local 76 dlen = rte_pktmbuf_pkt_len(m_src) + hdr_len; in process_outb_sa() 83 mdata = rte_pktmbuf_append(m_src, extend_tail + mdata_len); in process_outb_sa() 94 data = rte_pktmbuf_prepend(m_src, extend_head); in process_outb_sa() 107 hdr = (struct otx2_ipsec_po_out_hdr *)rte_pktmbuf_adj(m_src, in process_outb_sa() 124 req->ist.ei1 = rte_pktmbuf_iova(m_src); in process_outb_sa() 145 struct rte_mbuf *m_src = sym_op->m_src; in process_inb_sa() local 152 dlen = rte_pktmbuf_pkt_len(m_src); in process_inb_sa() 155 mdata = rte_pktmbuf_append(m_src, mdata_len); in process_inb_sa() 170 req->ist.ei1 = rte_pktmbuf_iova(m_src); in process_inb_sa()
|
| /f-stack/dpdk/drivers/crypto/aesni_gcm/ |
| H A D | aesni_gcm_pmd.c | 224 struct rte_mbuf *m_src = sym_op->m_src; in process_gcm_crypto_op() local 241 RTE_ASSERT(m_src != NULL); in process_gcm_crypto_op() 244 offset -= m_src->data_len; in process_gcm_crypto_op() 245 m_src = m_src->next; in process_gcm_crypto_op() 247 RTE_ASSERT(m_src != NULL); in process_gcm_crypto_op() 289 m_src = m_src->next; in process_gcm_crypto_op() 291 RTE_ASSERT(m_src != NULL); in process_gcm_crypto_op() 329 m_src = m_src->next; in process_gcm_crypto_op() 331 RTE_ASSERT(m_src != NULL); in process_gcm_crypto_op() 366 m_src = m_src->next; in process_gcm_crypto_op() [all …]
|
| /f-stack/dpdk/drivers/crypto/qat/ |
| H A D | qat_sym.c | 66 && (sym_op->m_dst != sym_op->m_src))) in qat_bpicipher_preprocess() 195 (op->sym->m_dst != op->sym->m_src)) || in qat_sym_build_request() 196 op->sym->m_src->nb_segs > 1)) { in qat_sym_build_request() 453 if (op->sym->m_src->nb_segs > 1 || in qat_sym_build_request() 465 (op->sym->m_dst != op->sym->m_src))) { in qat_sym_build_request() 485 rte_pktmbuf_headroom(op->sym->m_src)) in qat_sym_build_request() 491 rte_pktmbuf_iova_offset(op->sym->m_src, in qat_sym_build_request() 513 op->sym->m_src, auth_ofs) - src_buf_start; in qat_sym_build_request() 538 op->sym->m_src : op->sym->m_dst); in qat_sym_build_request() 582 ret = qat_sgl_fill_array(op->sym->m_src, in qat_sym_build_request() [all …]
|
| H A D | qat_sym.h | 98 last_block = (uint8_t *) rte_pktmbuf_mtod_offset(sym_op->m_src, in qat_bpicipher_postprocess() 151 crc_data = rte_pktmbuf_mtod_offset(sym_op->m_src, uint8_t *, in qat_crc_verify() 172 sym_op->m_src->nb_segs == 1) { in qat_crc_generate() 175 crc_data = rte_pktmbuf_mtod_offset(sym_op->m_src, uint8_t *, in qat_crc_generate()
|
| /f-stack/dpdk/lib/librte_port/ |
| H A D | rte_swx_port_source_sink.c | 142 struct rte_mbuf *m_dst, *m_src; in source_pkt_rx() local 146 m_src = p->pkts[p->pos]; in source_pkt_rx() 147 m_src_data = rte_pktmbuf_mtod(m_src, uint8_t *); in source_pkt_rx() 155 m_dst->data_len = m_src->data_len; in source_pkt_rx() 156 m_dst->pkt_len = m_src->pkt_len; in source_pkt_rx() 157 m_dst->data_off = m_src->data_off; in source_pkt_rx() 160 rte_memcpy(m_dst_data, m_src_data, m_src->data_len); in source_pkt_rx()
|
| H A D | rte_port_sym_crypto.c | 83 rte_pktmbuf_free(op->sym->m_src); in rte_port_sym_crypto_reader_rx() 87 pkts[n++] = op->sym->m_src; in rte_port_sym_crypto_reader_rx() 205 rte_pktmbuf_free(p->tx_buf[nb_tx]->sym->m_src); in send_burst() 416 rte_pktmbuf_free(p->tx_buf[nb_tx]->sym->m_src); in send_burst_nodrop()
|
| /f-stack/dpdk/lib/librte_vhost/ |
| H A D | vhost_crypto.c | 844 struct rte_mbuf *m_src = op->sym->m_src, *m_dst = op->sym->m_dst; in prepare_sym_cipher_op() local 866 m_src->buf_addr == NULL)) { in prepare_sym_cipher_op() 996 struct rte_mbuf *m_src = op->sym->m_src, *m_dst = op->sym->m_dst; in prepare_sym_chain_op() local 1021 if (unlikely(m_src->buf_iova == 0 || m_src->buf_addr == NULL)) { in prepare_sym_chain_op() 1108 digest_offset = m_src->data_len; in prepare_sym_chain_op() 1337 struct rte_mbuf *m_src = op->sym->m_src; in vhost_crypto_finalize_one_request() local 1362 rte_mempool_put(m_src->pool, (void *)m_src); in vhost_crypto_finalize_one_request() 1651 op->sym->m_src = mbufs[i * 2]; in rte_vhost_crypto_fetch_requests() 1653 op->sym->m_src->data_off = 0; in rte_vhost_crypto_fetch_requests() 1681 op->sym->m_src = mbufs[i]; in rte_vhost_crypto_fetch_requests() [all …]
|
| /f-stack/dpdk/app/test-crypto-perf/ |
| H A D | cperf_ops.c | 35 sym_op->m_src = (struct rte_mbuf *)((uint8_t *)ops[i] + in cperf_set_ops_security() 39 sym_op->m_src->buf_len = options->segment_sz; in cperf_set_ops_security() 40 sym_op->m_src->data_len = options->test_buffer_size; in cperf_set_ops_security() 41 sym_op->m_src->pkt_len = sym_op->m_src->data_len; in cperf_set_ops_security() 51 sym_op->m_src->buf_len = options->segment_sz; in cperf_set_ops_security() 52 sym_op->m_src->data_len = buf_sz; in cperf_set_ops_security() 53 sym_op->m_src->pkt_len = buf_sz; in cperf_set_ops_security() 97 sym_op->m_src = (struct rte_mbuf *)((uint8_t *)ops[i] + in cperf_set_ops_null_cipher() 262 tbuf = sym_op->m_src; in cperf_set_ops_auth() 369 tbuf = sym_op->m_src; in cperf_set_ops_cipher_auth() [all …]
|
| H A D | cperf_test_verify.c | 113 m = op->sym->m_src; in cperf_verify_op() 129 m = op->sym->m_src; in cperf_verify_op() 307 cperf_mbuf_set(ops[i]->sym->m_src, in cperf_verify_test_runner() 318 rte_pktmbuf_linearize(ops[i]->sym->m_src); in cperf_verify_test_runner()
|
| /f-stack/dpdk/drivers/crypto/virtio/ |
| H A D | virtqueue.c | 35 if (cop->sym->m_src) in virtqueue_detatch_unused() 36 rte_pktmbuf_free(cop->sym->m_src); in virtqueue_detatch_unused()
|
| /f-stack/dpdk/app/test/ |
| H A D | test_cryptodev_blockcipher.c | 293 sym_op->m_src = ibuf; in test_blockcipher_one_case() 430 (sym_op->m_src, digest_offset); in test_blockcipher_one_case() 432 pktmbuf_iova_offset(sym_op->m_src, in test_blockcipher_one_case() 469 sym_op->m_src->buf_addr, sym_op->m_src->buf_len); in test_blockcipher_one_case() 470 rte_memcpy(tmp_src_buf, sym_op->m_src->buf_addr, in test_blockcipher_one_case() 471 sym_op->m_src->buf_len); in test_blockcipher_one_case() 513 sym_op->m_src->buf_addr, sym_op->m_src->buf_len); in test_blockcipher_one_case() 586 mbuf = sym_op->m_src; in test_blockcipher_one_case() 670 mbuf = sym_op->m_src; in test_blockcipher_one_case()
|
| H A D | test_event_crypto_adapter.c | 123 struct rte_mbuf *m = op->sym->m_src; in send_recv_ev() 126 rte_pktmbuf_free(op->sym->m_src); in send_recv_ev() 237 sym_op->m_src = m; in test_op_forward_mode() 353 struct rte_mbuf *m = recv_op->sym->m_src; in send_op_recv_ev() 356 rte_pktmbuf_free(recv_op->sym->m_src); in send_op_recv_ev() 429 sym_op->m_src = m; in test_op_new_mode()
|
| /f-stack/dpdk/drivers/crypto/caam_jr/ |
| H A D | caam_jr.c | 643 ctx->op->sym->m_src->pkt_len = in hw_poll_job_ring() 645 ctx->op->sym->m_src->data_len = in hw_poll_job_ring() 713 struct rte_mbuf *mbuf = sym->m_src; in build_auth_only_sg() 878 mbuf = sym->m_src; in build_cipher_only_sg() 912 mbuf = sym->m_src; in build_cipher_only_sg() 940 mbuf = sym->m_src; in build_cipher_only_sg() 1073 mbuf = sym->m_src; in build_cipher_auth_sg() 1102 mbuf = sym->m_src; in build_cipher_auth_sg() 1142 mbuf = sym->m_src; in build_cipher_auth_sg() 1337 sym->m_src->buf_len - sym->m_src->data_off); in build_proto() [all …]
|
| /f-stack/dpdk/drivers/compress/octeontx/ |
| H A D | otx_zip.h | 167 struct rte_mbuf *m_src; in zipvf_prepare_in_buf() local 172 m_src = op->m_src; in zipvf_prepare_in_buf() 177 rte_pktmbuf_iova_offset(m_src, offset); in zipvf_prepare_in_buf()
|
| /f-stack/dpdk/drivers/crypto/dpaa2_sec/ |
| H A D | dpaa2_sec_dpseci.c | 78 mbuf = sym_op->m_src; in build_proto_compound_sg_fd() 138 mbuf = sym_op->m_src; in build_proto_compound_sg_fd() 309 mbuf = sym_op->m_src; in build_authenc_gcm_sg_fd() 379 mbuf = sym_op->m_src; in build_authenc_gcm_sg_fd() 454 dst = sym_op->m_src; in build_authenc_gcm_fd() 598 mbuf = sym_op->m_src; in build_authenc_sg_fd() 671 mbuf = sym_op->m_src; in build_authenc_sg_fd() 746 dst = sym_op->m_src; in build_authenc_fd() 894 mbuf = sym_op->m_src; in build_auth_sg_fd() 1185 mbuf = sym_op->m_src; in build_cipher_sg_fd() [all …]
|
| /f-stack/dpdk/drivers/compress/isal/ |
| H A D | isal_compress_pmd.c | 236 struct rte_mbuf *src = op->m_src; in chained_mbuf_compression() 326 struct rte_mbuf *src = op->m_src; in chained_mbuf_decompression() 449 if (op->m_src->pkt_len < (op->src.length + op->src.offset)) { in process_isal_deflate() 463 if (op->m_src->nb_segs > 1 || op->m_dst->nb_segs > 1) { in process_isal_deflate() 472 qp->stream->next_in = rte_pktmbuf_mtod_offset(op->m_src, in process_isal_deflate() 541 if (op->m_src->pkt_len < (op->src.length + op->src.offset)) { in process_isal_inflate() 555 if (op->m_src->nb_segs > 1 || op->m_dst->nb_segs > 1) { in process_isal_inflate() 563 qp->state->next_in = rte_pktmbuf_mtod_offset(op->m_src, in process_isal_inflate()
|
| /f-stack/dpdk/drivers/crypto/dpaa_sec/ |
| H A D | dpaa_sec.c | 950 mbuf = sym->m_src; in build_cipher_only_sg() 992 mbuf = sym->m_src; in build_cipher_only_sg() 1114 mbuf = sym->m_src; in build_cipher_auth_gcm_sg() 1176 mbuf = sym->m_src; in build_cipher_auth_gcm_sg() 1247 src_start_addr = sym->m_src->buf_iova + sym->m_src->data_off; in build_cipher_auth_gcm() 1363 mbuf = sym->m_src; in build_cipher_auth_sg() 1422 mbuf = sym->m_src; in build_cipher_auth_sg() 1485 src_start_addr = sym->m_src->buf_iova + sym->m_src->data_off; in build_cipher_auth() 1602 sg->length = sym->m_src->buf_len - sym->m_src->data_off; in build_proto() 1622 mbuf = sym->m_src; in build_proto_sg() [all …]
|
| /f-stack/dpdk/drivers/common/cpt/ |
| H A D | cpt_ucode.h | 2821 if (likely(m_src && (m_src->nb_segs == 1))) { in alloc_op_meta() 2828 mdata = (uint8_t *)m_src->buf_addr + m_src->buf_len; in alloc_op_meta() 2829 mphys = m_src->buf_iova + m_src->buf_len; in alloc_op_meta() 2841 RTE_SET_USED(m_src); in alloc_op_meta() 3001 struct rte_mbuf *m_src, *m_dst; in fill_fc_params() local 3035 m_src = sym_op->m_src; in fill_fc_params() 3074 m_src; in fill_fc_params() 3077 m = m_src; in fill_fc_params() 3120 m = m_src; in fill_fc_params() 3321 struct rte_mbuf *m_src, *m_dst; in fill_digest_params() local [all …]
|
| /f-stack/dpdk/drivers/crypto/kasumi/ |
| H A D | rte_kasumi_pmd.c | 185 src[i] = rte_pktmbuf_mtod(ops[i]->sym->m_src, uint8_t *) + in process_kasumi_cipher_op() 190 rte_pktmbuf_mtod(ops[i]->sym->m_src, uint8_t *) + in process_kasumi_cipher_op() 218 src = rte_pktmbuf_mtod(op->sym->m_src, uint8_t *); in process_kasumi_cipher_op_bit() 256 src = rte_pktmbuf_mtod(ops[i]->sym->m_src, uint8_t *) + in process_kasumi_hash_op() 414 if (!rte_pktmbuf_is_contiguous(curr_c_op->sym->m_src) || in kasumi_pmd_enqueue_burst()
|
| /f-stack/dpdk/lib/librte_eventdev/ |
| H A D | rte_event_crypto_adapter.c | 356 rte_pktmbuf_free(crypto_op->sym->m_src); in eca_enq_to_cryptodev() 365 rte_pktmbuf_free(crypto_op->sym->m_src); in eca_enq_to_cryptodev() 381 rte_pktmbuf_free(crypto_op->sym->m_src); in eca_enq_to_cryptodev() 389 rte_pktmbuf_free(crypto_op->sym->m_src); in eca_enq_to_cryptodev() 407 rte_pktmbuf_free(op->sym->m_src); in eca_enq_to_cryptodev() 458 rte_pktmbuf_free(op->sym->m_src); in eca_crypto_enq_flush() 532 rte_pktmbuf_free(ops[i]->sym->m_src); in eca_ops_enqueue_burst() 557 rte_pktmbuf_free(op->sym->m_src); in eca_ops_enqueue_burst()
|
| /f-stack/dpdk/drivers/compress/zlib/ |
| H A D | zlib_pmd.c | 22 struct rte_mbuf *mbuf_src = op->m_src; in process_zlib_deflate() 112 struct rte_mbuf *mbuf_src = op->m_src; in process_zlib_inflate() 202 (op->src.offset > rte_pktmbuf_data_len(op->m_src)) || in process_zlib_op()
|
| /f-stack/dpdk/drivers/crypto/snow3g/ |
| H A D | rte_snow3g_pmd.c | 196 src[i] = rte_pktmbuf_mtod(ops[i]->sym->m_src, uint8_t *) + in process_snow3g_cipher_op() 201 rte_pktmbuf_mtod(ops[i]->sym->m_src, uint8_t *) + in process_snow3g_cipher_op() 227 src = rte_pktmbuf_mtod(op->sym->m_src, uint8_t *); in process_snow3g_cipher_op_bit() 266 src = rte_pktmbuf_mtod(ops[i]->sym->m_src, uint8_t *) + in process_snow3g_hash_op() 305 if (!rte_pktmbuf_is_contiguous(ops[i]->sym->m_src) || in process_ops()
|
| /f-stack/dpdk/drivers/crypto/aesni_mb/ |
| H A D | rte_aesni_mb_pmd.c | 1062 struct rte_mbuf *m_src, *m_dst; in auth_start_offset() local 1071 m_src = op->sym->m_src; in auth_start_offset() 1074 p_src = rte_pktmbuf_mtod(m_src, uint8_t *); in auth_start_offset() 1246 struct rte_mbuf *m_src = op->sym->m_src, *m_dst; in set_mb_job_params() local 1375 m_dst = m_src; in set_mb_job_params() 1377 } else if (op->sym->m_dst == op->sym->m_src) { in set_mb_job_params() 1379 m_dst = m_src; in set_mb_job_params() 1415 job->src = rte_pktmbuf_mtod(m_src, uint8_t *); in set_mb_job_params() 1503 struct rte_mbuf *m_src, *m_dst; in set_sec_mb_job_params() local 1521 m_src = sym->m_src; in set_sec_mb_job_params() [all …]
|
| /f-stack/dpdk/drivers/compress/qat/ |
| H A D | qat_comp.c | 203 if (op->m_src->next != NULL || op->m_dst->next != NULL) { in qat_comp_build_request() 210 if (unlikely(op->m_src->nb_segs > cookie->src_nb_elems)) { in qat_comp_build_request() 217 op->m_src->nb_segs, 64, in qat_comp_build_request() 223 op->m_src->nb_segs); in qat_comp_build_request() 232 cookie->src_nb_elems = op->m_src->nb_segs; in qat_comp_build_request() 237 ret = qat_sgl_fill_array(op->m_src, in qat_comp_build_request() 308 rte_pktmbuf_iova_offset(op->m_src, op->src.offset); in qat_comp_build_request()
|
| /f-stack/dpdk/drivers/crypto/zuc/ |
| H A D | rte_zuc_pmd.c | 198 if (!rte_pktmbuf_is_contiguous(ops[i]->sym->m_src) || in process_zuc_cipher_op() 210 src[i] = rte_pktmbuf_mtod(ops[i]->sym->m_src, uint8_t *) + in process_zuc_cipher_op() 215 rte_pktmbuf_mtod(ops[i]->sym->m_src, uint8_t *) + in process_zuc_cipher_op() 260 src[i] = rte_pktmbuf_mtod(ops[i]->sym->m_src, uint8_t *) + in process_zuc_hash_op()
|