| /dpdk/drivers/crypto/dpaa_sec/ |
| H A D | dpaa_sec_raw_dp.c | 118 data_len = total_len - ofs.ofs.auth.head - ofs.ofs.auth.tail; in build_dpaa_raw_dp_auth_fd() 119 data_offset = ofs.ofs.auth.head; in build_dpaa_raw_dp_auth_fd() 246 aead_len = data_len - ofs.ofs.cipher.head - ofs.ofs.cipher.tail; in build_raw_cipher_auth_gcm_sg() 283 sg->offset = ofs.ofs.cipher.head; in build_raw_cipher_auth_gcm_sg() 296 sg->offset = ofs.ofs.cipher.head; in build_raw_cipher_auth_gcm_sg() 351 sg->offset = ofs.ofs.cipher.head; in build_raw_cipher_auth_gcm_sg() 399 ofs.ofs.auth.head; in build_dpaa_raw_dp_chain_fd() 407 cipher_len = data_len - ofs.ofs.cipher.head - ofs.ofs.cipher.tail; in build_dpaa_raw_dp_chain_fd() 408 auth_len = data_len - ofs.ofs.auth.head - ofs.ofs.auth.tail; in build_dpaa_raw_dp_chain_fd() 503 sg->offset = ofs.ofs.auth.head; in build_dpaa_raw_dp_chain_fd() [all …]
|
| /dpdk/drivers/crypto/dpaa2_sec/ |
| H A D | dpaa2_sec_raw_dp.c | 44 uint16_t auth_hdr_len = ofs.ofs.cipher.head - in build_raw_dp_chain_fd() 45 ofs.ofs.auth.head; in build_raw_dp_chain_fd() 56 cipher_len = data_len - ofs.ofs.cipher.head - ofs.ofs.cipher.tail; in build_raw_dp_chain_fd() 57 auth_len = data_len - ofs.ofs.auth.head - ofs.ofs.auth.tail; in build_raw_dp_chain_fd() 109 sge->length -= ofs.ofs.cipher.tail; in build_raw_dp_chain_fd() 123 sge->length -= ofs.ofs.cipher.tail; in build_raw_dp_chain_fd() 207 aead_len = data_len - ofs.ofs.cipher.head - ofs.ofs.cipher.tail; in build_raw_dp_aead_fd() 359 data_len = total_len - ofs.ofs.auth.head - ofs.ofs.auth.tail; in build_raw_dp_auth_fd() 360 data_offset = ofs.ofs.auth.head; in build_raw_dp_auth_fd() 604 data_len = total_len - ofs.ofs.cipher.head - ofs.ofs.cipher.tail; in build_raw_dp_cipher_fd() [all …]
|
| /dpdk/drivers/crypto/qat/dev/ |
| H A D | qat_crypto_pmd_gens.h | 487 ofs.ofs.cipher.head = cipher_ofs; in qat_sym_convert_op_to_vec_chain() 489 ofs.ofs.auth.head = auth_ofs; in qat_sym_convert_op_to_vec_chain() 584 ofs.ofs.cipher.tail; in enqueue_one_cipher_job_gen1() 601 auth_param->auth_off = ofs.ofs.auth.head; in enqueue_one_auth_job_gen1() 603 ofs.ofs.auth.tail; in enqueue_one_auth_job_gen1() 650 ofs.ofs.cipher.tail; in enqueue_one_chain_job_gen1() 651 auth_len = data_len - ofs.ofs.auth.head - ofs.ofs.auth.tail; in enqueue_one_chain_job_gen1() 660 auth_param->auth_off = ofs.ofs.auth.head; in enqueue_one_chain_job_gen1() 748 ofs.ofs.cipher.head); in enqueue_one_aead_job_gen1() 815 ofs.ofs.cipher.tail; in enqueue_one_aead_job_gen1() [all …]
|
| H A D | qat_crypto_pmd_gen3.c | 164 cipher_param->cipher_offset = ofs.ofs.cipher.head; in enqueue_one_aead_job_gen3() 165 cipher_param->cipher_length = data_len - ofs.ofs.cipher.head - in enqueue_one_aead_job_gen3() 166 ofs.ofs.cipher.tail; in enqueue_one_aead_job_gen3() 188 uint32_t auth_data_len = data_len - ofs.ofs.auth.head - in enqueue_one_auth_job_gen3() 189 ofs.ofs.auth.tail; in enqueue_one_auth_job_gen3() 274 union rte_crypto_sym_ofs ofs; in qat_sym_build_op_aead_gen3() local 285 if (unlikely(ofs.raw == UINT64_MAX)) { in qat_sym_build_op_aead_gen3() 320 union rte_crypto_sym_ofs ofs; in qat_sym_build_op_auth_gen3() local 344 ofs, total_len); in qat_sym_build_op_auth_gen3() 403 union rte_crypto_sym_ofs ofs, in qat_sym_dp_enqueue_single_aead_gen3() argument [all …]
|
| H A D | qat_crypto_pmd_gen4.c | 112 union rte_crypto_sym_ofs ofs, uint32_t data_len) in enqueue_one_aead_job_gen4() argument 125 cipher_param->cipher_offset = ofs.ofs.cipher.head; in enqueue_one_aead_job_gen4() 127 ofs.ofs.cipher.head - ofs.ofs.cipher.tail; in enqueue_one_aead_job_gen4() 135 enqueue_one_aead_job_gen1(ctx, req, iv, digest, aad, ofs, data_len); in enqueue_one_aead_job_gen4() 151 union rte_crypto_sym_ofs ofs; in qat_sym_build_op_aead_gen4() local 160 ofs.raw = qat_sym_convert_op_to_vec_aead(op, ctx, &in_sgl, &out_sgl, in qat_sym_build_op_aead_gen4() 162 if (unlikely(ofs.raw == UINT64_MAX)) { in qat_sym_build_op_aead_gen4() 232 union rte_crypto_sym_ofs ofs, in qat_sym_dp_enqueue_single_aead_gen4() argument 259 enqueue_one_aead_job_gen4(ctx, req, iv, digest, aad, ofs, in qat_sym_dp_enqueue_single_aead_gen4() 274 struct rte_crypto_sym_vec *vec, union rte_crypto_sym_ofs ofs, in qat_sym_dp_enqueue_aead_jobs_gen4() argument [all …]
|
| H A D | qat_sym_pmd_gen1.c | 193 union rte_crypto_sym_ofs ofs; in qat_sym_build_op_cipher_gen1() local 204 if (unlikely(ofs.raw == UINT64_MAX)) { in qat_sym_build_op_cipher_gen1() 238 union rte_crypto_sym_ofs ofs; in qat_sym_build_op_auth_gen1() local 249 if (unlikely(ofs.raw == UINT64_MAX)) { in qat_sym_build_op_auth_gen1() 285 union rte_crypto_sym_ofs ofs; in qat_sym_build_op_aead_gen1() local 332 union rte_crypto_sym_ofs ofs; in qat_sym_build_op_chain_gen1() local 357 ofs, total_len); in qat_sym_build_op_chain_gen1() 458 union rte_crypto_sym_ofs ofs, in qat_sym_dp_enqueue_single_cipher_gen1() argument 566 union rte_crypto_sym_ofs ofs, in qat_sym_dp_enqueue_single_auth_gen1() argument 674 union rte_crypto_sym_ofs ofs, in qat_sym_dp_enqueue_single_chain_gen1() argument [all …]
|
| /dpdk/lib/eal/common/ |
| H A D | eal_common_hexdump.c | 13 unsigned int i, out, ofs; in rte_hexdump() local 19 ofs = 0; in rte_hexdump() 20 while (ofs < len) { in rte_hexdump() 22 out = snprintf(line, LINE_LEN, "%08X:", ofs); in rte_hexdump() 24 if (ofs + i < len) in rte_hexdump() 26 " %02X", (data[ofs + i] & 0xff)); in rte_hexdump() 36 for (i = 0; ofs < len && i < 16; i++, ofs++) { in rte_hexdump() 37 unsigned char c = data[ofs]; in rte_hexdump()
|
| /dpdk/lib/ipsec/ |
| H A D | misc.h | 47 mbuf_get_seg_ofs(struct rte_mbuf *mb, uint32_t *ofs) in mbuf_get_seg_ofs() argument 53 n = *ofs; in mbuf_get_seg_ofs() 67 *ofs = n; in mbuf_get_seg_ofs() 84 mbuf_cut_seg_ofs(struct rte_mbuf *mb, struct rte_mbuf *ms, uint32_t ofs, in mbuf_cut_seg_ofs() argument 91 ms->data_len = ofs; in mbuf_cut_seg_ofs() 94 if (slen < ofs + len) { in mbuf_cut_seg_ofs() 114 union rte_crypto_sym_ofs ofs, struct rte_mbuf *mb[], in cpu_crypto_bulk() argument 148 ss->crypto.dev_id, ss->crypto.ses, ofs, in cpu_crypto_bulk() 171 ss->crypto.ses, ofs, &symvec); in cpu_crypto_bulk()
|
| H A D | esp_inb.c | 186 move_icv(struct rte_mbuf *ml, uint32_t ofs) in move_icv() argument 194 n = ml->data_len - ofs; in move_icv() 196 prev = rte_pktmbuf_mtod_offset(ml, const void *, ofs); in move_icv() 417 uint32_t ofs; in process_step1() local 419 ofs = mb->pkt_len - tlen; in process_step1() 421 ml[0] = mbuf_get_seg_ofs(mb, &ofs); in process_step1() 423 tofs[0] = ofs; in process_step1() 432 check_pad_bytes(struct rte_mbuf *mb, uint32_t ofs, uint32_t len) in check_pad_bytes() argument 438 k = mb->data_len - ofs; in check_pad_bytes() 440 pd = rte_pktmbuf_mtod_offset(mb, const uint8_t *, ofs); in check_pad_bytes() [all …]
|
| H A D | sa.c | 273 sa->cofs.ofs.cipher.tail = sa->sqh_len; in esp_inb_init() 277 sa->cofs.ofs.cipher.head = sa->ctp.cipher.offset - sa->ctp.auth.offset; in esp_inb_init() 345 sa->cofs.ofs.cipher.head = sa->ctp.cipher.offset - sa->ctp.auth.offset; in esp_outb_init() 346 sa->cofs.ofs.cipher.tail = (sa->ctp.auth.offset + sa->ctp.auth.length) - in esp_outb_init()
|
| /dpdk/lib/ip_frag/ |
| H A D | rte_ipv6_reassembly.c | 38 uint32_t i, n, ofs, first_len; in ipv6_frag_reassemble() local 47 ofs = fp->frags[IP_LAST_FRAG_IDX].ofs; in ipv6_frag_reassemble() 51 payload_len = ofs + last_len; in ipv6_frag_reassemble() 53 while (ofs != first_len) { in ipv6_frag_reassemble() 57 for (i = n; i != IP_FIRST_FRAG_IDX && ofs != first_len; i--) { in ipv6_frag_reassemble() 60 if (fp->frags[i].ofs + fp->frags[i].len == ofs) { in ipv6_frag_reassemble() 75 ofs = fp->frags[i].ofs; in ipv6_frag_reassemble()
|
| H A D | rte_ipv4_reassembly.c | 19 uint32_t i, n, ofs, first_len; in ipv4_frag_reassemble() local 27 ofs = fp->frags[IP_LAST_FRAG_IDX].ofs; in ipv4_frag_reassemble() 30 while (ofs != first_len) { in ipv4_frag_reassemble() 34 for (i = n; i != IP_FIRST_FRAG_IDX && ofs != first_len; i--) { in ipv4_frag_reassemble() 37 if(fp->frags[i].ofs + fp->frags[i].len == ofs) { in ipv4_frag_reassemble() 52 ofs = fp->frags[i].ofs; in ipv4_frag_reassemble()
|
| H A D | ip_frag_internal.c | 99 if (ofs == 0) { in ip_frag_process() 105 fp->total_size = ofs + len; in ip_frag_process() 130 fp->frags[IP_FIRST_FRAG_IDX].ofs, in ip_frag_process() 132 fp->frags[IP_LAST_FRAG_IDX].ofs, in ip_frag_process() 143 fp->frags[IP_FIRST_FRAG_IDX].ofs, in ip_frag_process() 145 fp->frags[IP_LAST_FRAG_IDX].ofs, in ip_frag_process() 156 fp->frags[idx].ofs = ofs; in ip_frag_process() 188 fp->frags[IP_FIRST_FRAG_IDX].ofs, in ip_frag_process() 190 fp->frags[IP_LAST_FRAG_IDX].ofs, in ip_frag_process() 201 fp->frags[IP_FIRST_FRAG_IDX].ofs, in ip_frag_process() [all …]
|
| H A D | ip_frag_common.h | 38 uint16_t ofs, uint16_t len, uint16_t more_frags); 141 .ofs = 0, in ip_frag_reset()
|
| H A D | ip_reassembly.h | 25 uint16_t ofs; /* offset into the packet */ member
|
| /dpdk/lib/bpf/ |
| H A D | bpf_jit_x86.c | 568 int32_t ofs) in emit_ld_reg() argument 590 emit_imm(st, ofs, imm_size(ofs)); in emit_ld_reg() 643 int32_t ofs) in emit_st_imm() argument 650 int32_t ofs) in emit_st_reg() argument 712 joff = ofs - st->sz; in emit_abs_jmp() 830 joff = ofs - st->sz; in emit_abs_jcc() 1183 RTE_VERIFY(ofs[LDMB_FIN_OFS] - ofs[LDMB_FSP_OFS] <= INT8_MAX); in emit_ld_mbuf() 1197 int32_t spil, ofs; in emit_prolog() local 1211 ofs = 0; in emit_prolog() 1241 int32_t spil, ofs; in emit_epilog() local [all …]
|
| H A D | bpf_load_elf.c | 57 resolve_xsym(const char *sn, size_t ofs, struct ebpf_insn *ins, size_t ins_sz, in resolve_xsym() argument 63 if (ofs % sizeof(ins[0]) != 0 || ofs >= ins_sz) in resolve_xsym() 66 idx = ofs / sizeof(ins[0]); in resolve_xsym() 70 ofs < ins_sz - sizeof(ins[idx])) in resolve_xsym() 189 size_t ofs, sym; in process_reloc() local 208 ofs = re[i].r_offset; in process_reloc() 217 rc = resolve_xsym(sn, ofs, ins, ins_sz, prm); in process_reloc() 221 sn, ofs, rc); in process_reloc()
|
| /dpdk/lib/cryptodev/ |
| H A D | rte_crypto_sym.h | 109 } ofs; member 954 rte_crypto_mbuf_to_vec(const struct rte_mbuf *mb, uint32_t ofs, uint32_t len, in rte_crypto_mbuf_to_vec() argument 963 RTE_ASSERT(mb->data_len > ofs); in rte_crypto_mbuf_to_vec() 968 vec[0].base = rte_pktmbuf_mtod_offset(mb, void *, ofs); in rte_crypto_mbuf_to_vec() 969 vec[0].iova = rte_pktmbuf_iova_offset(mb, ofs); in rte_crypto_mbuf_to_vec() 970 vec[0].tot_len = mb->buf_len - rte_pktmbuf_headroom(mb) - ofs; in rte_crypto_mbuf_to_vec() 973 seglen = mb->data_len - ofs; in rte_crypto_mbuf_to_vec() 986 vec[i].tot_len = mb->buf_len - rte_pktmbuf_headroom(mb) - ofs; in rte_crypto_mbuf_to_vec()
|
| H A D | rte_cryptodev.h | 1256 struct rte_cryptodev_sym_session *sess, union rte_crypto_sym_ofs ofs, 1309 union rte_crypto_sym_ofs ofs, void *user_data[], int *enqueue_status); 1334 uint16_t n_data_vecs, union rte_crypto_sym_ofs ofs, 1529 struct rte_crypto_sym_vec *vec, union rte_crypto_sym_ofs ofs, 1556 union rte_crypto_sym_ofs ofs, in rte_cryptodev_raw_enqueue() argument 1563 n_data_vecs, ofs, iv, digest, aad_or_auth_iv, user_data); in rte_cryptodev_raw_enqueue()
|
| /dpdk/drivers/net/netvsc/ |
| H A D | rndis.h | 165 #define RNDIS_PACKET_MSG_OFFSET_ABS(ofs) \ argument 166 ((ofs) + offsetof(struct rndis_packet_msg, dataoffset)) 297 #define RNDIS_QUERY_COMP_INFOBUFOFFSET_ABS(ofs) \ argument 298 ((ofs) + offsetof(struct rndis_query_comp, rid)) 369 #define RNDIS_STBUFOFFSET_ABS(ofs) \ argument 370 ((ofs) + offsetof(struct rndis_status_msg, status))
|
| H A D | hn_rndis.c | 275 sg.ofs = addr & PAGE_MASK; in hn_nvs_send_rndis_ctrl() 278 if (sg.ofs + reqlen > rte_mem_page_size()) { in hn_nvs_send_rndis_ctrl() 473 unsigned int ofs; in hn_rndis_query() local 525 ofs = RNDIS_QUERY_COMP_INFOBUFOFFSET_ABS(comp->infobufoffset); in hn_rndis_query() 526 if (ofs < sizeof(*comp) || ofs + comp->infobuflen > comp_len) { in hn_rndis_query() 538 memcpy(odata, (const char *)comp + ofs, odlen); in hn_rndis_query()
|
| H A D | hn_rxtx.c | 145 hn_rndis_pktmsg_offset(uint32_t ofs) in hn_rndis_pktmsg_offset() argument 147 return ofs - offsetof(struct rndis_packet_msg, dataoffset); in hn_rndis_pktmsg_offset() 815 unsigned int ofs, len; in hn_nvs_handle_rxbuf() local 817 ofs = pkt->rxbuf[i].ofs; in hn_nvs_handle_rxbuf() 820 if (unlikely(ofs + len > rxbuf_sz)) { in hn_nvs_handle_rxbuf() 823 i, ofs, len); in hn_nvs_handle_rxbuf() 833 rxbuf + ofs, len); in hn_nvs_handle_rxbuf() 1418 sg[segs].ofs = offset; in hn_fill_sg() 1459 sg[0].ofs = addr & PAGE_MASK; in hn_xmit_sg()
|
| /dpdk/examples/ipsec-secgw/ |
| H A D | ipsec_process.c | 148 uint32_t j, ofs, s; in copy_to_trf() local 158 ofs = offsetof(struct ip, ip_p); in copy_to_trf() 161 ofs = offsetof(struct ip6_hdr, ip6_nxt); in copy_to_trf() 165 ofs = offsetof(struct ip, ip_p); in copy_to_trf() 168 ofs = offsetof(struct ip6_hdr, ip6_nxt); in copy_to_trf() 173 void *, ofs); in copy_to_trf()
|
| /dpdk/drivers/bus/vmbus/ |
| H A D | rte_vmbus_reg.h | 142 uint32_t ofs; member 149 uint32_t ofs; member 174 uint32_t ofs; member
|
| /dpdk/lib/acl/ |
| H A D | acl_bld.c | 1443 size_t ofs, sz; in acl_build_rules() local 1447 ofs = n * sizeof(*br); in acl_build_rules() 1448 sz = ofs + n * fn * sizeof(*wp); in acl_build_rules() 1452 wp = (uint32_t *)((uintptr_t)br + ofs); in acl_build_rules() 1482 uint32_t i, n, ofs; in acl_set_data_indexes() local 1484 ofs = 0; in acl_set_data_indexes() 1489 ctx->trie[i].data_index = ctx->data_indexes + ofs; in acl_set_data_indexes() 1490 ofs += RTE_ACL_MAX_FIELDS; in acl_set_data_indexes() 1600 uint32_t i, max_ofs, ofs; in get_first_load_size() local 1602 ofs = 0; in get_first_load_size() [all …]
|